var/home/core/zuul-output/0000755000175000017500000000000015116422425014527 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015116434156015477 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000004751326515116434150017712 0ustar rootrootDec 11 02:00:57 crc systemd[1]: Starting Kubernetes Kubelet... Dec 11 02:00:57 crc restorecon[4689]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 11 02:00:57 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 11 02:00:58 crc restorecon[4689]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 11 02:00:58 crc restorecon[4689]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Dec 11 02:00:58 crc kubenswrapper[4824]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 11 02:00:58 crc kubenswrapper[4824]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Dec 11 02:00:58 crc kubenswrapper[4824]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 11 02:00:58 crc kubenswrapper[4824]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 11 02:00:58 crc kubenswrapper[4824]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Dec 11 02:00:58 crc kubenswrapper[4824]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.443851 4824 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.450674 4824 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.450723 4824 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.450734 4824 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.450747 4824 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.450761 4824 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.450771 4824 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.450780 4824 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.450790 4824 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.450799 4824 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.450808 4824 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.450817 4824 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.450825 4824 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.450833 4824 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.450841 4824 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.450851 4824 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.450860 4824 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.450869 4824 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.450877 4824 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.450885 4824 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.450895 4824 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.450905 4824 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.450917 4824 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.450929 4824 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.450939 4824 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.450950 4824 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.450961 4824 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.450971 4824 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.450982 4824 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.450993 4824 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.451004 4824 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.451014 4824 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.451025 4824 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.451036 4824 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.451049 4824 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.451060 4824 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.451069 4824 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.451078 4824 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.451086 4824 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.451094 4824 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.451103 4824 feature_gate.go:330] unrecognized feature gate: Example Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.451154 4824 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.451163 4824 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.451174 4824 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.451182 4824 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.451190 4824 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.451199 4824 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.451207 4824 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.451216 4824 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.451225 4824 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.451237 4824 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.451248 4824 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.451256 4824 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.451264 4824 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.451273 4824 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.451281 4824 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.451290 4824 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.451298 4824 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.451306 4824 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.451314 4824 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.451323 4824 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.451335 4824 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.451346 4824 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.451356 4824 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.451366 4824 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.451375 4824 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.451384 4824 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.451394 4824 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.451402 4824 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.451416 4824 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.451427 4824 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.451436 4824 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.451609 4824 flags.go:64] FLAG: --address="0.0.0.0" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.451635 4824 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.451650 4824 flags.go:64] FLAG: --anonymous-auth="true" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.451662 4824 flags.go:64] FLAG: --application-metrics-count-limit="100" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.451674 4824 flags.go:64] FLAG: --authentication-token-webhook="false" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.451685 4824 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.451697 4824 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.451709 4824 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.451719 4824 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.451729 4824 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.451739 4824 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.451750 4824 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.451760 4824 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.451770 4824 flags.go:64] FLAG: --cgroup-root="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.451780 4824 flags.go:64] FLAG: --cgroups-per-qos="true" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.451790 4824 flags.go:64] FLAG: --client-ca-file="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.451800 4824 flags.go:64] FLAG: --cloud-config="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.451809 4824 flags.go:64] FLAG: --cloud-provider="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.451819 4824 flags.go:64] FLAG: --cluster-dns="[]" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.451830 4824 flags.go:64] FLAG: --cluster-domain="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.451840 4824 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.451850 4824 flags.go:64] FLAG: --config-dir="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.451859 4824 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.451870 4824 flags.go:64] FLAG: --container-log-max-files="5" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.451882 4824 flags.go:64] FLAG: --container-log-max-size="10Mi" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.451892 4824 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.451903 4824 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.451913 4824 flags.go:64] FLAG: --containerd-namespace="k8s.io" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.451923 4824 flags.go:64] FLAG: --contention-profiling="false" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.451933 4824 flags.go:64] FLAG: --cpu-cfs-quota="true" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.451943 4824 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.451956 4824 flags.go:64] FLAG: --cpu-manager-policy="none" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.451969 4824 flags.go:64] FLAG: --cpu-manager-policy-options="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.451985 4824 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.451999 4824 flags.go:64] FLAG: --enable-controller-attach-detach="true" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452010 4824 flags.go:64] FLAG: --enable-debugging-handlers="true" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452020 4824 flags.go:64] FLAG: --enable-load-reader="false" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452030 4824 flags.go:64] FLAG: --enable-server="true" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452040 4824 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452052 4824 flags.go:64] FLAG: --event-burst="100" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452062 4824 flags.go:64] FLAG: --event-qps="50" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452072 4824 flags.go:64] FLAG: --event-storage-age-limit="default=0" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452083 4824 flags.go:64] FLAG: --event-storage-event-limit="default=0" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452093 4824 flags.go:64] FLAG: --eviction-hard="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452105 4824 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452145 4824 flags.go:64] FLAG: --eviction-minimum-reclaim="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452155 4824 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452168 4824 flags.go:64] FLAG: --eviction-soft="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452178 4824 flags.go:64] FLAG: --eviction-soft-grace-period="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452187 4824 flags.go:64] FLAG: --exit-on-lock-contention="false" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452197 4824 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452207 4824 flags.go:64] FLAG: --experimental-mounter-path="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452217 4824 flags.go:64] FLAG: --fail-cgroupv1="false" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452228 4824 flags.go:64] FLAG: --fail-swap-on="true" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452238 4824 flags.go:64] FLAG: --feature-gates="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452251 4824 flags.go:64] FLAG: --file-check-frequency="20s" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452261 4824 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452272 4824 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452282 4824 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452292 4824 flags.go:64] FLAG: --healthz-port="10248" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452301 4824 flags.go:64] FLAG: --help="false" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452311 4824 flags.go:64] FLAG: --hostname-override="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452320 4824 flags.go:64] FLAG: --housekeeping-interval="10s" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452330 4824 flags.go:64] FLAG: --http-check-frequency="20s" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452340 4824 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452350 4824 flags.go:64] FLAG: --image-credential-provider-config="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452360 4824 flags.go:64] FLAG: --image-gc-high-threshold="85" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452370 4824 flags.go:64] FLAG: --image-gc-low-threshold="80" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452380 4824 flags.go:64] FLAG: --image-service-endpoint="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452389 4824 flags.go:64] FLAG: --kernel-memcg-notification="false" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452399 4824 flags.go:64] FLAG: --kube-api-burst="100" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452408 4824 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452419 4824 flags.go:64] FLAG: --kube-api-qps="50" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452429 4824 flags.go:64] FLAG: --kube-reserved="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452438 4824 flags.go:64] FLAG: --kube-reserved-cgroup="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452447 4824 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452458 4824 flags.go:64] FLAG: --kubelet-cgroups="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452467 4824 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452477 4824 flags.go:64] FLAG: --lock-file="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452486 4824 flags.go:64] FLAG: --log-cadvisor-usage="false" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452496 4824 flags.go:64] FLAG: --log-flush-frequency="5s" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452505 4824 flags.go:64] FLAG: --log-json-info-buffer-size="0" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452520 4824 flags.go:64] FLAG: --log-json-split-stream="false" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452531 4824 flags.go:64] FLAG: --log-text-info-buffer-size="0" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452542 4824 flags.go:64] FLAG: --log-text-split-stream="false" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452551 4824 flags.go:64] FLAG: --logging-format="text" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452561 4824 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452571 4824 flags.go:64] FLAG: --make-iptables-util-chains="true" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452581 4824 flags.go:64] FLAG: --manifest-url="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452591 4824 flags.go:64] FLAG: --manifest-url-header="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452604 4824 flags.go:64] FLAG: --max-housekeeping-interval="15s" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452614 4824 flags.go:64] FLAG: --max-open-files="1000000" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452625 4824 flags.go:64] FLAG: --max-pods="110" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452635 4824 flags.go:64] FLAG: --maximum-dead-containers="-1" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452645 4824 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452655 4824 flags.go:64] FLAG: --memory-manager-policy="None" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452666 4824 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452676 4824 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452686 4824 flags.go:64] FLAG: --node-ip="192.168.126.11" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452697 4824 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452720 4824 flags.go:64] FLAG: --node-status-max-images="50" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452730 4824 flags.go:64] FLAG: --node-status-update-frequency="10s" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452783 4824 flags.go:64] FLAG: --oom-score-adj="-999" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452794 4824 flags.go:64] FLAG: --pod-cidr="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452804 4824 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452819 4824 flags.go:64] FLAG: --pod-manifest-path="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452829 4824 flags.go:64] FLAG: --pod-max-pids="-1" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452839 4824 flags.go:64] FLAG: --pods-per-core="0" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452848 4824 flags.go:64] FLAG: --port="10250" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452859 4824 flags.go:64] FLAG: --protect-kernel-defaults="false" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452868 4824 flags.go:64] FLAG: --provider-id="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452878 4824 flags.go:64] FLAG: --qos-reserved="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452888 4824 flags.go:64] FLAG: --read-only-port="10255" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452897 4824 flags.go:64] FLAG: --register-node="true" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452907 4824 flags.go:64] FLAG: --register-schedulable="true" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452917 4824 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452933 4824 flags.go:64] FLAG: --registry-burst="10" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452943 4824 flags.go:64] FLAG: --registry-qps="5" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452952 4824 flags.go:64] FLAG: --reserved-cpus="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452964 4824 flags.go:64] FLAG: --reserved-memory="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452977 4824 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452986 4824 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.452996 4824 flags.go:64] FLAG: --rotate-certificates="false" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.453006 4824 flags.go:64] FLAG: --rotate-server-certificates="false" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.453015 4824 flags.go:64] FLAG: --runonce="false" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.453025 4824 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.453035 4824 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.453045 4824 flags.go:64] FLAG: --seccomp-default="false" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.453056 4824 flags.go:64] FLAG: --serialize-image-pulls="true" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.453066 4824 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.453076 4824 flags.go:64] FLAG: --storage-driver-db="cadvisor" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.453087 4824 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.453097 4824 flags.go:64] FLAG: --storage-driver-password="root" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.453107 4824 flags.go:64] FLAG: --storage-driver-secure="false" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.453150 4824 flags.go:64] FLAG: --storage-driver-table="stats" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.453164 4824 flags.go:64] FLAG: --storage-driver-user="root" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.453176 4824 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.453189 4824 flags.go:64] FLAG: --sync-frequency="1m0s" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.453201 4824 flags.go:64] FLAG: --system-cgroups="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.453213 4824 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.453233 4824 flags.go:64] FLAG: --system-reserved-cgroup="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.453243 4824 flags.go:64] FLAG: --tls-cert-file="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.453253 4824 flags.go:64] FLAG: --tls-cipher-suites="[]" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.453266 4824 flags.go:64] FLAG: --tls-min-version="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.453276 4824 flags.go:64] FLAG: --tls-private-key-file="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.453285 4824 flags.go:64] FLAG: --topology-manager-policy="none" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.453295 4824 flags.go:64] FLAG: --topology-manager-policy-options="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.453305 4824 flags.go:64] FLAG: --topology-manager-scope="container" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.453316 4824 flags.go:64] FLAG: --v="2" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.453328 4824 flags.go:64] FLAG: --version="false" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.453343 4824 flags.go:64] FLAG: --vmodule="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.453358 4824 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.453371 4824 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.453610 4824 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.453623 4824 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.453636 4824 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.453647 4824 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.453656 4824 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.453666 4824 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.453675 4824 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.453684 4824 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.453695 4824 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.453706 4824 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.453716 4824 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.453725 4824 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.453733 4824 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.453742 4824 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.453752 4824 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.453760 4824 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.453769 4824 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.453778 4824 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.453788 4824 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.453797 4824 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.453805 4824 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.453815 4824 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.453825 4824 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.453833 4824 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.453842 4824 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.453850 4824 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.453859 4824 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.453879 4824 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.453888 4824 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.453897 4824 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.453905 4824 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.453913 4824 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.453921 4824 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.453930 4824 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.453938 4824 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.453947 4824 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.453955 4824 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.453963 4824 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.453973 4824 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.453981 4824 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.453990 4824 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.453998 4824 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.454006 4824 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.454014 4824 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.454022 4824 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.454031 4824 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.454040 4824 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.454048 4824 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.454056 4824 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.454064 4824 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.454073 4824 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.454081 4824 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.454089 4824 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.454097 4824 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.454105 4824 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.454147 4824 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.454155 4824 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.454163 4824 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.454175 4824 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.454190 4824 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.454201 4824 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.454211 4824 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.454220 4824 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.454230 4824 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.454239 4824 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.454250 4824 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.454260 4824 feature_gate.go:330] unrecognized feature gate: Example Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.454270 4824 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.454278 4824 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.454287 4824 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.454295 4824 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.454310 4824 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.466361 4824 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.466402 4824 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.466548 4824 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.466563 4824 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.466573 4824 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.466583 4824 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.466595 4824 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.466607 4824 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.466616 4824 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.466626 4824 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.466635 4824 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.466644 4824 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.466653 4824 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.466661 4824 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.466670 4824 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.466678 4824 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.466686 4824 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.466695 4824 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.466703 4824 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.466711 4824 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.466719 4824 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.466729 4824 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.466740 4824 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.466750 4824 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.466761 4824 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.466771 4824 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.466783 4824 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.466794 4824 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.466805 4824 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.466816 4824 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.466826 4824 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.466837 4824 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.466847 4824 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.466858 4824 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.466869 4824 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.466881 4824 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.466894 4824 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.466907 4824 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.466919 4824 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.466930 4824 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.466940 4824 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.466957 4824 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.466972 4824 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.466988 4824 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.467003 4824 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.467017 4824 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.467030 4824 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.467042 4824 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.467055 4824 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.467069 4824 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.467080 4824 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.467089 4824 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.467097 4824 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.467106 4824 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.467151 4824 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.467163 4824 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.467173 4824 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.467184 4824 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.467194 4824 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.467203 4824 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.467211 4824 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.467219 4824 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.467228 4824 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.467236 4824 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.467245 4824 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.467253 4824 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.467262 4824 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.467271 4824 feature_gate.go:330] unrecognized feature gate: Example Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.467279 4824 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.467287 4824 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.467295 4824 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.467304 4824 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.467316 4824 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.467330 4824 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.467582 4824 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.467595 4824 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.467605 4824 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.467615 4824 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.467624 4824 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.467633 4824 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.467642 4824 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.467651 4824 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.467659 4824 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.467667 4824 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.467676 4824 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.467684 4824 feature_gate.go:330] unrecognized feature gate: Example Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.467694 4824 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.467702 4824 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.467710 4824 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.467719 4824 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.467727 4824 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.467735 4824 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.467743 4824 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.467752 4824 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.467760 4824 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.467769 4824 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.467777 4824 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.467788 4824 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.467796 4824 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.467804 4824 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.467813 4824 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.467821 4824 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.467829 4824 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.467840 4824 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.467851 4824 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.467862 4824 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.467872 4824 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.467882 4824 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.467893 4824 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.467905 4824 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.467916 4824 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.467925 4824 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.467934 4824 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.467942 4824 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.467951 4824 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.467959 4824 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.467968 4824 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.467976 4824 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.467984 4824 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.467992 4824 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.468001 4824 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.468010 4824 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.468018 4824 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.468026 4824 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.468034 4824 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.468042 4824 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.468051 4824 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.468059 4824 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.468067 4824 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.468076 4824 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.468084 4824 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.468093 4824 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.468101 4824 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.468135 4824 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.468144 4824 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.468155 4824 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.468165 4824 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.468175 4824 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.468184 4824 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.468193 4824 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.468202 4824 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.468213 4824 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.468223 4824 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.468231 4824 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.468241 4824 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.468253 4824 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.468483 4824 server.go:940] "Client rotation is on, will bootstrap in background" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.473022 4824 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.473213 4824 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.474035 4824 server.go:997] "Starting client certificate rotation" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.474083 4824 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.474408 4824 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-11-23 03:52:32.916485525 +0000 UTC Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.474562 4824 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.481216 4824 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 11 02:00:58 crc kubenswrapper[4824]: E1211 02:00:58.482220 4824 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.50:6443: connect: connection refused" logger="UnhandledError" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.483638 4824 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.496280 4824 log.go:25] "Validated CRI v1 runtime API" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.518985 4824 log.go:25] "Validated CRI v1 image API" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.521233 4824 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.524379 4824 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-12-11-01-56-31-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.524434 4824 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:41 fsType:tmpfs blockSize:0}] Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.550224 4824 manager.go:217] Machine: {Timestamp:2025-12-11 02:00:58.54821759 +0000 UTC m=+0.237255049 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2799998 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:f76b32e4-2764-4219-bbf4-f7e56d76df3b BootID:d57bece6-6fc6-4a84-b8a8-e1678f9e50db Filesystems:[{Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:41 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:c9:71:94 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:c9:71:94 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:e5:5d:c8 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:cc:a9:97 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:68:2c:81 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:ac:10:6b Speed:-1 Mtu:1496} {Name:eth10 MacAddress:4e:b9:87:ea:8d:04 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:d6:3a:a4:fe:f1:37 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.550601 4824 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.550795 4824 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.551660 4824 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.551956 4824 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.552010 4824 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.552370 4824 topology_manager.go:138] "Creating topology manager with none policy" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.552388 4824 container_manager_linux.go:303] "Creating device plugin manager" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.552695 4824 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.552744 4824 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.553003 4824 state_mem.go:36] "Initialized new in-memory state store" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.553161 4824 server.go:1245] "Using root directory" path="/var/lib/kubelet" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.554004 4824 kubelet.go:418] "Attempting to sync node with API server" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.554035 4824 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.554081 4824 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.554102 4824 kubelet.go:324] "Adding apiserver pod source" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.554186 4824 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.556197 4824 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.556527 4824 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.50:6443: connect: connection refused Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.556625 4824 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Dec 11 02:00:58 crc kubenswrapper[4824]: E1211 02:00:58.556666 4824 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.50:6443: connect: connection refused" logger="UnhandledError" Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.556961 4824 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.50:6443: connect: connection refused Dec 11 02:00:58 crc kubenswrapper[4824]: E1211 02:00:58.557081 4824 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.50:6443: connect: connection refused" logger="UnhandledError" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.557751 4824 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.558622 4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.558663 4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.558678 4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.558692 4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.558714 4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.558727 4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.558741 4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.558763 4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.558780 4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.558794 4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.558813 4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.558826 4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.559101 4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.560211 4824 server.go:1280] "Started kubelet" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.560378 4824 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.50:6443: connect: connection refused Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.560659 4824 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.560886 4824 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.561530 4824 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 11 02:00:58 crc systemd[1]: Started Kubernetes Kubelet. Dec 11 02:00:58 crc kubenswrapper[4824]: E1211 02:00:58.564017 4824 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.50:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.188006bc065e804a default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 02:00:58.559709258 +0000 UTC m=+0.248746677,LastTimestamp:2025-12-11 02:00:58.559709258 +0000 UTC m=+0.248746677,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.565586 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.565648 4824 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.565697 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-27 06:05:34.976344744 +0000 UTC Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.566177 4824 volume_manager.go:287] "The desired_state_of_world populator starts" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.566212 4824 volume_manager.go:289] "Starting Kubelet Volume Manager" Dec 11 02:00:58 crc kubenswrapper[4824]: E1211 02:00:58.566203 4824 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.566288 4824 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.567889 4824 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.567944 4824 factory.go:55] Registering systemd factory Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.567967 4824 factory.go:221] Registration of the systemd container factory successfully Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.568280 4824 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.50:6443: connect: connection refused Dec 11 02:00:58 crc kubenswrapper[4824]: E1211 02:00:58.568433 4824 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.50:6443: connect: connection refused" logger="UnhandledError" Dec 11 02:00:58 crc kubenswrapper[4824]: E1211 02:00:58.569600 4824 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.50:6443: connect: connection refused" interval="200ms" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.572328 4824 server.go:460] "Adding debug handlers to kubelet server" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.576221 4824 factory.go:153] Registering CRI-O factory Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.576270 4824 factory.go:221] Registration of the crio container factory successfully Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.576301 4824 factory.go:103] Registering Raw factory Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.576324 4824 manager.go:1196] Started watching for new ooms in manager Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.577261 4824 manager.go:319] Starting recovery of all containers Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.577882 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.577933 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.577946 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.577959 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.577971 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.577983 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.577994 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.578006 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.578019 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.578030 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.578044 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.578057 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.578068 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.578081 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.578095 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.578106 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.578172 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.578184 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.578197 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.578209 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.578756 4824 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.578798 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.578813 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.578834 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.578847 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.578860 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.578872 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.578891 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.578904 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.578918 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.578929 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.579167 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.579192 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.579204 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.579216 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.579230 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.579279 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.579295 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.579309 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.579322 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.579334 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.579346 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.579358 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.579369 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.579380 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.579395 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.579407 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.579421 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.579432 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.579444 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.579456 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.579468 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.579479 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.579546 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.579563 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.579575 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.579599 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.579610 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.579623 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.579635 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.579646 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.579660 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.579672 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.579684 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.579697 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.579709 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.579720 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.579733 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.579745 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.579757 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.579769 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.579781 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.579791 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.579803 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.579814 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.579826 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.579838 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.579852 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.579863 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.579873 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.579884 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.579895 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.579907 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.579919 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.579930 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.579941 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.579953 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.579966 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.579978 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.579989 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.580001 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.580012 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.580025 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.580038 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.580049 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.580060 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.580071 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.580082 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.580094 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.580105 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.580134 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.580144 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.580156 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.580170 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.580182 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.580201 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.580214 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.580227 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.580239 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.580254 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.580265 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.580278 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.580290 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.580302 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.580314 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.580325 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.580337 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.580348 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.580359 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.580370 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.580381 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.580392 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.580404 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.580417 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.580600 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.580620 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.580632 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.580644 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.580656 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.580668 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.580680 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.580692 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.580705 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.580716 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.580727 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.580740 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.580753 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.580765 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.580778 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.580789 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.580800 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.580831 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.580843 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.580855 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.580882 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.580894 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.580906 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.580918 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.580931 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.580944 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.580955 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.580968 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.580981 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.580993 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.581036 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.581049 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.581061 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.581075 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.581087 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.581099 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.581126 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.581138 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.581150 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.581163 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.581177 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.581189 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.581202 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.581214 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.581226 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.581237 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.581248 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.581261 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.581273 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.581291 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.581303 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.581316 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.581328 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.581339 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.581352 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.581363 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.581375 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.581386 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.581398 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.581410 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.581421 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.581433 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.581444 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.581455 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.581467 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.581481 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.581493 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.581506 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.581517 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.581528 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.581539 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.581555 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.581567 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.581579 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.581592 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.581603 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.581615 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.581626 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.581639 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.581652 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.581666 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.581683 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.581696 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.581711 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.581723 4824 reconstruct.go:97] "Volume reconstruction finished" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.581732 4824 reconciler.go:26] "Reconciler: start to sync state" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.610346 4824 manager.go:324] Recovery completed Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.625666 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.627578 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.627625 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.627636 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.628169 4824 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.628594 4824 cpu_manager.go:225] "Starting CPU manager" policy="none" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.628615 4824 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.628645 4824 state_mem.go:36] "Initialized new in-memory state store" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.631017 4824 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.631230 4824 status_manager.go:217] "Starting to sync pod status with apiserver" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.631275 4824 kubelet.go:2335] "Starting kubelet main sync loop" Dec 11 02:00:58 crc kubenswrapper[4824]: E1211 02:00:58.631321 4824 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.631904 4824 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.50:6443: connect: connection refused Dec 11 02:00:58 crc kubenswrapper[4824]: E1211 02:00:58.631981 4824 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.50:6443: connect: connection refused" logger="UnhandledError" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.640565 4824 policy_none.go:49] "None policy: Start" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.644511 4824 memory_manager.go:170] "Starting memorymanager" policy="None" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.644544 4824 state_mem.go:35] "Initializing new in-memory state store" Dec 11 02:00:58 crc kubenswrapper[4824]: E1211 02:00:58.667000 4824 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.718993 4824 manager.go:334] "Starting Device Plugin manager" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.719075 4824 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.719096 4824 server.go:79] "Starting device plugin registration server" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.719712 4824 eviction_manager.go:189] "Eviction manager: starting control loop" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.719737 4824 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.720047 4824 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.720184 4824 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.720202 4824 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.731658 4824 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.731768 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.733224 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.733283 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.733302 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.733545 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.733901 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.733958 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.734820 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.734861 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.734879 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.735107 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.735243 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.735285 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.736074 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.736459 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.736498 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.736518 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.736553 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.736574 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.736670 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.736729 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:00:58 crc kubenswrapper[4824]: E1211 02:00:58.736735 4824 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.736751 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.736687 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.736826 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.736874 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.738602 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.738660 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.738686 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.738676 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.738767 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.738781 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.738972 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.739141 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.739223 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.740526 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.740543 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.740576 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.740588 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.740577 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.740613 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.740876 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.740918 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.742040 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.742107 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.742167 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:00:58 crc kubenswrapper[4824]: E1211 02:00:58.770176 4824 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.50:6443: connect: connection refused" interval="400ms" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.783776 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.783853 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.783901 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.783984 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.784134 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.784238 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.784315 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.784347 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.784399 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.784436 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.784466 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.784496 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.784586 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.784657 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.784692 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.819902 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.821586 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.821636 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.821650 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.821684 4824 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 11 02:00:58 crc kubenswrapper[4824]: E1211 02:00:58.822290 4824 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.50:6443: connect: connection refused" node="crc" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.885711 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.885783 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.885819 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.885850 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.885882 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.885912 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.885938 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.885968 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.885999 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.886028 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.886056 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.886085 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.886135 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.886166 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.886296 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.886327 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.886378 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.886438 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.886390 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.886399 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.886440 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.886481 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.886512 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.886519 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.886148 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.886649 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.886710 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.886751 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.886766 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.887451 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.895442 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 02:00:58 crc kubenswrapper[4824]: I1211 02:00:58.907151 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.934571 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-ca184d8de448270bb890a6174c618d07cd52baaafc94bd156172ab72ca5fee7b WatchSource:0}: Error finding container ca184d8de448270bb890a6174c618d07cd52baaafc94bd156172ab72ca5fee7b: Status 404 returned error can't find the container with id ca184d8de448270bb890a6174c618d07cd52baaafc94bd156172ab72ca5fee7b Dec 11 02:00:58 crc kubenswrapper[4824]: W1211 02:00:58.938417 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-9c69caebc758f3ce84f5a02996e99f6dd7262c8323502988a794d3da0b4d3359 WatchSource:0}: Error finding container 9c69caebc758f3ce84f5a02996e99f6dd7262c8323502988a794d3da0b4d3359: Status 404 returned error can't find the container with id 9c69caebc758f3ce84f5a02996e99f6dd7262c8323502988a794d3da0b4d3359 Dec 11 02:00:59 crc kubenswrapper[4824]: I1211 02:00:59.023079 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 02:00:59 crc kubenswrapper[4824]: I1211 02:00:59.025046 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:00:59 crc kubenswrapper[4824]: I1211 02:00:59.025140 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:00:59 crc kubenswrapper[4824]: I1211 02:00:59.025163 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:00:59 crc kubenswrapper[4824]: I1211 02:00:59.025210 4824 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 11 02:00:59 crc kubenswrapper[4824]: E1211 02:00:59.026017 4824 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.50:6443: connect: connection refused" node="crc" Dec 11 02:00:59 crc kubenswrapper[4824]: I1211 02:00:59.100399 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 11 02:00:59 crc kubenswrapper[4824]: W1211 02:00:59.120068 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-466052fc031190b5f9f39fd0de4679d7f40f6cf756cf75c0bb9a3e8f4bbfbdf8 WatchSource:0}: Error finding container 466052fc031190b5f9f39fd0de4679d7f40f6cf756cf75c0bb9a3e8f4bbfbdf8: Status 404 returned error can't find the container with id 466052fc031190b5f9f39fd0de4679d7f40f6cf756cf75c0bb9a3e8f4bbfbdf8 Dec 11 02:00:59 crc kubenswrapper[4824]: I1211 02:00:59.133994 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 11 02:00:59 crc kubenswrapper[4824]: W1211 02:00:59.148452 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-7f4bb87a3a668ea065ad0435d36285e8705dfa4fa1677afe3417fd4473645ed9 WatchSource:0}: Error finding container 7f4bb87a3a668ea065ad0435d36285e8705dfa4fa1677afe3417fd4473645ed9: Status 404 returned error can't find the container with id 7f4bb87a3a668ea065ad0435d36285e8705dfa4fa1677afe3417fd4473645ed9 Dec 11 02:00:59 crc kubenswrapper[4824]: I1211 02:00:59.162080 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 02:00:59 crc kubenswrapper[4824]: E1211 02:00:59.171664 4824 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.50:6443: connect: connection refused" interval="800ms" Dec 11 02:00:59 crc kubenswrapper[4824]: W1211 02:00:59.178573 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-a1154b0752f77d378fadb4e171acebc3c24aa9c7559bf92ed49096f8765d0e67 WatchSource:0}: Error finding container a1154b0752f77d378fadb4e171acebc3c24aa9c7559bf92ed49096f8765d0e67: Status 404 returned error can't find the container with id a1154b0752f77d378fadb4e171acebc3c24aa9c7559bf92ed49096f8765d0e67 Dec 11 02:00:59 crc kubenswrapper[4824]: E1211 02:00:59.403128 4824 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.50:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.188006bc065e804a default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 02:00:58.559709258 +0000 UTC m=+0.248746677,LastTimestamp:2025-12-11 02:00:58.559709258 +0000 UTC m=+0.248746677,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 02:00:59 crc kubenswrapper[4824]: I1211 02:00:59.426187 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 02:00:59 crc kubenswrapper[4824]: I1211 02:00:59.427935 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:00:59 crc kubenswrapper[4824]: I1211 02:00:59.427973 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:00:59 crc kubenswrapper[4824]: I1211 02:00:59.427988 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:00:59 crc kubenswrapper[4824]: I1211 02:00:59.428014 4824 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 11 02:00:59 crc kubenswrapper[4824]: E1211 02:00:59.428508 4824 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.50:6443: connect: connection refused" node="crc" Dec 11 02:00:59 crc kubenswrapper[4824]: I1211 02:00:59.561921 4824 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.50:6443: connect: connection refused Dec 11 02:00:59 crc kubenswrapper[4824]: I1211 02:00:59.566579 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-10 11:35:55.310889053 +0000 UTC Dec 11 02:00:59 crc kubenswrapper[4824]: I1211 02:00:59.636737 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"922af7f67b99a0ce08776f3f75c0c377e6ea69af1720b6178a6d1e2f9b4b8ba3"} Dec 11 02:00:59 crc kubenswrapper[4824]: I1211 02:00:59.636950 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"9c69caebc758f3ce84f5a02996e99f6dd7262c8323502988a794d3da0b4d3359"} Dec 11 02:00:59 crc kubenswrapper[4824]: I1211 02:00:59.639386 4824 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62" exitCode=0 Dec 11 02:00:59 crc kubenswrapper[4824]: I1211 02:00:59.639478 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62"} Dec 11 02:00:59 crc kubenswrapper[4824]: I1211 02:00:59.639555 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"a1154b0752f77d378fadb4e171acebc3c24aa9c7559bf92ed49096f8765d0e67"} Dec 11 02:00:59 crc kubenswrapper[4824]: I1211 02:00:59.639714 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 02:00:59 crc kubenswrapper[4824]: I1211 02:00:59.640785 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:00:59 crc kubenswrapper[4824]: I1211 02:00:59.640832 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:00:59 crc kubenswrapper[4824]: I1211 02:00:59.640852 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:00:59 crc kubenswrapper[4824]: I1211 02:00:59.641294 4824 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="f1c9f0af0451936c800a4b5e764759b7c3b07a957566188e7a7166d1ecf0cf4e" exitCode=0 Dec 11 02:00:59 crc kubenswrapper[4824]: I1211 02:00:59.641402 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"f1c9f0af0451936c800a4b5e764759b7c3b07a957566188e7a7166d1ecf0cf4e"} Dec 11 02:00:59 crc kubenswrapper[4824]: I1211 02:00:59.641445 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"7f4bb87a3a668ea065ad0435d36285e8705dfa4fa1677afe3417fd4473645ed9"} Dec 11 02:00:59 crc kubenswrapper[4824]: I1211 02:00:59.641571 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 02:00:59 crc kubenswrapper[4824]: W1211 02:00:59.642474 4824 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.50:6443: connect: connection refused Dec 11 02:00:59 crc kubenswrapper[4824]: E1211 02:00:59.642564 4824 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.50:6443: connect: connection refused" logger="UnhandledError" Dec 11 02:00:59 crc kubenswrapper[4824]: I1211 02:00:59.642963 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 02:00:59 crc kubenswrapper[4824]: I1211 02:00:59.643002 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:00:59 crc kubenswrapper[4824]: I1211 02:00:59.643031 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:00:59 crc kubenswrapper[4824]: I1211 02:00:59.643048 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:00:59 crc kubenswrapper[4824]: I1211 02:00:59.644629 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:00:59 crc kubenswrapper[4824]: I1211 02:00:59.644673 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:00:59 crc kubenswrapper[4824]: I1211 02:00:59.644692 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:00:59 crc kubenswrapper[4824]: I1211 02:00:59.644913 4824 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="18a2507c91dd7ab306ef38abf724e07f468190caba6cb110bf3224448f631aad" exitCode=0 Dec 11 02:00:59 crc kubenswrapper[4824]: I1211 02:00:59.645013 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"18a2507c91dd7ab306ef38abf724e07f468190caba6cb110bf3224448f631aad"} Dec 11 02:00:59 crc kubenswrapper[4824]: I1211 02:00:59.645051 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"466052fc031190b5f9f39fd0de4679d7f40f6cf756cf75c0bb9a3e8f4bbfbdf8"} Dec 11 02:00:59 crc kubenswrapper[4824]: I1211 02:00:59.645203 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 02:00:59 crc kubenswrapper[4824]: I1211 02:00:59.652749 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:00:59 crc kubenswrapper[4824]: I1211 02:00:59.652790 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:00:59 crc kubenswrapper[4824]: I1211 02:00:59.652807 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:00:59 crc kubenswrapper[4824]: I1211 02:00:59.654671 4824 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="5e27af6c438009a44288f8bc10f8ebf2b61d9309f87e8d955e8d255a7190174f" exitCode=0 Dec 11 02:00:59 crc kubenswrapper[4824]: I1211 02:00:59.654721 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"5e27af6c438009a44288f8bc10f8ebf2b61d9309f87e8d955e8d255a7190174f"} Dec 11 02:00:59 crc kubenswrapper[4824]: I1211 02:00:59.654760 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"ca184d8de448270bb890a6174c618d07cd52baaafc94bd156172ab72ca5fee7b"} Dec 11 02:00:59 crc kubenswrapper[4824]: I1211 02:00:59.654899 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 02:00:59 crc kubenswrapper[4824]: I1211 02:00:59.655930 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:00:59 crc kubenswrapper[4824]: I1211 02:00:59.655976 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:00:59 crc kubenswrapper[4824]: I1211 02:00:59.655992 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:00:59 crc kubenswrapper[4824]: W1211 02:00:59.673197 4824 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.50:6443: connect: connection refused Dec 11 02:00:59 crc kubenswrapper[4824]: E1211 02:00:59.673296 4824 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.50:6443: connect: connection refused" logger="UnhandledError" Dec 11 02:00:59 crc kubenswrapper[4824]: W1211 02:00:59.900038 4824 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.50:6443: connect: connection refused Dec 11 02:00:59 crc kubenswrapper[4824]: E1211 02:00:59.900174 4824 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.50:6443: connect: connection refused" logger="UnhandledError" Dec 11 02:00:59 crc kubenswrapper[4824]: E1211 02:00:59.972718 4824 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.50:6443: connect: connection refused" interval="1.6s" Dec 11 02:01:00 crc kubenswrapper[4824]: W1211 02:01:00.006227 4824 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.50:6443: connect: connection refused Dec 11 02:01:00 crc kubenswrapper[4824]: E1211 02:01:00.006308 4824 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.50:6443: connect: connection refused" logger="UnhandledError" Dec 11 02:01:00 crc kubenswrapper[4824]: I1211 02:01:00.229152 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 02:01:00 crc kubenswrapper[4824]: I1211 02:01:00.234258 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:00 crc kubenswrapper[4824]: I1211 02:01:00.234304 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:00 crc kubenswrapper[4824]: I1211 02:01:00.234313 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:00 crc kubenswrapper[4824]: I1211 02:01:00.234357 4824 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 11 02:01:00 crc kubenswrapper[4824]: I1211 02:01:00.489261 4824 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 11 02:01:00 crc kubenswrapper[4824]: I1211 02:01:00.567059 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-07 06:15:55.119241348 +0000 UTC Dec 11 02:01:00 crc kubenswrapper[4824]: I1211 02:01:00.567135 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 652h14m54.552110829s for next certificate rotation Dec 11 02:01:00 crc kubenswrapper[4824]: I1211 02:01:00.659808 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"f14a940a72b2404a6cbd52a69e150b03266bcd7f5c1d3f4d07b51c0c69e51786"} Dec 11 02:01:00 crc kubenswrapper[4824]: I1211 02:01:00.659866 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"20af9b10a67c8b21a7abcd23e42bc412798d159c8422e7c8ff58f5b997ca19a9"} Dec 11 02:01:00 crc kubenswrapper[4824]: I1211 02:01:00.659880 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"e9bbfc6edfe050b3fa9dbeee5f98d416bcb2a1533a38608a881b75396fd0b059"} Dec 11 02:01:00 crc kubenswrapper[4824]: I1211 02:01:00.659891 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"c26f06d2ed5e6e98b61a51f073130675f2c49acda106b8aedcbb2a6d4f7580c1"} Dec 11 02:01:00 crc kubenswrapper[4824]: I1211 02:01:00.661836 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"51533689fc74e35e0db9e2ef1b51a225ae3bbd49669afacce01841b3343cc658"} Dec 11 02:01:00 crc kubenswrapper[4824]: I1211 02:01:00.661938 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 02:01:00 crc kubenswrapper[4824]: I1211 02:01:00.662953 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:00 crc kubenswrapper[4824]: I1211 02:01:00.662985 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:00 crc kubenswrapper[4824]: I1211 02:01:00.662997 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:00 crc kubenswrapper[4824]: I1211 02:01:00.665319 4824 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="be8f3f2460ba45c67ac7c6e7eac2d6ada631e90f4c7de645a41c134539437f82" exitCode=0 Dec 11 02:01:00 crc kubenswrapper[4824]: I1211 02:01:00.665377 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"be8f3f2460ba45c67ac7c6e7eac2d6ada631e90f4c7de645a41c134539437f82"} Dec 11 02:01:00 crc kubenswrapper[4824]: I1211 02:01:00.665476 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 02:01:00 crc kubenswrapper[4824]: I1211 02:01:00.666296 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:00 crc kubenswrapper[4824]: I1211 02:01:00.666323 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:00 crc kubenswrapper[4824]: I1211 02:01:00.666335 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:00 crc kubenswrapper[4824]: I1211 02:01:00.670040 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"339d619250c1c0a12f988a9b0ebc03874db3ac41a9d3b3f8573795467578cd0f"} Dec 11 02:01:00 crc kubenswrapper[4824]: I1211 02:01:00.670065 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"335a774b8a3aff18695de4aba0b522d346aecc8ed8d96a80150301f5debce32f"} Dec 11 02:01:00 crc kubenswrapper[4824]: I1211 02:01:00.670078 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"f410073059c5898660c50f775a270bebe51836f73e2a2f02db38a2aabf15931d"} Dec 11 02:01:00 crc kubenswrapper[4824]: I1211 02:01:00.670176 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 02:01:00 crc kubenswrapper[4824]: I1211 02:01:00.671031 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:00 crc kubenswrapper[4824]: I1211 02:01:00.671059 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:00 crc kubenswrapper[4824]: I1211 02:01:00.671069 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:00 crc kubenswrapper[4824]: I1211 02:01:00.675009 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"0c89afaf56b4a9597f6543ec2e3b176c7a6dab7dc341941fc3bc8ec85cc4edaf"} Dec 11 02:01:00 crc kubenswrapper[4824]: I1211 02:01:00.675042 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"136c68020325db8e6930d0f40b0d0d7e013c14c4dd5e7d1449d01d6d458104d7"} Dec 11 02:01:00 crc kubenswrapper[4824]: I1211 02:01:00.675055 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"44b29bb7a6eda7cb890d36f86375b5f3f197ae5cf390f16e71139b0904873e39"} Dec 11 02:01:00 crc kubenswrapper[4824]: I1211 02:01:00.675152 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 02:01:00 crc kubenswrapper[4824]: I1211 02:01:00.675812 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:00 crc kubenswrapper[4824]: I1211 02:01:00.675837 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:00 crc kubenswrapper[4824]: I1211 02:01:00.675851 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:01 crc kubenswrapper[4824]: I1211 02:01:01.269467 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 11 02:01:01 crc kubenswrapper[4824]: I1211 02:01:01.682315 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 02:01:01 crc kubenswrapper[4824]: I1211 02:01:01.682368 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"8d45819f0611c91b20c83eeee1092031a3441fb69f3015241632ad419880227d"} Dec 11 02:01:01 crc kubenswrapper[4824]: I1211 02:01:01.683959 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:01 crc kubenswrapper[4824]: I1211 02:01:01.684021 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:01 crc kubenswrapper[4824]: I1211 02:01:01.684097 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:01 crc kubenswrapper[4824]: I1211 02:01:01.686360 4824 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="016b733da5b0e8f0e3830ee9746c4ab5decfcac7573f42aec257a842a393c60d" exitCode=0 Dec 11 02:01:01 crc kubenswrapper[4824]: I1211 02:01:01.686471 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 02:01:01 crc kubenswrapper[4824]: I1211 02:01:01.686459 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"016b733da5b0e8f0e3830ee9746c4ab5decfcac7573f42aec257a842a393c60d"} Dec 11 02:01:01 crc kubenswrapper[4824]: I1211 02:01:01.686618 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 02:01:01 crc kubenswrapper[4824]: I1211 02:01:01.686820 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 02:01:01 crc kubenswrapper[4824]: I1211 02:01:01.687486 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:01 crc kubenswrapper[4824]: I1211 02:01:01.687593 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:01 crc kubenswrapper[4824]: I1211 02:01:01.687620 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:01 crc kubenswrapper[4824]: I1211 02:01:01.688268 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:01 crc kubenswrapper[4824]: I1211 02:01:01.688474 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:01 crc kubenswrapper[4824]: I1211 02:01:01.688639 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:01 crc kubenswrapper[4824]: I1211 02:01:01.688272 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:01 crc kubenswrapper[4824]: I1211 02:01:01.688878 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:01 crc kubenswrapper[4824]: I1211 02:01:01.688906 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:02 crc kubenswrapper[4824]: I1211 02:01:02.695762 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"09587408482a9771161017d1246b051cb9ad2961bd949414210b3c0ad58ae3fc"} Dec 11 02:01:02 crc kubenswrapper[4824]: I1211 02:01:02.695847 4824 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 11 02:01:02 crc kubenswrapper[4824]: I1211 02:01:02.695877 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"bae1b2b71f76385e3144a18a4335f125adaf2b40a59e9f8c353ab2fd5efd5549"} Dec 11 02:01:02 crc kubenswrapper[4824]: I1211 02:01:02.695904 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"0426825294b0be03de99d39e8a1584f7717017b99d70a12af858ff8ea0699302"} Dec 11 02:01:02 crc kubenswrapper[4824]: I1211 02:01:02.695908 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 02:01:02 crc kubenswrapper[4824]: I1211 02:01:02.697200 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:02 crc kubenswrapper[4824]: I1211 02:01:02.697270 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:02 crc kubenswrapper[4824]: I1211 02:01:02.697296 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:02 crc kubenswrapper[4824]: I1211 02:01:02.864492 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 02:01:02 crc kubenswrapper[4824]: I1211 02:01:02.864689 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 02:01:02 crc kubenswrapper[4824]: I1211 02:01:02.866358 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:02 crc kubenswrapper[4824]: I1211 02:01:02.866400 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:02 crc kubenswrapper[4824]: I1211 02:01:02.866420 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:03 crc kubenswrapper[4824]: I1211 02:01:03.704325 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"af95eed3f8b150973c8f151d41b7a8ce2c41d90ede17bf0f027b721f4f3858d2"} Dec 11 02:01:03 crc kubenswrapper[4824]: I1211 02:01:03.704396 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"e42434f5fe6b847211ad32b3fee7e6a6a6adbe9c5caa30235ac07a25507f7871"} Dec 11 02:01:03 crc kubenswrapper[4824]: I1211 02:01:03.704492 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 02:01:03 crc kubenswrapper[4824]: I1211 02:01:03.705790 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:03 crc kubenswrapper[4824]: I1211 02:01:03.705842 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:03 crc kubenswrapper[4824]: I1211 02:01:03.705860 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:04 crc kubenswrapper[4824]: I1211 02:01:04.104094 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 02:01:04 crc kubenswrapper[4824]: I1211 02:01:04.104282 4824 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 11 02:01:04 crc kubenswrapper[4824]: I1211 02:01:04.104324 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 02:01:04 crc kubenswrapper[4824]: I1211 02:01:04.105472 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:04 crc kubenswrapper[4824]: I1211 02:01:04.105525 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:04 crc kubenswrapper[4824]: I1211 02:01:04.105542 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:04 crc kubenswrapper[4824]: I1211 02:01:04.604948 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 02:01:04 crc kubenswrapper[4824]: I1211 02:01:04.707602 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 02:01:04 crc kubenswrapper[4824]: I1211 02:01:04.707640 4824 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 11 02:01:04 crc kubenswrapper[4824]: I1211 02:01:04.707704 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 02:01:04 crc kubenswrapper[4824]: I1211 02:01:04.708961 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:04 crc kubenswrapper[4824]: I1211 02:01:04.709015 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:04 crc kubenswrapper[4824]: I1211 02:01:04.709035 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:04 crc kubenswrapper[4824]: I1211 02:01:04.709196 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:04 crc kubenswrapper[4824]: I1211 02:01:04.709242 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:04 crc kubenswrapper[4824]: I1211 02:01:04.709260 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:05 crc kubenswrapper[4824]: I1211 02:01:05.393419 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Dec 11 02:01:05 crc kubenswrapper[4824]: I1211 02:01:05.710576 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 02:01:05 crc kubenswrapper[4824]: I1211 02:01:05.711734 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:05 crc kubenswrapper[4824]: I1211 02:01:05.711782 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:05 crc kubenswrapper[4824]: I1211 02:01:05.711801 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:07 crc kubenswrapper[4824]: I1211 02:01:07.070687 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 02:01:07 crc kubenswrapper[4824]: I1211 02:01:07.071329 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 02:01:07 crc kubenswrapper[4824]: I1211 02:01:07.072869 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:07 crc kubenswrapper[4824]: I1211 02:01:07.073157 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:07 crc kubenswrapper[4824]: I1211 02:01:07.073319 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:07 crc kubenswrapper[4824]: I1211 02:01:07.230025 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 02:01:07 crc kubenswrapper[4824]: I1211 02:01:07.237899 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 02:01:07 crc kubenswrapper[4824]: I1211 02:01:07.716224 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 02:01:07 crc kubenswrapper[4824]: I1211 02:01:07.717471 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:07 crc kubenswrapper[4824]: I1211 02:01:07.717533 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:07 crc kubenswrapper[4824]: I1211 02:01:07.717551 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:07 crc kubenswrapper[4824]: I1211 02:01:07.935255 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 02:01:07 crc kubenswrapper[4824]: I1211 02:01:07.935552 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 02:01:07 crc kubenswrapper[4824]: I1211 02:01:07.937165 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:07 crc kubenswrapper[4824]: I1211 02:01:07.937224 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:07 crc kubenswrapper[4824]: I1211 02:01:07.937245 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:08 crc kubenswrapper[4824]: I1211 02:01:08.718678 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 02:01:08 crc kubenswrapper[4824]: I1211 02:01:08.720237 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:08 crc kubenswrapper[4824]: I1211 02:01:08.720309 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:08 crc kubenswrapper[4824]: I1211 02:01:08.720329 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:08 crc kubenswrapper[4824]: E1211 02:01:08.737141 4824 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 11 02:01:10 crc kubenswrapper[4824]: E1211 02:01:10.235022 4824 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": net/http: TLS handshake timeout" node="crc" Dec 11 02:01:10 crc kubenswrapper[4824]: E1211 02:01:10.491520 4824 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 11 02:01:10 crc kubenswrapper[4824]: I1211 02:01:10.530531 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 02:01:10 crc kubenswrapper[4824]: I1211 02:01:10.530789 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 02:01:10 crc kubenswrapper[4824]: I1211 02:01:10.532418 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:10 crc kubenswrapper[4824]: I1211 02:01:10.532497 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:10 crc kubenswrapper[4824]: I1211 02:01:10.532518 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:10 crc kubenswrapper[4824]: I1211 02:01:10.537758 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 02:01:10 crc kubenswrapper[4824]: I1211 02:01:10.561100 4824 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Dec 11 02:01:10 crc kubenswrapper[4824]: I1211 02:01:10.724225 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 02:01:10 crc kubenswrapper[4824]: I1211 02:01:10.725540 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:10 crc kubenswrapper[4824]: I1211 02:01:10.725610 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:10 crc kubenswrapper[4824]: I1211 02:01:10.725631 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:11 crc kubenswrapper[4824]: E1211 02:01:11.574429 4824 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="3.2s" Dec 11 02:01:11 crc kubenswrapper[4824]: I1211 02:01:11.835843 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 02:01:11 crc kubenswrapper[4824]: I1211 02:01:11.837316 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:11 crc kubenswrapper[4824]: I1211 02:01:11.837362 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:11 crc kubenswrapper[4824]: I1211 02:01:11.837380 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:11 crc kubenswrapper[4824]: I1211 02:01:11.837413 4824 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 11 02:01:12 crc kubenswrapper[4824]: I1211 02:01:12.030597 4824 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 11 02:01:12 crc kubenswrapper[4824]: I1211 02:01:12.030765 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 11 02:01:12 crc kubenswrapper[4824]: I1211 02:01:12.037868 4824 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 11 02:01:12 crc kubenswrapper[4824]: I1211 02:01:12.037934 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 11 02:01:13 crc kubenswrapper[4824]: I1211 02:01:13.243245 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Dec 11 02:01:13 crc kubenswrapper[4824]: I1211 02:01:13.243603 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 02:01:13 crc kubenswrapper[4824]: I1211 02:01:13.245364 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:13 crc kubenswrapper[4824]: I1211 02:01:13.245432 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:13 crc kubenswrapper[4824]: I1211 02:01:13.245453 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:13 crc kubenswrapper[4824]: I1211 02:01:13.274456 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Dec 11 02:01:13 crc kubenswrapper[4824]: I1211 02:01:13.531237 4824 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 11 02:01:13 crc kubenswrapper[4824]: I1211 02:01:13.531324 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 11 02:01:13 crc kubenswrapper[4824]: I1211 02:01:13.733684 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 02:01:13 crc kubenswrapper[4824]: I1211 02:01:13.734833 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:13 crc kubenswrapper[4824]: I1211 02:01:13.734872 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:13 crc kubenswrapper[4824]: I1211 02:01:13.734883 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:13 crc kubenswrapper[4824]: I1211 02:01:13.759639 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Dec 11 02:01:14 crc kubenswrapper[4824]: I1211 02:01:14.611914 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 02:01:14 crc kubenswrapper[4824]: I1211 02:01:14.612198 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 02:01:14 crc kubenswrapper[4824]: I1211 02:01:14.613528 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:14 crc kubenswrapper[4824]: I1211 02:01:14.613636 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:14 crc kubenswrapper[4824]: I1211 02:01:14.613657 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:14 crc kubenswrapper[4824]: I1211 02:01:14.619029 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 02:01:14 crc kubenswrapper[4824]: I1211 02:01:14.710669 4824 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 11 02:01:14 crc kubenswrapper[4824]: I1211 02:01:14.726818 4824 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Dec 11 02:01:14 crc kubenswrapper[4824]: I1211 02:01:14.735628 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 02:01:14 crc kubenswrapper[4824]: I1211 02:01:14.735652 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 02:01:14 crc kubenswrapper[4824]: I1211 02:01:14.737336 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:14 crc kubenswrapper[4824]: I1211 02:01:14.737394 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:14 crc kubenswrapper[4824]: I1211 02:01:14.737416 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:14 crc kubenswrapper[4824]: I1211 02:01:14.737434 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:14 crc kubenswrapper[4824]: I1211 02:01:14.737456 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:14 crc kubenswrapper[4824]: I1211 02:01:14.737438 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:14 crc kubenswrapper[4824]: I1211 02:01:14.974062 4824 csr.go:261] certificate signing request csr-jrjrq is approved, waiting to be issued Dec 11 02:01:15 crc kubenswrapper[4824]: I1211 02:01:15.013454 4824 csr.go:257] certificate signing request csr-jrjrq is issued Dec 11 02:01:16 crc kubenswrapper[4824]: I1211 02:01:16.014386 4824 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-12-11 01:56:14 +0000 UTC, rotation deadline is 2026-09-13 03:56:34.443127204 +0000 UTC Dec 11 02:01:16 crc kubenswrapper[4824]: I1211 02:01:16.014444 4824 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 6625h55m18.428688106s for next certificate rotation Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.018276 4824 trace.go:236] Trace[1004283099]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (11-Dec-2025 02:01:02.268) (total time: 14749ms): Dec 11 02:01:17 crc kubenswrapper[4824]: Trace[1004283099]: ---"Objects listed" error: 14749ms (02:01:17.018) Dec 11 02:01:17 crc kubenswrapper[4824]: Trace[1004283099]: [14.749424941s] [14.749424941s] END Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.018335 4824 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.018683 4824 trace.go:236] Trace[398281671]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (11-Dec-2025 02:01:02.515) (total time: 14503ms): Dec 11 02:01:17 crc kubenswrapper[4824]: Trace[398281671]: ---"Objects listed" error: 14503ms (02:01:17.018) Dec 11 02:01:17 crc kubenswrapper[4824]: Trace[398281671]: [14.503574399s] [14.503574399s] END Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.018902 4824 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.020449 4824 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Dec 11 02:01:17 crc kubenswrapper[4824]: E1211 02:01:17.021794 4824 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.021840 4824 trace.go:236] Trace[507549281]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (11-Dec-2025 02:01:02.870) (total time: 14150ms): Dec 11 02:01:17 crc kubenswrapper[4824]: Trace[507549281]: ---"Objects listed" error: 14150ms (02:01:17.021) Dec 11 02:01:17 crc kubenswrapper[4824]: Trace[507549281]: [14.150995886s] [14.150995886s] END Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.021880 4824 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.022429 4824 trace.go:236] Trace[1546829759]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (11-Dec-2025 02:01:02.042) (total time: 14979ms): Dec 11 02:01:17 crc kubenswrapper[4824]: Trace[1546829759]: ---"Objects listed" error: 14979ms (02:01:17.022) Dec 11 02:01:17 crc kubenswrapper[4824]: Trace[1546829759]: [14.979663005s] [14.979663005s] END Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.022449 4824 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.069706 4824 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:38720->192.168.126.11:17697: read: connection reset by peer" start-of-body= Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.069819 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:38720->192.168.126.11:17697: read: connection reset by peer" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.070312 4824 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.070379 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.381248 4824 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.381293 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.568162 4824 apiserver.go:52] "Watching apiserver" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.578713 4824 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.578997 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-kptws","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf"] Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.579389 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.579473 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.579492 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.579577 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 02:01:17 crc kubenswrapper[4824]: E1211 02:01:17.579576 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 02:01:17 crc kubenswrapper[4824]: E1211 02:01:17.579643 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.579738 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.579890 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-kptws" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.580261 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 02:01:17 crc kubenswrapper[4824]: E1211 02:01:17.586233 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.586966 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.587246 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.587583 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.587638 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.587798 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.588097 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.588273 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.588472 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.588641 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.588708 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.589367 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.589765 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.608755 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.623549 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.641612 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.652076 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.657733 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kptws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68568dba-83a5-46c1-8d01-337359679d91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd7nf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kptws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.665948 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.667442 4824 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.673767 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.681905 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.714518 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-kh8g9"] Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.715162 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-6bkc2"] Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.715345 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-kh8g9" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.715390 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-6bkc2" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.716202 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-f6pgc"] Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.716928 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.718052 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-gx6xt"] Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.718461 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.721130 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.721306 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.721393 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.721493 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.721572 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.721647 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.721308 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.721824 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.721137 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.721868 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.721502 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.722002 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.722045 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.723697 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.723993 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.724187 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.724311 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.724415 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.724495 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.724513 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.724589 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.724621 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.724651 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.724678 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.724703 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.724725 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.724749 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.724747 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.724770 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.724816 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.724840 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.724863 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.724887 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.724928 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.724951 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.724978 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.725000 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.725038 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.725064 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.725089 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.725135 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.725166 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.725198 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.725222 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.725457 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.725718 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.725969 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.726276 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.726683 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.726774 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.727054 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.727134 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.727155 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.727343 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.727524 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.727574 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.727412 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.727768 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.728061 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.728447 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.728747 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.728800 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.728828 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.728854 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.728880 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.728903 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.728927 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.728950 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.728974 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.728994 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.729014 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.729036 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.729346 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.729388 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.729414 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.729436 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.729487 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.729511 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.729532 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.729556 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.729603 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.729628 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.729652 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.729675 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.729699 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.729897 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.729921 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.729945 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.729967 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.729990 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.730014 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.730038 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.730063 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.730086 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.730126 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.730153 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.730176 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.730208 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.730231 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.730254 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.730278 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.730299 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.730323 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.730344 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.730368 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.730389 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.730409 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.730431 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.730454 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.730476 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.730500 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.730528 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.730550 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.730571 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.730593 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.730615 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.730637 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.730658 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.730680 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.730700 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.730746 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.730769 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.730792 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.730815 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.730844 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.730865 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.730887 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.730909 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.730945 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.730972 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.730993 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.731014 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.731037 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.731061 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.731085 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.731107 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.731149 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.731172 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.731193 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.731214 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.731235 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.731259 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.731282 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.731304 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.731327 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.731350 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.731374 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.731395 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.731417 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.731438 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.731460 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.731486 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.731512 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.731536 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.731561 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.731583 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.731608 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.731631 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.731654 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.731677 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.731700 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.731725 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.731747 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.731771 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.731794 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.731820 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.731843 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.731868 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.731892 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.731915 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.731940 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.731963 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.731986 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.732028 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.732058 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.732083 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.732124 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.732149 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.732171 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.732194 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.732215 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.732236 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.732260 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.732284 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.732306 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.732328 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.732428 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.732454 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.732476 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.732510 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.732534 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.732555 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.732579 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.732601 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.732623 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.732644 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.732681 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.732703 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.732724 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.732745 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.732765 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.732786 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.732808 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.732827 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.732848 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.732871 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.732896 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.732919 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.732944 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.732968 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.732993 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.733017 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.733040 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.733064 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.733085 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.733105 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.729506 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.729889 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.729984 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.730267 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.730453 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.730703 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.730727 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.730885 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.731065 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.731540 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.731661 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.731760 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.732090 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.733337 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.734073 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.734237 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.734567 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.734683 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.734844 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.735233 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.735257 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.735401 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.735580 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.735638 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.735685 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.735907 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.738850 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.738935 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.738967 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.738999 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.739028 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.739056 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.739085 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.739088 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.739136 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.739159 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.739180 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.739199 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.739262 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.739290 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.739311 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.739333 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.739353 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.739373 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.739395 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.739429 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.739462 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.739485 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/68568dba-83a5-46c1-8d01-337359679d91-hosts-file\") pod \"node-resolver-kptws\" (UID: \"68568dba-83a5-46c1-8d01-337359679d91\") " pod="openshift-dns/node-resolver-kptws" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.739508 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.739529 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.739551 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gd7nf\" (UniqueName: \"kubernetes.io/projected/68568dba-83a5-46c1-8d01-337359679d91-kube-api-access-gd7nf\") pod \"node-resolver-kptws\" (UID: \"68568dba-83a5-46c1-8d01-337359679d91\") " pod="openshift-dns/node-resolver-kptws" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.739580 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.739611 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.739638 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.739713 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.739726 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.739741 4824 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.739752 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.739762 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.739772 4824 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.739786 4824 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.739798 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.739809 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.739822 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.739832 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.739844 4824 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.739854 4824 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.739866 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.739876 4824 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.739886 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.739895 4824 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.739909 4824 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.739919 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.739929 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.739941 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.739952 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.739962 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.739973 4824 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.740038 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.740050 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.740063 4824 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.740078 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.740093 4824 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.739371 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.740268 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.752671 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.740789 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.741520 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.741737 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.741917 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.741918 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.742148 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.742348 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.742448 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.742545 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.742610 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.742822 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.742917 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: E1211 02:01:17.743072 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 02:01:18.243010017 +0000 UTC m=+19.932047396 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.743959 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.744527 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.744563 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.744899 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.745070 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.745442 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.745447 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.745703 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.745779 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.745789 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.746126 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.746157 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.746373 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.746414 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.746445 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.746743 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.746830 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.747134 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.747146 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.747219 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.747413 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.746640 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.747543 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.747698 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.747717 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.747778 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.748249 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.748307 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.748409 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.748685 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.748877 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.748970 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.748993 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.749163 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.747916 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.749419 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.749608 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.750188 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.750243 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.750245 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.750335 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.750429 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.750495 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.750662 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.750705 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.750736 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.750742 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.750962 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.751332 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.751816 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.751917 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.751964 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: E1211 02:01:17.751988 4824 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.752055 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.752339 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.752423 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: E1211 02:01:17.752771 4824 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.753010 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.752626 4824 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.753592 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.753673 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.753796 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.754094 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.754296 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.754387 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.754492 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.755464 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.753280 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.756478 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.756508 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.756538 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.756757 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.756773 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.756876 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.757086 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.757092 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.753028 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.757490 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.757540 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.757699 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.757708 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: E1211 02:01:17.757767 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-11 02:01:18.257747307 +0000 UTC m=+19.946784686 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 11 02:01:17 crc kubenswrapper[4824]: E1211 02:01:17.758177 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-11 02:01:18.258145308 +0000 UTC m=+19.947182727 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.759332 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.760023 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.761891 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.762293 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.762847 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.764616 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.764835 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.764868 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.764899 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.765161 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.765408 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.765782 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.765849 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.765931 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.766604 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.766768 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.767259 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.767522 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.767784 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.768088 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.768475 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.768687 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.768722 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.768838 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.769217 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.769238 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.769506 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.769562 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.769821 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.769875 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.770144 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.770157 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.770228 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.770234 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.770293 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.770466 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.770718 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.770780 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.770801 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.770798 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.771148 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.771190 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.771568 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.771841 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.771923 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.771960 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.771984 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.773157 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.773436 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: E1211 02:01:17.772161 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.773543 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: E1211 02:01:17.773560 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.772181 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.772516 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.772869 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.772861 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.772952 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.773641 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.773857 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.773971 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.774180 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.774252 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.774287 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.776252 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: E1211 02:01:17.777422 4824 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 02:01:17 crc kubenswrapper[4824]: E1211 02:01:17.777540 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-11 02:01:18.277496504 +0000 UTC m=+19.966533973 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 02:01:17 crc kubenswrapper[4824]: E1211 02:01:17.777804 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 11 02:01:17 crc kubenswrapper[4824]: E1211 02:01:17.777825 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 11 02:01:17 crc kubenswrapper[4824]: E1211 02:01:17.777838 4824 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 02:01:17 crc kubenswrapper[4824]: E1211 02:01:17.777875 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-11 02:01:18.277863104 +0000 UTC m=+19.966900483 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.778263 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.780345 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.781270 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.781945 4824 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="8d45819f0611c91b20c83eeee1092031a3441fb69f3015241632ad419880227d" exitCode=255 Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.782004 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"8d45819f0611c91b20c83eeee1092031a3441fb69f3015241632ad419880227d"} Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.782523 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.787232 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.787520 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.793013 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.794891 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.796664 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.798099 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.798494 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kh8g9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"50c69d41-a65b-463e-a829-ea349bd2fe46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kh8g9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.808486 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.822056 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.832453 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.840522 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kptws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68568dba-83a5-46c1-8d01-337359679d91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd7nf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kptws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.840684 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.840787 4824 scope.go:117] "RemoveContainer" containerID="8d45819f0611c91b20c83eeee1092031a3441fb69f3015241632ad419880227d" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.840805 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/8022f696-d0e7-437b-848a-3cd25bd1f364-host-run-multus-certs\") pod \"multus-6bkc2\" (UID: \"8022f696-d0e7-437b-848a-3cd25bd1f364\") " pod="openshift-multus/multus-6bkc2" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.840866 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/68568dba-83a5-46c1-8d01-337359679d91-hosts-file\") pod \"node-resolver-kptws\" (UID: \"68568dba-83a5-46c1-8d01-337359679d91\") " pod="openshift-dns/node-resolver-kptws" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.840894 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.840920 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/44b156e4-64a4-4d45-aa5e-9b10a862faed-mcd-auth-proxy-config\") pod \"machine-config-daemon-gx6xt\" (UID: \"44b156e4-64a4-4d45-aa5e-9b10a862faed\") " pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.840956 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/50c69d41-a65b-463e-a829-ea349bd2fe46-tuning-conf-dir\") pod \"multus-additional-cni-plugins-kh8g9\" (UID: \"50c69d41-a65b-463e-a829-ea349bd2fe46\") " pod="openshift-multus/multus-additional-cni-plugins-kh8g9" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.840977 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-run-systemd\") pod \"ovnkube-node-f6pgc\" (UID: \"5b193bd2-b536-4056-92f1-94c9836ab2eb\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.840998 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w62dt\" (UniqueName: \"kubernetes.io/projected/44b156e4-64a4-4d45-aa5e-9b10a862faed-kube-api-access-w62dt\") pod \"machine-config-daemon-gx6xt\" (UID: \"44b156e4-64a4-4d45-aa5e-9b10a862faed\") " pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.841021 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/8022f696-d0e7-437b-848a-3cd25bd1f364-os-release\") pod \"multus-6bkc2\" (UID: \"8022f696-d0e7-437b-848a-3cd25bd1f364\") " pod="openshift-multus/multus-6bkc2" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.841040 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8022f696-d0e7-437b-848a-3cd25bd1f364-host-var-lib-cni-bin\") pod \"multus-6bkc2\" (UID: \"8022f696-d0e7-437b-848a-3cd25bd1f364\") " pod="openshift-multus/multus-6bkc2" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.841060 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/44b156e4-64a4-4d45-aa5e-9b10a862faed-rootfs\") pod \"machine-config-daemon-gx6xt\" (UID: \"44b156e4-64a4-4d45-aa5e-9b10a862faed\") " pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.841087 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.841124 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/50c69d41-a65b-463e-a829-ea349bd2fe46-os-release\") pod \"multus-additional-cni-plugins-kh8g9\" (UID: \"50c69d41-a65b-463e-a829-ea349bd2fe46\") " pod="openshift-multus/multus-additional-cni-plugins-kh8g9" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.841150 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/50c69d41-a65b-463e-a829-ea349bd2fe46-cni-binary-copy\") pod \"multus-additional-cni-plugins-kh8g9\" (UID: \"50c69d41-a65b-463e-a829-ea349bd2fe46\") " pod="openshift-multus/multus-additional-cni-plugins-kh8g9" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.841171 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-systemd-units\") pod \"ovnkube-node-f6pgc\" (UID: \"5b193bd2-b536-4056-92f1-94c9836ab2eb\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.841193 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-host-run-netns\") pod \"ovnkube-node-f6pgc\" (UID: \"5b193bd2-b536-4056-92f1-94c9836ab2eb\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.841213 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-var-lib-openvswitch\") pod \"ovnkube-node-f6pgc\" (UID: \"5b193bd2-b536-4056-92f1-94c9836ab2eb\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.841258 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/8022f696-d0e7-437b-848a-3cd25bd1f364-cni-binary-copy\") pod \"multus-6bkc2\" (UID: \"8022f696-d0e7-437b-848a-3cd25bd1f364\") " pod="openshift-multus/multus-6bkc2" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.841280 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5b193bd2-b536-4056-92f1-94c9836ab2eb-env-overrides\") pod \"ovnkube-node-f6pgc\" (UID: \"5b193bd2-b536-4056-92f1-94c9836ab2eb\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.841344 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5b193bd2-b536-4056-92f1-94c9836ab2eb-ovn-node-metrics-cert\") pod \"ovnkube-node-f6pgc\" (UID: \"5b193bd2-b536-4056-92f1-94c9836ab2eb\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.841376 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/8022f696-d0e7-437b-848a-3cd25bd1f364-host-var-lib-cni-multus\") pod \"multus-6bkc2\" (UID: \"8022f696-d0e7-437b-848a-3cd25bd1f364\") " pod="openshift-multus/multus-6bkc2" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.841415 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/68568dba-83a5-46c1-8d01-337359679d91-hosts-file\") pod \"node-resolver-kptws\" (UID: \"68568dba-83a5-46c1-8d01-337359679d91\") " pod="openshift-dns/node-resolver-kptws" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.841425 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/50c69d41-a65b-463e-a829-ea349bd2fe46-cnibin\") pod \"multus-additional-cni-plugins-kh8g9\" (UID: \"50c69d41-a65b-463e-a829-ea349bd2fe46\") " pod="openshift-multus/multus-additional-cni-plugins-kh8g9" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.841469 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-host-cni-bin\") pod \"ovnkube-node-f6pgc\" (UID: \"5b193bd2-b536-4056-92f1-94c9836ab2eb\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.841520 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/8022f696-d0e7-437b-848a-3cd25bd1f364-cnibin\") pod \"multus-6bkc2\" (UID: \"8022f696-d0e7-437b-848a-3cd25bd1f364\") " pod="openshift-multus/multus-6bkc2" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.841535 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.841544 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/50c69d41-a65b-463e-a829-ea349bd2fe46-system-cni-dir\") pod \"multus-additional-cni-plugins-kh8g9\" (UID: \"50c69d41-a65b-463e-a829-ea349bd2fe46\") " pod="openshift-multus/multus-additional-cni-plugins-kh8g9" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.841602 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-host-slash\") pod \"ovnkube-node-f6pgc\" (UID: \"5b193bd2-b536-4056-92f1-94c9836ab2eb\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.841611 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.841624 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-node-log\") pod \"ovnkube-node-f6pgc\" (UID: \"5b193bd2-b536-4056-92f1-94c9836ab2eb\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.841728 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-host-run-ovn-kubernetes\") pod \"ovnkube-node-f6pgc\" (UID: \"5b193bd2-b536-4056-92f1-94c9836ab2eb\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.841810 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/8022f696-d0e7-437b-848a-3cd25bd1f364-hostroot\") pod \"multus-6bkc2\" (UID: \"8022f696-d0e7-437b-848a-3cd25bd1f364\") " pod="openshift-multus/multus-6bkc2" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.841834 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8022f696-d0e7-437b-848a-3cd25bd1f364-etc-kubernetes\") pod \"multus-6bkc2\" (UID: \"8022f696-d0e7-437b-848a-3cd25bd1f364\") " pod="openshift-multus/multus-6bkc2" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.841853 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5b193bd2-b536-4056-92f1-94c9836ab2eb-ovnkube-config\") pod \"ovnkube-node-f6pgc\" (UID: \"5b193bd2-b536-4056-92f1-94c9836ab2eb\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.841880 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8022f696-d0e7-437b-848a-3cd25bd1f364-host-run-netns\") pod \"multus-6bkc2\" (UID: \"8022f696-d0e7-437b-848a-3cd25bd1f364\") " pod="openshift-multus/multus-6bkc2" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.841973 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-host-kubelet\") pod \"ovnkube-node-f6pgc\" (UID: \"5b193bd2-b536-4056-92f1-94c9836ab2eb\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.842147 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-host-cni-netd\") pod \"ovnkube-node-f6pgc\" (UID: \"5b193bd2-b536-4056-92f1-94c9836ab2eb\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.842279 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gd7nf\" (UniqueName: \"kubernetes.io/projected/68568dba-83a5-46c1-8d01-337359679d91-kube-api-access-gd7nf\") pod \"node-resolver-kptws\" (UID: \"68568dba-83a5-46c1-8d01-337359679d91\") " pod="openshift-dns/node-resolver-kptws" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.842381 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/8022f696-d0e7-437b-848a-3cd25bd1f364-multus-socket-dir-parent\") pod \"multus-6bkc2\" (UID: \"8022f696-d0e7-437b-848a-3cd25bd1f364\") " pod="openshift-multus/multus-6bkc2" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.842439 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/8022f696-d0e7-437b-848a-3cd25bd1f364-host-run-k8s-cni-cncf-io\") pod \"multus-6bkc2\" (UID: \"8022f696-d0e7-437b-848a-3cd25bd1f364\") " pod="openshift-multus/multus-6bkc2" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.842467 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-run-ovn\") pod \"ovnkube-node-f6pgc\" (UID: \"5b193bd2-b536-4056-92f1-94c9836ab2eb\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.842489 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qpqjk\" (UniqueName: \"kubernetes.io/projected/5b193bd2-b536-4056-92f1-94c9836ab2eb-kube-api-access-qpqjk\") pod \"ovnkube-node-f6pgc\" (UID: \"5b193bd2-b536-4056-92f1-94c9836ab2eb\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.842513 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8022f696-d0e7-437b-848a-3cd25bd1f364-multus-cni-dir\") pod \"multus-6bkc2\" (UID: \"8022f696-d0e7-437b-848a-3cd25bd1f364\") " pod="openshift-multus/multus-6bkc2" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.842534 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hh5kn\" (UniqueName: \"kubernetes.io/projected/8022f696-d0e7-437b-848a-3cd25bd1f364-kube-api-access-hh5kn\") pod \"multus-6bkc2\" (UID: \"8022f696-d0e7-437b-848a-3cd25bd1f364\") " pod="openshift-multus/multus-6bkc2" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.842879 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/50c69d41-a65b-463e-a829-ea349bd2fe46-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-kh8g9\" (UID: \"50c69d41-a65b-463e-a829-ea349bd2fe46\") " pod="openshift-multus/multus-additional-cni-plugins-kh8g9" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.844023 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6qdtz\" (UniqueName: \"kubernetes.io/projected/50c69d41-a65b-463e-a829-ea349bd2fe46-kube-api-access-6qdtz\") pod \"multus-additional-cni-plugins-kh8g9\" (UID: \"50c69d41-a65b-463e-a829-ea349bd2fe46\") " pod="openshift-multus/multus-additional-cni-plugins-kh8g9" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.844122 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8022f696-d0e7-437b-848a-3cd25bd1f364-system-cni-dir\") pod \"multus-6bkc2\" (UID: \"8022f696-d0e7-437b-848a-3cd25bd1f364\") " pod="openshift-multus/multus-6bkc2" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.844150 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-etc-openvswitch\") pod \"ovnkube-node-f6pgc\" (UID: \"5b193bd2-b536-4056-92f1-94c9836ab2eb\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.844621 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-f6pgc\" (UID: \"5b193bd2-b536-4056-92f1-94c9836ab2eb\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.844690 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/8022f696-d0e7-437b-848a-3cd25bd1f364-host-var-lib-kubelet\") pod \"multus-6bkc2\" (UID: \"8022f696-d0e7-437b-848a-3cd25bd1f364\") " pod="openshift-multus/multus-6bkc2" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.844759 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/8022f696-d0e7-437b-848a-3cd25bd1f364-multus-conf-dir\") pod \"multus-6bkc2\" (UID: \"8022f696-d0e7-437b-848a-3cd25bd1f364\") " pod="openshift-multus/multus-6bkc2" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.844793 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-log-socket\") pod \"ovnkube-node-f6pgc\" (UID: \"5b193bd2-b536-4056-92f1-94c9836ab2eb\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.844862 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/5b193bd2-b536-4056-92f1-94c9836ab2eb-ovnkube-script-lib\") pod \"ovnkube-node-f6pgc\" (UID: \"5b193bd2-b536-4056-92f1-94c9836ab2eb\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.844934 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-run-openvswitch\") pod \"ovnkube-node-f6pgc\" (UID: \"5b193bd2-b536-4056-92f1-94c9836ab2eb\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.845001 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/44b156e4-64a4-4d45-aa5e-9b10a862faed-proxy-tls\") pod \"machine-config-daemon-gx6xt\" (UID: \"44b156e4-64a4-4d45-aa5e-9b10a862faed\") " pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.845074 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/8022f696-d0e7-437b-848a-3cd25bd1f364-multus-daemon-config\") pod \"multus-6bkc2\" (UID: \"8022f696-d0e7-437b-848a-3cd25bd1f364\") " pod="openshift-multus/multus-6bkc2" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.845326 4824 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.845347 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.845396 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.845412 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.845425 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.845476 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.845493 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.845506 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.845518 4824 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.845554 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.845569 4824 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.845580 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.845593 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.845605 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.845642 4824 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.845654 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.845665 4824 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.845677 4824 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.845712 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.845726 4824 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.845738 4824 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.845750 4824 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.845762 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.845798 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.845813 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.845826 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.845838 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.845850 4824 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.845888 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.845901 4824 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.845914 4824 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.845925 4824 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.845962 4824 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.845978 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.845994 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.846008 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.846063 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.846077 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.846090 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.846101 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.846151 4824 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.846163 4824 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.846175 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.846186 4824 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.846222 4824 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.846236 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.846248 4824 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.846261 4824 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.846316 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.846332 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.846343 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.846354 4824 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.846368 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.846407 4824 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.846418 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.846430 4824 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.846441 4824 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.846451 4824 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.846489 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.846501 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.846515 4824 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.846526 4824 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.846562 4824 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.846575 4824 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.846587 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.846599 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.846612 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.846648 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.846662 4824 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.846673 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.846684 4824 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.846695 4824 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.846736 4824 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.846747 4824 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.846760 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.846771 4824 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.846807 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.846820 4824 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.846832 4824 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.846845 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.846856 4824 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.846894 4824 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.846907 4824 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.846918 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.846930 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.846966 4824 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.846981 4824 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.846992 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.847002 4824 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.847012 4824 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.847045 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.847060 4824 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.847071 4824 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.847093 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.847143 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.847159 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.847174 4824 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.847238 4824 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.847251 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.847262 4824 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.847276 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.847287 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.847325 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.847337 4824 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.847348 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.847359 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.847370 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.847407 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.847417 4824 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.847428 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.847439 4824 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.847449 4824 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.847486 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.847507 4824 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.847518 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.847528 4824 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.847562 4824 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.847577 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.847588 4824 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.847599 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.847611 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.847621 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.847657 4824 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.847668 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.847679 4824 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.847729 4824 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.847742 4824 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.847752 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.847763 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.847775 4824 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.847786 4824 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.847823 4824 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.847836 4824 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.847847 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.847859 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.847892 4824 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.847907 4824 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.847918 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.847929 4824 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.847942 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.847979 4824 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.847994 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.848007 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.848018 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.848030 4824 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.848066 4824 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.848079 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.848090 4824 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.848102 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.848148 4824 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.848163 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.848175 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.848186 4824 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.848222 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.848235 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.848245 4824 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.848257 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.848268 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.848305 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.848317 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.848329 4824 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.848339 4824 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.848350 4824 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.848385 4824 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.848400 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.856147 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.859216 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gd7nf\" (UniqueName: \"kubernetes.io/projected/68568dba-83a5-46c1-8d01-337359679d91-kube-api-access-gd7nf\") pod \"node-resolver-kptws\" (UID: \"68568dba-83a5-46c1-8d01-337359679d91\") " pod="openshift-dns/node-resolver-kptws" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.866447 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.877621 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kh8g9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"50c69d41-a65b-463e-a829-ea349bd2fe46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kh8g9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.888716 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6bkc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8022f696-d0e7-437b-848a-3cd25bd1f364\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hh5kn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6bkc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.902707 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.906989 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b193bd2-b536-4056-92f1-94c9836ab2eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f6pgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.910664 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.926013 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.932502 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-kptws" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.933799 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 02:01:17 crc kubenswrapper[4824]: W1211 02:01:17.939346 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-e39c981e597138f82a759254cf02b48e8f10947074ea1521b2775bd56a5f4bf2 WatchSource:0}: Error finding container e39c981e597138f82a759254cf02b48e8f10947074ea1521b2775bd56a5f4bf2: Status 404 returned error can't find the container with id e39c981e597138f82a759254cf02b48e8f10947074ea1521b2775bd56a5f4bf2 Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.949591 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/50c69d41-a65b-463e-a829-ea349bd2fe46-cni-binary-copy\") pod \"multus-additional-cni-plugins-kh8g9\" (UID: \"50c69d41-a65b-463e-a829-ea349bd2fe46\") " pod="openshift-multus/multus-additional-cni-plugins-kh8g9" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.949627 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-systemd-units\") pod \"ovnkube-node-f6pgc\" (UID: \"5b193bd2-b536-4056-92f1-94c9836ab2eb\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.949644 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-host-run-netns\") pod \"ovnkube-node-f6pgc\" (UID: \"5b193bd2-b536-4056-92f1-94c9836ab2eb\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.949661 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-var-lib-openvswitch\") pod \"ovnkube-node-f6pgc\" (UID: \"5b193bd2-b536-4056-92f1-94c9836ab2eb\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.949676 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/50c69d41-a65b-463e-a829-ea349bd2fe46-os-release\") pod \"multus-additional-cni-plugins-kh8g9\" (UID: \"50c69d41-a65b-463e-a829-ea349bd2fe46\") " pod="openshift-multus/multus-additional-cni-plugins-kh8g9" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.949693 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5b193bd2-b536-4056-92f1-94c9836ab2eb-env-overrides\") pod \"ovnkube-node-f6pgc\" (UID: \"5b193bd2-b536-4056-92f1-94c9836ab2eb\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.949707 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5b193bd2-b536-4056-92f1-94c9836ab2eb-ovn-node-metrics-cert\") pod \"ovnkube-node-f6pgc\" (UID: \"5b193bd2-b536-4056-92f1-94c9836ab2eb\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.949723 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/8022f696-d0e7-437b-848a-3cd25bd1f364-cni-binary-copy\") pod \"multus-6bkc2\" (UID: \"8022f696-d0e7-437b-848a-3cd25bd1f364\") " pod="openshift-multus/multus-6bkc2" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.949736 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-host-cni-bin\") pod \"ovnkube-node-f6pgc\" (UID: \"5b193bd2-b536-4056-92f1-94c9836ab2eb\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.949752 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/8022f696-d0e7-437b-848a-3cd25bd1f364-host-var-lib-cni-multus\") pod \"multus-6bkc2\" (UID: \"8022f696-d0e7-437b-848a-3cd25bd1f364\") " pod="openshift-multus/multus-6bkc2" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.949767 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/50c69d41-a65b-463e-a829-ea349bd2fe46-cnibin\") pod \"multus-additional-cni-plugins-kh8g9\" (UID: \"50c69d41-a65b-463e-a829-ea349bd2fe46\") " pod="openshift-multus/multus-additional-cni-plugins-kh8g9" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.949781 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-host-slash\") pod \"ovnkube-node-f6pgc\" (UID: \"5b193bd2-b536-4056-92f1-94c9836ab2eb\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.949796 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-node-log\") pod \"ovnkube-node-f6pgc\" (UID: \"5b193bd2-b536-4056-92f1-94c9836ab2eb\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.949810 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-host-run-ovn-kubernetes\") pod \"ovnkube-node-f6pgc\" (UID: \"5b193bd2-b536-4056-92f1-94c9836ab2eb\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.949824 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/8022f696-d0e7-437b-848a-3cd25bd1f364-cnibin\") pod \"multus-6bkc2\" (UID: \"8022f696-d0e7-437b-848a-3cd25bd1f364\") " pod="openshift-multus/multus-6bkc2" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.949838 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/50c69d41-a65b-463e-a829-ea349bd2fe46-system-cni-dir\") pod \"multus-additional-cni-plugins-kh8g9\" (UID: \"50c69d41-a65b-463e-a829-ea349bd2fe46\") " pod="openshift-multus/multus-additional-cni-plugins-kh8g9" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.949852 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/8022f696-d0e7-437b-848a-3cd25bd1f364-hostroot\") pod \"multus-6bkc2\" (UID: \"8022f696-d0e7-437b-848a-3cd25bd1f364\") " pod="openshift-multus/multus-6bkc2" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.949866 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8022f696-d0e7-437b-848a-3cd25bd1f364-etc-kubernetes\") pod \"multus-6bkc2\" (UID: \"8022f696-d0e7-437b-848a-3cd25bd1f364\") " pod="openshift-multus/multus-6bkc2" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.949889 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5b193bd2-b536-4056-92f1-94c9836ab2eb-ovnkube-config\") pod \"ovnkube-node-f6pgc\" (UID: \"5b193bd2-b536-4056-92f1-94c9836ab2eb\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.949909 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-host-cni-netd\") pod \"ovnkube-node-f6pgc\" (UID: \"5b193bd2-b536-4056-92f1-94c9836ab2eb\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.949922 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8022f696-d0e7-437b-848a-3cd25bd1f364-host-run-netns\") pod \"multus-6bkc2\" (UID: \"8022f696-d0e7-437b-848a-3cd25bd1f364\") " pod="openshift-multus/multus-6bkc2" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.949937 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-host-kubelet\") pod \"ovnkube-node-f6pgc\" (UID: \"5b193bd2-b536-4056-92f1-94c9836ab2eb\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.949950 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/8022f696-d0e7-437b-848a-3cd25bd1f364-host-run-k8s-cni-cncf-io\") pod \"multus-6bkc2\" (UID: \"8022f696-d0e7-437b-848a-3cd25bd1f364\") " pod="openshift-multus/multus-6bkc2" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.949965 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-run-ovn\") pod \"ovnkube-node-f6pgc\" (UID: \"5b193bd2-b536-4056-92f1-94c9836ab2eb\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.949981 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qpqjk\" (UniqueName: \"kubernetes.io/projected/5b193bd2-b536-4056-92f1-94c9836ab2eb-kube-api-access-qpqjk\") pod \"ovnkube-node-f6pgc\" (UID: \"5b193bd2-b536-4056-92f1-94c9836ab2eb\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.949996 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/8022f696-d0e7-437b-848a-3cd25bd1f364-multus-socket-dir-parent\") pod \"multus-6bkc2\" (UID: \"8022f696-d0e7-437b-848a-3cd25bd1f364\") " pod="openshift-multus/multus-6bkc2" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.950012 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/50c69d41-a65b-463e-a829-ea349bd2fe46-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-kh8g9\" (UID: \"50c69d41-a65b-463e-a829-ea349bd2fe46\") " pod="openshift-multus/multus-additional-cni-plugins-kh8g9" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.950028 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6qdtz\" (UniqueName: \"kubernetes.io/projected/50c69d41-a65b-463e-a829-ea349bd2fe46-kube-api-access-6qdtz\") pod \"multus-additional-cni-plugins-kh8g9\" (UID: \"50c69d41-a65b-463e-a829-ea349bd2fe46\") " pod="openshift-multus/multus-additional-cni-plugins-kh8g9" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.950041 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8022f696-d0e7-437b-848a-3cd25bd1f364-multus-cni-dir\") pod \"multus-6bkc2\" (UID: \"8022f696-d0e7-437b-848a-3cd25bd1f364\") " pod="openshift-multus/multus-6bkc2" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.950055 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hh5kn\" (UniqueName: \"kubernetes.io/projected/8022f696-d0e7-437b-848a-3cd25bd1f364-kube-api-access-hh5kn\") pod \"multus-6bkc2\" (UID: \"8022f696-d0e7-437b-848a-3cd25bd1f364\") " pod="openshift-multus/multus-6bkc2" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.950073 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-etc-openvswitch\") pod \"ovnkube-node-f6pgc\" (UID: \"5b193bd2-b536-4056-92f1-94c9836ab2eb\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.950087 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-f6pgc\" (UID: \"5b193bd2-b536-4056-92f1-94c9836ab2eb\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.950102 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8022f696-d0e7-437b-848a-3cd25bd1f364-system-cni-dir\") pod \"multus-6bkc2\" (UID: \"8022f696-d0e7-437b-848a-3cd25bd1f364\") " pod="openshift-multus/multus-6bkc2" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.950153 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-log-socket\") pod \"ovnkube-node-f6pgc\" (UID: \"5b193bd2-b536-4056-92f1-94c9836ab2eb\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.950170 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/5b193bd2-b536-4056-92f1-94c9836ab2eb-ovnkube-script-lib\") pod \"ovnkube-node-f6pgc\" (UID: \"5b193bd2-b536-4056-92f1-94c9836ab2eb\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.950184 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/8022f696-d0e7-437b-848a-3cd25bd1f364-host-var-lib-kubelet\") pod \"multus-6bkc2\" (UID: \"8022f696-d0e7-437b-848a-3cd25bd1f364\") " pod="openshift-multus/multus-6bkc2" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.950200 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/8022f696-d0e7-437b-848a-3cd25bd1f364-multus-conf-dir\") pod \"multus-6bkc2\" (UID: \"8022f696-d0e7-437b-848a-3cd25bd1f364\") " pod="openshift-multus/multus-6bkc2" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.950214 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-run-openvswitch\") pod \"ovnkube-node-f6pgc\" (UID: \"5b193bd2-b536-4056-92f1-94c9836ab2eb\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.950228 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/44b156e4-64a4-4d45-aa5e-9b10a862faed-proxy-tls\") pod \"machine-config-daemon-gx6xt\" (UID: \"44b156e4-64a4-4d45-aa5e-9b10a862faed\") " pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.950249 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/8022f696-d0e7-437b-848a-3cd25bd1f364-multus-daemon-config\") pod \"multus-6bkc2\" (UID: \"8022f696-d0e7-437b-848a-3cd25bd1f364\") " pod="openshift-multus/multus-6bkc2" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.950264 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/8022f696-d0e7-437b-848a-3cd25bd1f364-host-run-multus-certs\") pod \"multus-6bkc2\" (UID: \"8022f696-d0e7-437b-848a-3cd25bd1f364\") " pod="openshift-multus/multus-6bkc2" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.950279 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/44b156e4-64a4-4d45-aa5e-9b10a862faed-mcd-auth-proxy-config\") pod \"machine-config-daemon-gx6xt\" (UID: \"44b156e4-64a4-4d45-aa5e-9b10a862faed\") " pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.950300 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/50c69d41-a65b-463e-a829-ea349bd2fe46-tuning-conf-dir\") pod \"multus-additional-cni-plugins-kh8g9\" (UID: \"50c69d41-a65b-463e-a829-ea349bd2fe46\") " pod="openshift-multus/multus-additional-cni-plugins-kh8g9" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.950314 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-run-systemd\") pod \"ovnkube-node-f6pgc\" (UID: \"5b193bd2-b536-4056-92f1-94c9836ab2eb\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.950328 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w62dt\" (UniqueName: \"kubernetes.io/projected/44b156e4-64a4-4d45-aa5e-9b10a862faed-kube-api-access-w62dt\") pod \"machine-config-daemon-gx6xt\" (UID: \"44b156e4-64a4-4d45-aa5e-9b10a862faed\") " pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.950348 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8022f696-d0e7-437b-848a-3cd25bd1f364-host-var-lib-cni-bin\") pod \"multus-6bkc2\" (UID: \"8022f696-d0e7-437b-848a-3cd25bd1f364\") " pod="openshift-multus/multus-6bkc2" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.950362 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/44b156e4-64a4-4d45-aa5e-9b10a862faed-rootfs\") pod \"machine-config-daemon-gx6xt\" (UID: \"44b156e4-64a4-4d45-aa5e-9b10a862faed\") " pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.950377 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/8022f696-d0e7-437b-848a-3cd25bd1f364-os-release\") pod \"multus-6bkc2\" (UID: \"8022f696-d0e7-437b-848a-3cd25bd1f364\") " pod="openshift-multus/multus-6bkc2" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.950459 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/8022f696-d0e7-437b-848a-3cd25bd1f364-os-release\") pod \"multus-6bkc2\" (UID: \"8022f696-d0e7-437b-848a-3cd25bd1f364\") " pod="openshift-multus/multus-6bkc2" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.950771 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/8022f696-d0e7-437b-848a-3cd25bd1f364-host-run-k8s-cni-cncf-io\") pod \"multus-6bkc2\" (UID: \"8022f696-d0e7-437b-848a-3cd25bd1f364\") " pod="openshift-multus/multus-6bkc2" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.950824 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/50c69d41-a65b-463e-a829-ea349bd2fe46-cnibin\") pod \"multus-additional-cni-plugins-kh8g9\" (UID: \"50c69d41-a65b-463e-a829-ea349bd2fe46\") " pod="openshift-multus/multus-additional-cni-plugins-kh8g9" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.950857 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-node-log\") pod \"ovnkube-node-f6pgc\" (UID: \"5b193bd2-b536-4056-92f1-94c9836ab2eb\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.950881 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-host-run-ovn-kubernetes\") pod \"ovnkube-node-f6pgc\" (UID: \"5b193bd2-b536-4056-92f1-94c9836ab2eb\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.950923 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/8022f696-d0e7-437b-848a-3cd25bd1f364-cnibin\") pod \"multus-6bkc2\" (UID: \"8022f696-d0e7-437b-848a-3cd25bd1f364\") " pod="openshift-multus/multus-6bkc2" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.950981 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-systemd-units\") pod \"ovnkube-node-f6pgc\" (UID: \"5b193bd2-b536-4056-92f1-94c9836ab2eb\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.951018 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-host-run-netns\") pod \"ovnkube-node-f6pgc\" (UID: \"5b193bd2-b536-4056-92f1-94c9836ab2eb\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.951072 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-var-lib-openvswitch\") pod \"ovnkube-node-f6pgc\" (UID: \"5b193bd2-b536-4056-92f1-94c9836ab2eb\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.951163 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/50c69d41-a65b-463e-a829-ea349bd2fe46-os-release\") pod \"multus-additional-cni-plugins-kh8g9\" (UID: \"50c69d41-a65b-463e-a829-ea349bd2fe46\") " pod="openshift-multus/multus-additional-cni-plugins-kh8g9" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.951178 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/8022f696-d0e7-437b-848a-3cd25bd1f364-host-var-lib-kubelet\") pod \"multus-6bkc2\" (UID: \"8022f696-d0e7-437b-848a-3cd25bd1f364\") " pod="openshift-multus/multus-6bkc2" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.951209 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/8022f696-d0e7-437b-848a-3cd25bd1f364-multus-conf-dir\") pod \"multus-6bkc2\" (UID: \"8022f696-d0e7-437b-848a-3cd25bd1f364\") " pod="openshift-multus/multus-6bkc2" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.951233 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-run-openvswitch\") pod \"ovnkube-node-f6pgc\" (UID: \"5b193bd2-b536-4056-92f1-94c9836ab2eb\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.951544 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/50c69d41-a65b-463e-a829-ea349bd2fe46-cni-binary-copy\") pod \"multus-additional-cni-plugins-kh8g9\" (UID: \"50c69d41-a65b-463e-a829-ea349bd2fe46\") " pod="openshift-multus/multus-additional-cni-plugins-kh8g9" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.951583 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/50c69d41-a65b-463e-a829-ea349bd2fe46-system-cni-dir\") pod \"multus-additional-cni-plugins-kh8g9\" (UID: \"50c69d41-a65b-463e-a829-ea349bd2fe46\") " pod="openshift-multus/multus-additional-cni-plugins-kh8g9" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.951623 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/8022f696-d0e7-437b-848a-3cd25bd1f364-hostroot\") pod \"multus-6bkc2\" (UID: \"8022f696-d0e7-437b-848a-3cd25bd1f364\") " pod="openshift-multus/multus-6bkc2" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.951630 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-run-systemd\") pod \"ovnkube-node-f6pgc\" (UID: \"5b193bd2-b536-4056-92f1-94c9836ab2eb\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.951645 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8022f696-d0e7-437b-848a-3cd25bd1f364-etc-kubernetes\") pod \"multus-6bkc2\" (UID: \"8022f696-d0e7-437b-848a-3cd25bd1f364\") " pod="openshift-multus/multus-6bkc2" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.951972 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8022f696-d0e7-437b-848a-3cd25bd1f364-host-var-lib-cni-bin\") pod \"multus-6bkc2\" (UID: \"8022f696-d0e7-437b-848a-3cd25bd1f364\") " pod="openshift-multus/multus-6bkc2" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.952002 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/44b156e4-64a4-4d45-aa5e-9b10a862faed-rootfs\") pod \"machine-config-daemon-gx6xt\" (UID: \"44b156e4-64a4-4d45-aa5e-9b10a862faed\") " pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.952034 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8022f696-d0e7-437b-848a-3cd25bd1f364-host-run-netns\") pod \"multus-6bkc2\" (UID: \"8022f696-d0e7-437b-848a-3cd25bd1f364\") " pod="openshift-multus/multus-6bkc2" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.952055 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-host-cni-netd\") pod \"ovnkube-node-f6pgc\" (UID: \"5b193bd2-b536-4056-92f1-94c9836ab2eb\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.952082 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-host-kubelet\") pod \"ovnkube-node-f6pgc\" (UID: \"5b193bd2-b536-4056-92f1-94c9836ab2eb\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.952226 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/8022f696-d0e7-437b-848a-3cd25bd1f364-multus-daemon-config\") pod \"multus-6bkc2\" (UID: \"8022f696-d0e7-437b-848a-3cd25bd1f364\") " pod="openshift-multus/multus-6bkc2" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.952609 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/44b156e4-64a4-4d45-aa5e-9b10a862faed-mcd-auth-proxy-config\") pod \"machine-config-daemon-gx6xt\" (UID: \"44b156e4-64a4-4d45-aa5e-9b10a862faed\") " pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.952648 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/8022f696-d0e7-437b-848a-3cd25bd1f364-host-run-multus-certs\") pod \"multus-6bkc2\" (UID: \"8022f696-d0e7-437b-848a-3cd25bd1f364\") " pod="openshift-multus/multus-6bkc2" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.952675 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-host-cni-bin\") pod \"ovnkube-node-f6pgc\" (UID: \"5b193bd2-b536-4056-92f1-94c9836ab2eb\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.953098 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/8022f696-d0e7-437b-848a-3cd25bd1f364-cni-binary-copy\") pod \"multus-6bkc2\" (UID: \"8022f696-d0e7-437b-848a-3cd25bd1f364\") " pod="openshift-multus/multus-6bkc2" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.953160 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5b193bd2-b536-4056-92f1-94c9836ab2eb-env-overrides\") pod \"ovnkube-node-f6pgc\" (UID: \"5b193bd2-b536-4056-92f1-94c9836ab2eb\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.953336 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8022f696-d0e7-437b-848a-3cd25bd1f364-multus-cni-dir\") pod \"multus-6bkc2\" (UID: \"8022f696-d0e7-437b-848a-3cd25bd1f364\") " pod="openshift-multus/multus-6bkc2" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.953266 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-f6pgc\" (UID: \"5b193bd2-b536-4056-92f1-94c9836ab2eb\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.953593 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/8022f696-d0e7-437b-848a-3cd25bd1f364-multus-socket-dir-parent\") pod \"multus-6bkc2\" (UID: \"8022f696-d0e7-437b-848a-3cd25bd1f364\") " pod="openshift-multus/multus-6bkc2" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.953616 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-run-ovn\") pod \"ovnkube-node-f6pgc\" (UID: \"5b193bd2-b536-4056-92f1-94c9836ab2eb\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.953810 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/8022f696-d0e7-437b-848a-3cd25bd1f364-host-var-lib-cni-multus\") pod \"multus-6bkc2\" (UID: \"8022f696-d0e7-437b-848a-3cd25bd1f364\") " pod="openshift-multus/multus-6bkc2" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.953835 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-etc-openvswitch\") pod \"ovnkube-node-f6pgc\" (UID: \"5b193bd2-b536-4056-92f1-94c9836ab2eb\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.953895 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/50c69d41-a65b-463e-a829-ea349bd2fe46-tuning-conf-dir\") pod \"multus-additional-cni-plugins-kh8g9\" (UID: \"50c69d41-a65b-463e-a829-ea349bd2fe46\") " pod="openshift-multus/multus-additional-cni-plugins-kh8g9" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.950836 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-host-slash\") pod \"ovnkube-node-f6pgc\" (UID: \"5b193bd2-b536-4056-92f1-94c9836ab2eb\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.954067 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8022f696-d0e7-437b-848a-3cd25bd1f364-system-cni-dir\") pod \"multus-6bkc2\" (UID: \"8022f696-d0e7-437b-848a-3cd25bd1f364\") " pod="openshift-multus/multus-6bkc2" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.954091 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-log-socket\") pod \"ovnkube-node-f6pgc\" (UID: \"5b193bd2-b536-4056-92f1-94c9836ab2eb\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.955354 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5b193bd2-b536-4056-92f1-94c9836ab2eb-ovn-node-metrics-cert\") pod \"ovnkube-node-f6pgc\" (UID: \"5b193bd2-b536-4056-92f1-94c9836ab2eb\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.955446 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/50c69d41-a65b-463e-a829-ea349bd2fe46-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-kh8g9\" (UID: \"50c69d41-a65b-463e-a829-ea349bd2fe46\") " pod="openshift-multus/multus-additional-cni-plugins-kh8g9" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.955484 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.957536 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/44b156e4-64a4-4d45-aa5e-9b10a862faed-proxy-tls\") pod \"machine-config-daemon-gx6xt\" (UID: \"44b156e4-64a4-4d45-aa5e-9b10a862faed\") " pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.957992 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/5b193bd2-b536-4056-92f1-94c9836ab2eb-ovnkube-script-lib\") pod \"ovnkube-node-f6pgc\" (UID: \"5b193bd2-b536-4056-92f1-94c9836ab2eb\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.958727 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5b193bd2-b536-4056-92f1-94c9836ab2eb-ovnkube-config\") pod \"ovnkube-node-f6pgc\" (UID: \"5b193bd2-b536-4056-92f1-94c9836ab2eb\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.973904 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44b156e4-64a4-4d45-aa5e-9b10a862faed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w62dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w62dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gx6xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.974474 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qpqjk\" (UniqueName: \"kubernetes.io/projected/5b193bd2-b536-4056-92f1-94c9836ab2eb-kube-api-access-qpqjk\") pod \"ovnkube-node-f6pgc\" (UID: \"5b193bd2-b536-4056-92f1-94c9836ab2eb\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.974787 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w62dt\" (UniqueName: \"kubernetes.io/projected/44b156e4-64a4-4d45-aa5e-9b10a862faed-kube-api-access-w62dt\") pod \"machine-config-daemon-gx6xt\" (UID: \"44b156e4-64a4-4d45-aa5e-9b10a862faed\") " pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.975662 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6qdtz\" (UniqueName: \"kubernetes.io/projected/50c69d41-a65b-463e-a829-ea349bd2fe46-kube-api-access-6qdtz\") pod \"multus-additional-cni-plugins-kh8g9\" (UID: \"50c69d41-a65b-463e-a829-ea349bd2fe46\") " pod="openshift-multus/multus-additional-cni-plugins-kh8g9" Dec 11 02:01:17 crc kubenswrapper[4824]: I1211 02:01:17.993907 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hh5kn\" (UniqueName: \"kubernetes.io/projected/8022f696-d0e7-437b-848a-3cd25bd1f364-kube-api-access-hh5kn\") pod \"multus-6bkc2\" (UID: \"8022f696-d0e7-437b-848a-3cd25bd1f364\") " pod="openshift-multus/multus-6bkc2" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:17.999986 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.040414 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"beab386b-6c65-4bc3-9ef0-dc3f68b4403f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c26f06d2ed5e6e98b61a51f073130675f2c49acda106b8aedcbb2a6d4f7580c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20af9b10a67c8b21a7abcd23e42bc412798d159c8422e7c8ff58f5b997ca19a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9bbfc6edfe050b3fa9dbeee5f98d416bcb2a1533a38608a881b75396fd0b059\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d45819f0611c91b20c83eeee1092031a3441fb69f3015241632ad419880227d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d45819f0611c91b20c83eeee1092031a3441fb69f3015241632ad419880227d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 02:01:11.070250 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 02:01:11.071673 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2752882565/tls.crt::/tmp/serving-cert-2752882565/tls.key\\\\\\\"\\\\nI1211 02:01:17.042610 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 02:01:17.048046 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 02:01:17.048080 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 02:01:17.048172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 02:01:17.048188 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 02:01:17.058248 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 02:01:17.058294 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 02:01:17.058303 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 02:01:17.058315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 02:01:17.058322 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 02:01:17.058328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 02:01:17.058335 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 02:01:17.058808 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 02:01:17.060225 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14a940a72b2404a6cbd52a69e150b03266bcd7f5c1d3f4d07b51c0c69e51786\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:00:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.044315 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-kh8g9" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.077405 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.085556 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-6bkc2" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.111001 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.111027 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.111609 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.144446 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kptws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68568dba-83a5-46c1-8d01-337359679d91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd7nf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kptws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 02:01:18 crc kubenswrapper[4824]: W1211 02:01:18.173155 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5b193bd2_b536_4056_92f1_94c9836ab2eb.slice/crio-b82eae5d23b57e21a6683458a036bb2e02cb72681d142360940afa5217ed72e3 WatchSource:0}: Error finding container b82eae5d23b57e21a6683458a036bb2e02cb72681d142360940afa5217ed72e3: Status 404 returned error can't find the container with id b82eae5d23b57e21a6683458a036bb2e02cb72681d142360940afa5217ed72e3 Dec 11 02:01:18 crc kubenswrapper[4824]: W1211 02:01:18.216997 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod44b156e4_64a4_4d45_aa5e_9b10a862faed.slice/crio-8425bb86d1c7f3a68b6eb965004bbc789a10556eb68397541435629d5b964a2c WatchSource:0}: Error finding container 8425bb86d1c7f3a68b6eb965004bbc789a10556eb68397541435629d5b964a2c: Status 404 returned error can't find the container with id 8425bb86d1c7f3a68b6eb965004bbc789a10556eb68397541435629d5b964a2c Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.253935 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 02:01:18 crc kubenswrapper[4824]: E1211 02:01:18.254319 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 02:01:19.25429493 +0000 UTC m=+20.943332309 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.354756 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.354816 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.354842 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.354861 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 02:01:18 crc kubenswrapper[4824]: E1211 02:01:18.354917 4824 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 11 02:01:18 crc kubenswrapper[4824]: E1211 02:01:18.354977 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 11 02:01:18 crc kubenswrapper[4824]: E1211 02:01:18.354992 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 11 02:01:18 crc kubenswrapper[4824]: E1211 02:01:18.354994 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 11 02:01:18 crc kubenswrapper[4824]: E1211 02:01:18.355003 4824 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 02:01:18 crc kubenswrapper[4824]: E1211 02:01:18.355014 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 11 02:01:18 crc kubenswrapper[4824]: E1211 02:01:18.355025 4824 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 02:01:18 crc kubenswrapper[4824]: E1211 02:01:18.354995 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-11 02:01:19.354978068 +0000 UTC m=+21.044015447 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 11 02:01:18 crc kubenswrapper[4824]: E1211 02:01:18.355057 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-11 02:01:19.35504501 +0000 UTC m=+21.044082389 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 02:01:18 crc kubenswrapper[4824]: E1211 02:01:18.355069 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-11 02:01:19.355063691 +0000 UTC m=+21.044101070 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 02:01:18 crc kubenswrapper[4824]: E1211 02:01:18.355353 4824 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 11 02:01:18 crc kubenswrapper[4824]: E1211 02:01:18.355391 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-11 02:01:19.355383729 +0000 UTC m=+21.044421108 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.475163 4824 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Dec 11 02:01:18 crc kubenswrapper[4824]: W1211 02:01:18.475485 4824 reflector.go:484] object-"openshift-network-node-identity"/"env-overrides": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-node-identity"/"env-overrides": Unexpected watch close - watch lasted less than a second and no items received Dec 11 02:01:18 crc kubenswrapper[4824]: W1211 02:01:18.475514 4824 reflector.go:484] object-"openshift-network-node-identity"/"ovnkube-identity-cm": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-node-identity"/"ovnkube-identity-cm": Unexpected watch close - watch lasted less than a second and no items received Dec 11 02:01:18 crc kubenswrapper[4824]: W1211 02:01:18.475498 4824 reflector.go:484] object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq": watch of *v1.Secret ended with: very short watch: object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq": Unexpected watch close - watch lasted less than a second and no items received Dec 11 02:01:18 crc kubenswrapper[4824]: W1211 02:01:18.475547 4824 reflector.go:484] object-"openshift-network-operator"/"metrics-tls": watch of *v1.Secret ended with: very short watch: object-"openshift-network-operator"/"metrics-tls": Unexpected watch close - watch lasted less than a second and no items received Dec 11 02:01:18 crc kubenswrapper[4824]: W1211 02:01:18.475566 4824 reflector.go:484] object-"openshift-machine-config-operator"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-machine-config-operator"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 11 02:01:18 crc kubenswrapper[4824]: W1211 02:01:18.475597 4824 reflector.go:484] object-"openshift-network-operator"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-operator"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 11 02:01:18 crc kubenswrapper[4824]: W1211 02:01:18.475615 4824 reflector.go:484] object-"openshift-multus"/"default-dockercfg-2q5b6": watch of *v1.Secret ended with: very short watch: object-"openshift-multus"/"default-dockercfg-2q5b6": Unexpected watch close - watch lasted less than a second and no items received Dec 11 02:01:18 crc kubenswrapper[4824]: W1211 02:01:18.475655 4824 reflector.go:484] object-"openshift-ovn-kubernetes"/"ovnkube-script-lib": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-ovn-kubernetes"/"ovnkube-script-lib": Unexpected watch close - watch lasted less than a second and no items received Dec 11 02:01:18 crc kubenswrapper[4824]: W1211 02:01:18.475697 4824 reflector.go:484] object-"openshift-machine-config-operator"/"proxy-tls": watch of *v1.Secret ended with: very short watch: object-"openshift-machine-config-operator"/"proxy-tls": Unexpected watch close - watch lasted less than a second and no items received Dec 11 02:01:18 crc kubenswrapper[4824]: W1211 02:01:18.475716 4824 reflector.go:484] object-"openshift-machine-config-operator"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-machine-config-operator"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 11 02:01:18 crc kubenswrapper[4824]: W1211 02:01:18.475726 4824 reflector.go:484] object-"openshift-network-node-identity"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-node-identity"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 11 02:01:18 crc kubenswrapper[4824]: W1211 02:01:18.475743 4824 reflector.go:484] object-"openshift-network-node-identity"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-node-identity"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 11 02:01:18 crc kubenswrapper[4824]: W1211 02:01:18.475752 4824 reflector.go:484] object-"openshift-network-node-identity"/"network-node-identity-cert": watch of *v1.Secret ended with: very short watch: object-"openshift-network-node-identity"/"network-node-identity-cert": Unexpected watch close - watch lasted less than a second and no items received Dec 11 02:01:18 crc kubenswrapper[4824]: W1211 02:01:18.475766 4824 reflector.go:484] object-"openshift-machine-config-operator"/"kube-rbac-proxy": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-machine-config-operator"/"kube-rbac-proxy": Unexpected watch close - watch lasted less than a second and no items received Dec 11 02:01:18 crc kubenswrapper[4824]: W1211 02:01:18.475776 4824 reflector.go:484] object-"openshift-network-operator"/"iptables-alerter-script": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-operator"/"iptables-alerter-script": Unexpected watch close - watch lasted less than a second and no items received Dec 11 02:01:18 crc kubenswrapper[4824]: W1211 02:01:18.475790 4824 reflector.go:484] object-"openshift-ovn-kubernetes"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-ovn-kubernetes"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 11 02:01:18 crc kubenswrapper[4824]: W1211 02:01:18.475800 4824 reflector.go:484] object-"openshift-multus"/"multus-daemon-config": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-multus"/"multus-daemon-config": Unexpected watch close - watch lasted less than a second and no items received Dec 11 02:01:18 crc kubenswrapper[4824]: W1211 02:01:18.475816 4824 reflector.go:484] object-"openshift-multus"/"default-cni-sysctl-allowlist": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-multus"/"default-cni-sysctl-allowlist": Unexpected watch close - watch lasted less than a second and no items received Dec 11 02:01:18 crc kubenswrapper[4824]: W1211 02:01:18.475817 4824 reflector.go:484] object-"openshift-ovn-kubernetes"/"env-overrides": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-ovn-kubernetes"/"env-overrides": Unexpected watch close - watch lasted less than a second and no items received Dec 11 02:01:18 crc kubenswrapper[4824]: W1211 02:01:18.475832 4824 reflector.go:484] object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 11 02:01:18 crc kubenswrapper[4824]: W1211 02:01:18.475843 4824 reflector.go:484] object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz": watch of *v1.Secret ended with: very short watch: object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz": Unexpected watch close - watch lasted less than a second and no items received Dec 11 02:01:18 crc kubenswrapper[4824]: W1211 02:01:18.475857 4824 reflector.go:484] object-"openshift-multus"/"cni-copy-resources": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-multus"/"cni-copy-resources": Unexpected watch close - watch lasted less than a second and no items received Dec 11 02:01:18 crc kubenswrapper[4824]: W1211 02:01:18.475879 4824 reflector.go:484] object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert": watch of *v1.Secret ended with: very short watch: object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert": Unexpected watch close - watch lasted less than a second and no items received Dec 11 02:01:18 crc kubenswrapper[4824]: W1211 02:01:18.475904 4824 reflector.go:484] object-"openshift-ovn-kubernetes"/"ovnkube-config": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-ovn-kubernetes"/"ovnkube-config": Unexpected watch close - watch lasted less than a second and no items received Dec 11 02:01:18 crc kubenswrapper[4824]: W1211 02:01:18.475926 4824 reflector.go:484] object-"openshift-dns"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-dns"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 11 02:01:18 crc kubenswrapper[4824]: W1211 02:01:18.475931 4824 reflector.go:484] object-"openshift-network-operator"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-operator"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 11 02:01:18 crc kubenswrapper[4824]: W1211 02:01:18.475934 4824 reflector.go:484] object-"openshift-multus"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-multus"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 11 02:01:18 crc kubenswrapper[4824]: W1211 02:01:18.475944 4824 reflector.go:484] object-"openshift-multus"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-multus"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 11 02:01:18 crc kubenswrapper[4824]: W1211 02:01:18.475955 4824 reflector.go:484] object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl": watch of *v1.Secret ended with: very short watch: object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl": Unexpected watch close - watch lasted less than a second and no items received Dec 11 02:01:18 crc kubenswrapper[4824]: W1211 02:01:18.476023 4824 reflector.go:484] object-"openshift-dns"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-dns"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 11 02:01:18 crc kubenswrapper[4824]: W1211 02:01:18.476177 4824 reflector.go:484] object-"openshift-dns"/"node-resolver-dockercfg-kz9s7": watch of *v1.Secret ended with: very short watch: object-"openshift-dns"/"node-resolver-dockercfg-kz9s7": Unexpected watch close - watch lasted less than a second and no items received Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.635722 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.636276 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.637242 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.638400 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.639463 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.639947 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.640549 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.641695 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.642320 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.643185 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.643686 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.644705 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.645222 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.646143 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.646638 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.647140 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.648022 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.648451 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.649377 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.649961 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.650017 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"beab386b-6c65-4bc3-9ef0-dc3f68b4403f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c26f06d2ed5e6e98b61a51f073130675f2c49acda106b8aedcbb2a6d4f7580c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20af9b10a67c8b21a7abcd23e42bc412798d159c8422e7c8ff58f5b997ca19a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9bbfc6edfe050b3fa9dbeee5f98d416bcb2a1533a38608a881b75396fd0b059\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d45819f0611c91b20c83eeee1092031a3441fb69f3015241632ad419880227d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d45819f0611c91b20c83eeee1092031a3441fb69f3015241632ad419880227d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 02:01:11.070250 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 02:01:11.071673 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2752882565/tls.crt::/tmp/serving-cert-2752882565/tls.key\\\\\\\"\\\\nI1211 02:01:17.042610 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 02:01:17.048046 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 02:01:17.048080 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 02:01:17.048172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 02:01:17.048188 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 02:01:17.058248 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 02:01:17.058294 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 02:01:17.058303 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 02:01:17.058315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 02:01:17.058322 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 02:01:17.058328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 02:01:17.058335 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 02:01:17.058808 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 02:01:17.060225 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14a940a72b2404a6cbd52a69e150b03266bcd7f5c1d3f4d07b51c0c69e51786\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:00:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:18Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.650444 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.651347 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.651771 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.652817 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.653258 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.654254 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.654841 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.655300 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.656186 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.656651 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.657479 4824 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.657584 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.659162 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.659977 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.660486 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.661905 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.662540 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.663396 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.663991 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.664018 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:18Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.664955 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.665491 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.666423 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.667405 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.668197 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.668970 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.669596 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.670528 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.671296 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.671811 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.673292 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.673944 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.675002 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.675564 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.676034 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.676677 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:18Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.685622 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kptws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68568dba-83a5-46c1-8d01-337359679d91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd7nf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kptws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:18Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.697195 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:18Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.709919 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kh8g9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"50c69d41-a65b-463e-a829-ea349bd2fe46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kh8g9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:18Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.720569 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6bkc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8022f696-d0e7-437b-848a-3cd25bd1f364\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hh5kn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6bkc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:18Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.742924 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b193bd2-b536-4056-92f1-94c9836ab2eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f6pgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:18Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.759264 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:18Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.777507 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:18Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.785623 4824 generic.go:334] "Generic (PLEG): container finished" podID="5b193bd2-b536-4056-92f1-94c9836ab2eb" containerID="4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df" exitCode=0 Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.785690 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" event={"ID":"5b193bd2-b536-4056-92f1-94c9836ab2eb","Type":"ContainerDied","Data":"4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df"} Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.785719 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" event={"ID":"5b193bd2-b536-4056-92f1-94c9836ab2eb","Type":"ContainerStarted","Data":"b82eae5d23b57e21a6683458a036bb2e02cb72681d142360940afa5217ed72e3"} Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.789811 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"55bdf5b815062892e0211cb4f47ed961d088e6a3db652fd475367973d56e7b0b"} Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.794003 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-kptws" event={"ID":"68568dba-83a5-46c1-8d01-337359679d91","Type":"ContainerStarted","Data":"f965a6d6db619995132f61b5d51b66f0b2b3fdbef3fbec63e231b9a62ccd0a9f"} Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.794040 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-kptws" event={"ID":"68568dba-83a5-46c1-8d01-337359679d91","Type":"ContainerStarted","Data":"83d92dcd3ead736380fc5989919243591e91c292bb092e180839602c5c12a5d7"} Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.795131 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-6bkc2" event={"ID":"8022f696-d0e7-437b-848a-3cd25bd1f364","Type":"ContainerStarted","Data":"f3e28f6778ec0dba6ea3e6b69e5a2c1c73c0d5db1abeead241fb58efdbfaefe4"} Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.795152 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-6bkc2" event={"ID":"8022f696-d0e7-437b-848a-3cd25bd1f364","Type":"ContainerStarted","Data":"c02098178d38d95de45a6e90d95e8910916100f2976af9cb4f0792d823067b98"} Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.796352 4824 generic.go:334] "Generic (PLEG): container finished" podID="50c69d41-a65b-463e-a829-ea349bd2fe46" containerID="69af1f33fd0bafc929b94e702b4f46c6c7d7a362d5552b121729c477892d91d4" exitCode=0 Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.796405 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-kh8g9" event={"ID":"50c69d41-a65b-463e-a829-ea349bd2fe46","Type":"ContainerDied","Data":"69af1f33fd0bafc929b94e702b4f46c6c7d7a362d5552b121729c477892d91d4"} Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.796422 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-kh8g9" event={"ID":"50c69d41-a65b-463e-a829-ea349bd2fe46","Type":"ContainerStarted","Data":"077fa9d064f04aef59a55b2757035b25ecd5be1fe3befb27e62ef1d063311077"} Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.800439 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44b156e4-64a4-4d45-aa5e-9b10a862faed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w62dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w62dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gx6xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:18Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.807325 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"76b10bbd1e891b8c682f4b2e17d023fd787a8ed44ed94fbf7d97314d641ccf36"} Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.807382 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"1f32838f33832b9b67e7dd9c53340680e17721cf1c7bdf0170230c5f8feafe64"} Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.807394 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"e39c981e597138f82a759254cf02b48e8f10947074ea1521b2775bd56a5f4bf2"} Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.817498 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:18Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.817802 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" event={"ID":"44b156e4-64a4-4d45-aa5e-9b10a862faed","Type":"ContainerStarted","Data":"7368d92144b188e95a7796fabf497b0f8d9c0b240293bdc7b4aec1ff75497bc5"} Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.817832 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" event={"ID":"44b156e4-64a4-4d45-aa5e-9b10a862faed","Type":"ContainerStarted","Data":"dcf289ac27dbda4f24aabbde3fb5119329e7b4fc7d6e018168da9a69661a4c57"} Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.817841 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" event={"ID":"44b156e4-64a4-4d45-aa5e-9b10a862faed","Type":"ContainerStarted","Data":"8425bb86d1c7f3a68b6eb965004bbc789a10556eb68397541435629d5b964a2c"} Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.823914 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"2ded146c045085be2b58821749052413a36850ede00ae1024fe2dae9fa5ac84b"} Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.823947 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"7036011d2f004e504c495ded1af78dcf9eca62c3d69de9013b3161a1ee7e8a36"} Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.832349 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.841699 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:18Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.844178 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"b7c766f8de73ca762ffde0c18b2d1c46bb152b2b37535a611c5a0e454a21b7d0"} Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.844750 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.856665 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"beab386b-6c65-4bc3-9ef0-dc3f68b4403f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c26f06d2ed5e6e98b61a51f073130675f2c49acda106b8aedcbb2a6d4f7580c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20af9b10a67c8b21a7abcd23e42bc412798d159c8422e7c8ff58f5b997ca19a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9bbfc6edfe050b3fa9dbeee5f98d416bcb2a1533a38608a881b75396fd0b059\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d45819f0611c91b20c83eeee1092031a3441fb69f3015241632ad419880227d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d45819f0611c91b20c83eeee1092031a3441fb69f3015241632ad419880227d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 02:01:11.070250 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 02:01:11.071673 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2752882565/tls.crt::/tmp/serving-cert-2752882565/tls.key\\\\\\\"\\\\nI1211 02:01:17.042610 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 02:01:17.048046 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 02:01:17.048080 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 02:01:17.048172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 02:01:17.048188 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 02:01:17.058248 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 02:01:17.058294 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 02:01:17.058303 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 02:01:17.058315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 02:01:17.058322 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 02:01:17.058328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 02:01:17.058335 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 02:01:17.058808 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 02:01:17.060225 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14a940a72b2404a6cbd52a69e150b03266bcd7f5c1d3f4d07b51c0c69e51786\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:00:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:18Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.868778 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:18Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.881483 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76b10bbd1e891b8c682f4b2e17d023fd787a8ed44ed94fbf7d97314d641ccf36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f32838f33832b9b67e7dd9c53340680e17721cf1c7bdf0170230c5f8feafe64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:18Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.893853 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kptws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68568dba-83a5-46c1-8d01-337359679d91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f965a6d6db619995132f61b5d51b66f0b2b3fdbef3fbec63e231b9a62ccd0a9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd7nf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kptws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:18Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.904155 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:18Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.917165 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kh8g9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"50c69d41-a65b-463e-a829-ea349bd2fe46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69af1f33fd0bafc929b94e702b4f46c6c7d7a362d5552b121729c477892d91d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69af1f33fd0bafc929b94e702b4f46c6c7d7a362d5552b121729c477892d91d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kh8g9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:18Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.929638 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6bkc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8022f696-d0e7-437b-848a-3cd25bd1f364\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3e28f6778ec0dba6ea3e6b69e5a2c1c73c0d5db1abeead241fb58efdbfaefe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hh5kn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6bkc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:18Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.944180 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b193bd2-b536-4056-92f1-94c9836ab2eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f6pgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:18Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.957094 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:18Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.971548 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:18Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.980440 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44b156e4-64a4-4d45-aa5e-9b10a862faed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w62dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w62dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gx6xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:18Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:18 crc kubenswrapper[4824]: I1211 02:01:18.992054 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:18Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:19 crc kubenswrapper[4824]: I1211 02:01:19.005518 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"beab386b-6c65-4bc3-9ef0-dc3f68b4403f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c26f06d2ed5e6e98b61a51f073130675f2c49acda106b8aedcbb2a6d4f7580c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20af9b10a67c8b21a7abcd23e42bc412798d159c8422e7c8ff58f5b997ca19a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9bbfc6edfe050b3fa9dbeee5f98d416bcb2a1533a38608a881b75396fd0b059\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7c766f8de73ca762ffde0c18b2d1c46bb152b2b37535a611c5a0e454a21b7d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d45819f0611c91b20c83eeee1092031a3441fb69f3015241632ad419880227d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 02:01:11.070250 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 02:01:11.071673 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2752882565/tls.crt::/tmp/serving-cert-2752882565/tls.key\\\\\\\"\\\\nI1211 02:01:17.042610 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 02:01:17.048046 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 02:01:17.048080 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 02:01:17.048172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 02:01:17.048188 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 02:01:17.058248 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 02:01:17.058294 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 02:01:17.058303 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 02:01:17.058315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 02:01:17.058322 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 02:01:17.058328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 02:01:17.058335 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 02:01:17.058808 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 02:01:17.060225 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14a940a72b2404a6cbd52a69e150b03266bcd7f5c1d3f4d07b51c0c69e51786\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:00:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:19Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:19 crc kubenswrapper[4824]: I1211 02:01:19.018870 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ded146c045085be2b58821749052413a36850ede00ae1024fe2dae9fa5ac84b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:19Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:19 crc kubenswrapper[4824]: I1211 02:01:19.030626 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76b10bbd1e891b8c682f4b2e17d023fd787a8ed44ed94fbf7d97314d641ccf36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f32838f33832b9b67e7dd9c53340680e17721cf1c7bdf0170230c5f8feafe64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:19Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:19 crc kubenswrapper[4824]: I1211 02:01:19.040740 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kptws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68568dba-83a5-46c1-8d01-337359679d91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f965a6d6db619995132f61b5d51b66f0b2b3fdbef3fbec63e231b9a62ccd0a9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd7nf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kptws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:19Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:19 crc kubenswrapper[4824]: I1211 02:01:19.056527 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kh8g9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"50c69d41-a65b-463e-a829-ea349bd2fe46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69af1f33fd0bafc929b94e702b4f46c6c7d7a362d5552b121729c477892d91d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69af1f33fd0bafc929b94e702b4f46c6c7d7a362d5552b121729c477892d91d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kh8g9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:19Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:19 crc kubenswrapper[4824]: I1211 02:01:19.070049 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6bkc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8022f696-d0e7-437b-848a-3cd25bd1f364\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3e28f6778ec0dba6ea3e6b69e5a2c1c73c0d5db1abeead241fb58efdbfaefe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hh5kn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6bkc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:19Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:19 crc kubenswrapper[4824]: I1211 02:01:19.091389 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b193bd2-b536-4056-92f1-94c9836ab2eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f6pgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:19Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:19 crc kubenswrapper[4824]: I1211 02:01:19.104537 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:19Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:19 crc kubenswrapper[4824]: I1211 02:01:19.116300 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:19Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:19 crc kubenswrapper[4824]: I1211 02:01:19.130586 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44b156e4-64a4-4d45-aa5e-9b10a862faed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7368d92144b188e95a7796fabf497b0f8d9c0b240293bdc7b4aec1ff75497bc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w62dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcf289ac27dbda4f24aabbde3fb5119329e7b4fc7d6e018168da9a69661a4c57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w62dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gx6xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:19Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:19 crc kubenswrapper[4824]: I1211 02:01:19.155862 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:19Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:19 crc kubenswrapper[4824]: I1211 02:01:19.263563 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 02:01:19 crc kubenswrapper[4824]: E1211 02:01:19.263727 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 02:01:21.26371235 +0000 UTC m=+22.952749729 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:01:19 crc kubenswrapper[4824]: I1211 02:01:19.302735 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 11 02:01:19 crc kubenswrapper[4824]: I1211 02:01:19.329898 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 11 02:01:19 crc kubenswrapper[4824]: I1211 02:01:19.335644 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 11 02:01:19 crc kubenswrapper[4824]: I1211 02:01:19.338621 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 11 02:01:19 crc kubenswrapper[4824]: I1211 02:01:19.354341 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 11 02:01:19 crc kubenswrapper[4824]: I1211 02:01:19.364335 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 02:01:19 crc kubenswrapper[4824]: E1211 02:01:19.364566 4824 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 11 02:01:19 crc kubenswrapper[4824]: E1211 02:01:19.364652 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-11 02:01:21.364632145 +0000 UTC m=+23.053669594 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 11 02:01:19 crc kubenswrapper[4824]: I1211 02:01:19.364591 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 02:01:19 crc kubenswrapper[4824]: I1211 02:01:19.364837 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 02:01:19 crc kubenswrapper[4824]: I1211 02:01:19.364939 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 02:01:19 crc kubenswrapper[4824]: E1211 02:01:19.365012 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 11 02:01:19 crc kubenswrapper[4824]: E1211 02:01:19.365034 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 11 02:01:19 crc kubenswrapper[4824]: E1211 02:01:19.365048 4824 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 02:01:19 crc kubenswrapper[4824]: E1211 02:01:19.365080 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-11 02:01:21.365070648 +0000 UTC m=+23.054108027 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 02:01:19 crc kubenswrapper[4824]: E1211 02:01:19.365167 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 11 02:01:19 crc kubenswrapper[4824]: E1211 02:01:19.365183 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 11 02:01:19 crc kubenswrapper[4824]: E1211 02:01:19.365193 4824 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 02:01:19 crc kubenswrapper[4824]: E1211 02:01:19.365219 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-11 02:01:21.365210951 +0000 UTC m=+23.054248420 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 02:01:19 crc kubenswrapper[4824]: E1211 02:01:19.365354 4824 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 11 02:01:19 crc kubenswrapper[4824]: E1211 02:01:19.365395 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-11 02:01:21.365384526 +0000 UTC m=+23.054421995 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 11 02:01:19 crc kubenswrapper[4824]: I1211 02:01:19.431965 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 11 02:01:19 crc kubenswrapper[4824]: I1211 02:01:19.501169 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 11 02:01:19 crc kubenswrapper[4824]: I1211 02:01:19.518792 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 11 02:01:19 crc kubenswrapper[4824]: I1211 02:01:19.518873 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 11 02:01:19 crc kubenswrapper[4824]: I1211 02:01:19.531215 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 11 02:01:19 crc kubenswrapper[4824]: I1211 02:01:19.568591 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 11 02:01:19 crc kubenswrapper[4824]: I1211 02:01:19.574857 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 11 02:01:19 crc kubenswrapper[4824]: I1211 02:01:19.621029 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 11 02:01:19 crc kubenswrapper[4824]: I1211 02:01:19.631310 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 11 02:01:19 crc kubenswrapper[4824]: I1211 02:01:19.631683 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 02:01:19 crc kubenswrapper[4824]: I1211 02:01:19.631744 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 02:01:19 crc kubenswrapper[4824]: E1211 02:01:19.631802 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 02:01:19 crc kubenswrapper[4824]: E1211 02:01:19.631866 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 02:01:19 crc kubenswrapper[4824]: I1211 02:01:19.631911 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 02:01:19 crc kubenswrapper[4824]: E1211 02:01:19.632031 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 02:01:19 crc kubenswrapper[4824]: I1211 02:01:19.667168 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 11 02:01:19 crc kubenswrapper[4824]: I1211 02:01:19.676347 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 11 02:01:19 crc kubenswrapper[4824]: I1211 02:01:19.691620 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 11 02:01:19 crc kubenswrapper[4824]: I1211 02:01:19.714058 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 11 02:01:19 crc kubenswrapper[4824]: I1211 02:01:19.761963 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 11 02:01:19 crc kubenswrapper[4824]: I1211 02:01:19.800773 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 11 02:01:19 crc kubenswrapper[4824]: I1211 02:01:19.840551 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 11 02:01:19 crc kubenswrapper[4824]: I1211 02:01:19.851323 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" event={"ID":"5b193bd2-b536-4056-92f1-94c9836ab2eb","Type":"ContainerStarted","Data":"4ed2ae40b1fe28111efc76d9d72345edecb5f3ea34522b238f3730812b29f2bd"} Dec 11 02:01:19 crc kubenswrapper[4824]: I1211 02:01:19.851359 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" event={"ID":"5b193bd2-b536-4056-92f1-94c9836ab2eb","Type":"ContainerStarted","Data":"ce7aa078800e6f299cfe03770fa7e3b64fa3742fe0c0f383c87d84cf48569ddf"} Dec 11 02:01:19 crc kubenswrapper[4824]: I1211 02:01:19.851370 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" event={"ID":"5b193bd2-b536-4056-92f1-94c9836ab2eb","Type":"ContainerStarted","Data":"c345850891744a0240781277e9a89a59f6899c4c944995f62e44650f807440a1"} Dec 11 02:01:19 crc kubenswrapper[4824]: I1211 02:01:19.851382 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" event={"ID":"5b193bd2-b536-4056-92f1-94c9836ab2eb","Type":"ContainerStarted","Data":"51f3c41d86ddc44f7eca118029876ad353317f753d0ad173ed4d30c2a15f10e6"} Dec 11 02:01:19 crc kubenswrapper[4824]: I1211 02:01:19.851393 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" event={"ID":"5b193bd2-b536-4056-92f1-94c9836ab2eb","Type":"ContainerStarted","Data":"6aa9fc1f1bff192c69bd59d32a05d4a72d09641d9ce4b265bc4517147a7e15dd"} Dec 11 02:01:19 crc kubenswrapper[4824]: I1211 02:01:19.851408 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" event={"ID":"5b193bd2-b536-4056-92f1-94c9836ab2eb","Type":"ContainerStarted","Data":"4b4a3bc50eb80777188206adab774e8acad8b2e4952a8415f91b1c29ba9a0ffd"} Dec 11 02:01:19 crc kubenswrapper[4824]: I1211 02:01:19.853141 4824 generic.go:334] "Generic (PLEG): container finished" podID="50c69d41-a65b-463e-a829-ea349bd2fe46" containerID="fd9675d0afded587bc660a3862fd0c5dd6885961afacda957aa9ee8501d68a69" exitCode=0 Dec 11 02:01:19 crc kubenswrapper[4824]: I1211 02:01:19.853161 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-kh8g9" event={"ID":"50c69d41-a65b-463e-a829-ea349bd2fe46","Type":"ContainerDied","Data":"fd9675d0afded587bc660a3862fd0c5dd6885961afacda957aa9ee8501d68a69"} Dec 11 02:01:19 crc kubenswrapper[4824]: I1211 02:01:19.871358 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:19Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:19 crc kubenswrapper[4824]: I1211 02:01:19.872331 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 11 02:01:19 crc kubenswrapper[4824]: I1211 02:01:19.874995 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 11 02:01:19 crc kubenswrapper[4824]: I1211 02:01:19.888780 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ded146c045085be2b58821749052413a36850ede00ae1024fe2dae9fa5ac84b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:19Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:19 crc kubenswrapper[4824]: I1211 02:01:19.909459 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76b10bbd1e891b8c682f4b2e17d023fd787a8ed44ed94fbf7d97314d641ccf36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f32838f33832b9b67e7dd9c53340680e17721cf1c7bdf0170230c5f8feafe64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:19Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:19 crc kubenswrapper[4824]: I1211 02:01:19.924652 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kptws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68568dba-83a5-46c1-8d01-337359679d91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f965a6d6db619995132f61b5d51b66f0b2b3fdbef3fbec63e231b9a62ccd0a9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd7nf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kptws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:19Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:19 crc kubenswrapper[4824]: I1211 02:01:19.936338 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 11 02:01:19 crc kubenswrapper[4824]: I1211 02:01:19.939940 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"beab386b-6c65-4bc3-9ef0-dc3f68b4403f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c26f06d2ed5e6e98b61a51f073130675f2c49acda106b8aedcbb2a6d4f7580c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20af9b10a67c8b21a7abcd23e42bc412798d159c8422e7c8ff58f5b997ca19a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9bbfc6edfe050b3fa9dbeee5f98d416bcb2a1533a38608a881b75396fd0b059\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7c766f8de73ca762ffde0c18b2d1c46bb152b2b37535a611c5a0e454a21b7d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d45819f0611c91b20c83eeee1092031a3441fb69f3015241632ad419880227d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 02:01:11.070250 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 02:01:11.071673 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2752882565/tls.crt::/tmp/serving-cert-2752882565/tls.key\\\\\\\"\\\\nI1211 02:01:17.042610 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 02:01:17.048046 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 02:01:17.048080 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 02:01:17.048172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 02:01:17.048188 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 02:01:17.058248 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 02:01:17.058294 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 02:01:17.058303 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 02:01:17.058315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 02:01:17.058322 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 02:01:17.058328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 02:01:17.058335 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 02:01:17.058808 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 02:01:17.060225 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14a940a72b2404a6cbd52a69e150b03266bcd7f5c1d3f4d07b51c0c69e51786\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:00:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:19Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:19 crc kubenswrapper[4824]: I1211 02:01:19.955281 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6bkc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8022f696-d0e7-437b-848a-3cd25bd1f364\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3e28f6778ec0dba6ea3e6b69e5a2c1c73c0d5db1abeead241fb58efdbfaefe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hh5kn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6bkc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:19Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:19 crc kubenswrapper[4824]: I1211 02:01:19.956185 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 11 02:01:19 crc kubenswrapper[4824]: I1211 02:01:19.963365 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 11 02:01:19 crc kubenswrapper[4824]: I1211 02:01:19.966945 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 11 02:01:19 crc kubenswrapper[4824]: I1211 02:01:19.976547 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b193bd2-b536-4056-92f1-94c9836ab2eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f6pgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:19Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:19 crc kubenswrapper[4824]: I1211 02:01:19.988599 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-br4j9"] Dec 11 02:01:19 crc kubenswrapper[4824]: I1211 02:01:19.989063 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-br4j9" Dec 11 02:01:19 crc kubenswrapper[4824]: I1211 02:01:19.990303 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:19Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:19 crc kubenswrapper[4824]: I1211 02:01:19.992258 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 11 02:01:19 crc kubenswrapper[4824]: I1211 02:01:19.994424 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 11 02:01:19 crc kubenswrapper[4824]: I1211 02:01:19.994503 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 11 02:01:19 crc kubenswrapper[4824]: I1211 02:01:19.995241 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 11 02:01:19 crc kubenswrapper[4824]: I1211 02:01:19.995322 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.009238 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kh8g9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"50c69d41-a65b-463e-a829-ea349bd2fe46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69af1f33fd0bafc929b94e702b4f46c6c7d7a362d5552b121729c477892d91d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69af1f33fd0bafc929b94e702b4f46c6c7d7a362d5552b121729c477892d91d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd9675d0afded587bc660a3862fd0c5dd6885961afacda957aa9ee8501d68a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd9675d0afded587bc660a3862fd0c5dd6885961afacda957aa9ee8501d68a69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kh8g9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:20Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.009760 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.051465 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.070466 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.096653 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44b156e4-64a4-4d45-aa5e-9b10a862faed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7368d92144b188e95a7796fabf497b0f8d9c0b240293bdc7b4aec1ff75497bc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w62dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcf289ac27dbda4f24aabbde3fb5119329e7b4fc7d6e018168da9a69661a4c57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w62dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gx6xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:20Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.141879 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:20Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.170372 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ce125194-9c9d-4ea4-8a72-2c30d5722295-host\") pod \"node-ca-br4j9\" (UID: \"ce125194-9c9d-4ea4-8a72-2c30d5722295\") " pod="openshift-image-registry/node-ca-br4j9" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.170478 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/ce125194-9c9d-4ea4-8a72-2c30d5722295-serviceca\") pod \"node-ca-br4j9\" (UID: \"ce125194-9c9d-4ea4-8a72-2c30d5722295\") " pod="openshift-image-registry/node-ca-br4j9" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.170589 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fc5s8\" (UniqueName: \"kubernetes.io/projected/ce125194-9c9d-4ea4-8a72-2c30d5722295-kube-api-access-fc5s8\") pod \"node-ca-br4j9\" (UID: \"ce125194-9c9d-4ea4-8a72-2c30d5722295\") " pod="openshift-image-registry/node-ca-br4j9" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.179425 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:20Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.222045 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.224852 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.224888 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.224900 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.224964 4824 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.225187 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:20Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.271430 4824 kubelet_node_status.go:115] "Node was previously registered" node="crc" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.271838 4824 kubelet_node_status.go:79] "Successfully registered node" node="crc" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.272219 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ce125194-9c9d-4ea4-8a72-2c30d5722295-host\") pod \"node-ca-br4j9\" (UID: \"ce125194-9c9d-4ea4-8a72-2c30d5722295\") " pod="openshift-image-registry/node-ca-br4j9" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.272320 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/ce125194-9c9d-4ea4-8a72-2c30d5722295-serviceca\") pod \"node-ca-br4j9\" (UID: \"ce125194-9c9d-4ea4-8a72-2c30d5722295\") " pod="openshift-image-registry/node-ca-br4j9" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.272362 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fc5s8\" (UniqueName: \"kubernetes.io/projected/ce125194-9c9d-4ea4-8a72-2c30d5722295-kube-api-access-fc5s8\") pod \"node-ca-br4j9\" (UID: \"ce125194-9c9d-4ea4-8a72-2c30d5722295\") " pod="openshift-image-registry/node-ca-br4j9" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.272362 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ce125194-9c9d-4ea4-8a72-2c30d5722295-host\") pod \"node-ca-br4j9\" (UID: \"ce125194-9c9d-4ea4-8a72-2c30d5722295\") " pod="openshift-image-registry/node-ca-br4j9" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.273707 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.273756 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.273773 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.273792 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.273809 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:20Z","lastTransitionTime":"2025-12-11T02:01:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.274626 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/ce125194-9c9d-4ea4-8a72-2c30d5722295-serviceca\") pod \"node-ca-br4j9\" (UID: \"ce125194-9c9d-4ea4-8a72-2c30d5722295\") " pod="openshift-image-registry/node-ca-br4j9" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.306545 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ded146c045085be2b58821749052413a36850ede00ae1024fe2dae9fa5ac84b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:20Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:20 crc kubenswrapper[4824]: E1211 02:01:20.307587 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d57bece6-6fc6-4a84-b8a8-e1678f9e50db\\\",\\\"systemUUID\\\":\\\"f76b32e4-2764-4219-bbf4-f7e56d76df3b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:20Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.313026 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.313080 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.313097 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.313149 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.313169 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:20Z","lastTransitionTime":"2025-12-11T02:01:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:20 crc kubenswrapper[4824]: E1211 02:01:20.336810 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d57bece6-6fc6-4a84-b8a8-e1678f9e50db\\\",\\\"systemUUID\\\":\\\"f76b32e4-2764-4219-bbf4-f7e56d76df3b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:20Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.342430 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.342499 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.342522 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.342553 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.342575 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:20Z","lastTransitionTime":"2025-12-11T02:01:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.347995 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fc5s8\" (UniqueName: \"kubernetes.io/projected/ce125194-9c9d-4ea4-8a72-2c30d5722295-kube-api-access-fc5s8\") pod \"node-ca-br4j9\" (UID: \"ce125194-9c9d-4ea4-8a72-2c30d5722295\") " pod="openshift-image-registry/node-ca-br4j9" Dec 11 02:01:20 crc kubenswrapper[4824]: E1211 02:01:20.360906 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d57bece6-6fc6-4a84-b8a8-e1678f9e50db\\\",\\\"systemUUID\\\":\\\"f76b32e4-2764-4219-bbf4-f7e56d76df3b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:20Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.364916 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76b10bbd1e891b8c682f4b2e17d023fd787a8ed44ed94fbf7d97314d641ccf36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f32838f33832b9b67e7dd9c53340680e17721cf1c7bdf0170230c5f8feafe64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:20Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.365565 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.365731 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.365861 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.365995 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.366195 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:20Z","lastTransitionTime":"2025-12-11T02:01:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:20 crc kubenswrapper[4824]: E1211 02:01:20.381671 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d57bece6-6fc6-4a84-b8a8-e1678f9e50db\\\",\\\"systemUUID\\\":\\\"f76b32e4-2764-4219-bbf4-f7e56d76df3b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:20Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.386133 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.386179 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.386191 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.386211 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.386222 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:20Z","lastTransitionTime":"2025-12-11T02:01:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.397661 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kptws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68568dba-83a5-46c1-8d01-337359679d91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f965a6d6db619995132f61b5d51b66f0b2b3fdbef3fbec63e231b9a62ccd0a9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd7nf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kptws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:20Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:20 crc kubenswrapper[4824]: E1211 02:01:20.399978 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d57bece6-6fc6-4a84-b8a8-e1678f9e50db\\\",\\\"systemUUID\\\":\\\"f76b32e4-2764-4219-bbf4-f7e56d76df3b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:20Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:20 crc kubenswrapper[4824]: E1211 02:01:20.400249 4824 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.401813 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.401883 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.401903 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.401928 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.401949 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:20Z","lastTransitionTime":"2025-12-11T02:01:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.444172 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"beab386b-6c65-4bc3-9ef0-dc3f68b4403f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c26f06d2ed5e6e98b61a51f073130675f2c49acda106b8aedcbb2a6d4f7580c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20af9b10a67c8b21a7abcd23e42bc412798d159c8422e7c8ff58f5b997ca19a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9bbfc6edfe050b3fa9dbeee5f98d416bcb2a1533a38608a881b75396fd0b059\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7c766f8de73ca762ffde0c18b2d1c46bb152b2b37535a611c5a0e454a21b7d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d45819f0611c91b20c83eeee1092031a3441fb69f3015241632ad419880227d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 02:01:11.070250 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 02:01:11.071673 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2752882565/tls.crt::/tmp/serving-cert-2752882565/tls.key\\\\\\\"\\\\nI1211 02:01:17.042610 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 02:01:17.048046 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 02:01:17.048080 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 02:01:17.048172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 02:01:17.048188 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 02:01:17.058248 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 02:01:17.058294 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 02:01:17.058303 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 02:01:17.058315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 02:01:17.058322 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 02:01:17.058328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 02:01:17.058335 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 02:01:17.058808 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 02:01:17.060225 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14a940a72b2404a6cbd52a69e150b03266bcd7f5c1d3f4d07b51c0c69e51786\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:00:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:20Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.484476 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6bkc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8022f696-d0e7-437b-848a-3cd25bd1f364\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3e28f6778ec0dba6ea3e6b69e5a2c1c73c0d5db1abeead241fb58efdbfaefe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hh5kn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6bkc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:20Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.504525 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.504591 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.504610 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.504639 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.504659 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:20Z","lastTransitionTime":"2025-12-11T02:01:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.526815 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b193bd2-b536-4056-92f1-94c9836ab2eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f6pgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:20Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.535891 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.539912 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.559676 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-br4j9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce125194-9c9d-4ea4-8a72-2c30d5722295\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:19Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:19Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fc5s8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-br4j9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:20Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.578761 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.606804 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-br4j9" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.607686 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.607876 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.607954 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.607988 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.608009 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:20Z","lastTransitionTime":"2025-12-11T02:01:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.627847 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:20Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:20 crc kubenswrapper[4824]: W1211 02:01:20.663927 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podce125194_9c9d_4ea4_8a72_2c30d5722295.slice/crio-ea4ea24783b38c70da0881498fd6fb6e63fa4883ded3766993846e6c2e30ae40 WatchSource:0}: Error finding container ea4ea24783b38c70da0881498fd6fb6e63fa4883ded3766993846e6c2e30ae40: Status 404 returned error can't find the container with id ea4ea24783b38c70da0881498fd6fb6e63fa4883ded3766993846e6c2e30ae40 Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.666781 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kh8g9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"50c69d41-a65b-463e-a829-ea349bd2fe46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69af1f33fd0bafc929b94e702b4f46c6c7d7a362d5552b121729c477892d91d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69af1f33fd0bafc929b94e702b4f46c6c7d7a362d5552b121729c477892d91d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd9675d0afded587bc660a3862fd0c5dd6885961afacda957aa9ee8501d68a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd9675d0afded587bc660a3862fd0c5dd6885961afacda957aa9ee8501d68a69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kh8g9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:20Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.702872 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44b156e4-64a4-4d45-aa5e-9b10a862faed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7368d92144b188e95a7796fabf497b0f8d9c0b240293bdc7b4aec1ff75497bc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w62dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcf289ac27dbda4f24aabbde3fb5119329e7b4fc7d6e018168da9a69661a4c57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w62dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gx6xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:20Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.712293 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.712363 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.712383 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.712416 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.712437 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:20Z","lastTransitionTime":"2025-12-11T02:01:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.742885 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:20Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.786890 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:20Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.815642 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.815687 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.815698 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.815719 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.815732 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:20Z","lastTransitionTime":"2025-12-11T02:01:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.827738 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:20Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.858358 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-br4j9" event={"ID":"ce125194-9c9d-4ea4-8a72-2c30d5722295","Type":"ContainerStarted","Data":"ea4ea24783b38c70da0881498fd6fb6e63fa4883ded3766993846e6c2e30ae40"} Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.864697 4824 generic.go:334] "Generic (PLEG): container finished" podID="50c69d41-a65b-463e-a829-ea349bd2fe46" containerID="aac94149458bd6321be90b26b48ee26428d273e7f61407d95f9c9d6b3aeae4b6" exitCode=0 Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.864724 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:20Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.864772 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-kh8g9" event={"ID":"50c69d41-a65b-463e-a829-ea349bd2fe46","Type":"ContainerDied","Data":"aac94149458bd6321be90b26b48ee26428d273e7f61407d95f9c9d6b3aeae4b6"} Dec 11 02:01:20 crc kubenswrapper[4824]: E1211 02:01:20.899737 4824 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-crc\" already exists" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.921373 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.921441 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.921460 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.921485 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.921501 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:20Z","lastTransitionTime":"2025-12-11T02:01:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.925518 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44b156e4-64a4-4d45-aa5e-9b10a862faed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7368d92144b188e95a7796fabf497b0f8d9c0b240293bdc7b4aec1ff75497bc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w62dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcf289ac27dbda4f24aabbde3fb5119329e7b4fc7d6e018168da9a69661a4c57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w62dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gx6xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:20Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:20 crc kubenswrapper[4824]: I1211 02:01:20.962847 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"253d4291-2b2f-487c-8862-3d305f0bdc55\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44b29bb7a6eda7cb890d36f86375b5f3f197ae5cf390f16e71139b0904873e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://922af7f67b99a0ce08776f3f75c0c377e6ea69af1720b6178a6d1e2f9b4b8ba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://136c68020325db8e6930d0f40b0d0d7e013c14c4dd5e7d1449d01d6d458104d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c89afaf56b4a9597f6543ec2e3b176c7a6dab7dc341941fc3bc8ec85cc4edaf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:00:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:20Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.005547 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:21Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.026254 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.026319 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.026341 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.026368 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.026389 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:21Z","lastTransitionTime":"2025-12-11T02:01:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.042812 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kptws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68568dba-83a5-46c1-8d01-337359679d91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f965a6d6db619995132f61b5d51b66f0b2b3fdbef3fbec63e231b9a62ccd0a9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd7nf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kptws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:21Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.085210 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"beab386b-6c65-4bc3-9ef0-dc3f68b4403f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c26f06d2ed5e6e98b61a51f073130675f2c49acda106b8aedcbb2a6d4f7580c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20af9b10a67c8b21a7abcd23e42bc412798d159c8422e7c8ff58f5b997ca19a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9bbfc6edfe050b3fa9dbeee5f98d416bcb2a1533a38608a881b75396fd0b059\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7c766f8de73ca762ffde0c18b2d1c46bb152b2b37535a611c5a0e454a21b7d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d45819f0611c91b20c83eeee1092031a3441fb69f3015241632ad419880227d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 02:01:11.070250 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 02:01:11.071673 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2752882565/tls.crt::/tmp/serving-cert-2752882565/tls.key\\\\\\\"\\\\nI1211 02:01:17.042610 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 02:01:17.048046 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 02:01:17.048080 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 02:01:17.048172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 02:01:17.048188 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 02:01:17.058248 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 02:01:17.058294 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 02:01:17.058303 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 02:01:17.058315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 02:01:17.058322 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 02:01:17.058328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 02:01:17.058335 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 02:01:17.058808 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 02:01:17.060225 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14a940a72b2404a6cbd52a69e150b03266bcd7f5c1d3f4d07b51c0c69e51786\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:00:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:21Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.122828 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ded146c045085be2b58821749052413a36850ede00ae1024fe2dae9fa5ac84b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:21Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.128760 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.128814 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.128826 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.128848 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.128861 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:21Z","lastTransitionTime":"2025-12-11T02:01:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.163072 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76b10bbd1e891b8c682f4b2e17d023fd787a8ed44ed94fbf7d97314d641ccf36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f32838f33832b9b67e7dd9c53340680e17721cf1c7bdf0170230c5f8feafe64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:21Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.199917 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-br4j9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce125194-9c9d-4ea4-8a72-2c30d5722295\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:19Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:19Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fc5s8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-br4j9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:21Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.232370 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.232437 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.232459 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.232494 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.232542 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:21Z","lastTransitionTime":"2025-12-11T02:01:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.243102 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:21Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.283544 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 02:01:21 crc kubenswrapper[4824]: E1211 02:01:21.283750 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 02:01:25.283707213 +0000 UTC m=+26.972744632 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.289401 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kh8g9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"50c69d41-a65b-463e-a829-ea349bd2fe46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69af1f33fd0bafc929b94e702b4f46c6c7d7a362d5552b121729c477892d91d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69af1f33fd0bafc929b94e702b4f46c6c7d7a362d5552b121729c477892d91d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd9675d0afded587bc660a3862fd0c5dd6885961afacda957aa9ee8501d68a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd9675d0afded587bc660a3862fd0c5dd6885961afacda957aa9ee8501d68a69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kh8g9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:21Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.325310 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6bkc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8022f696-d0e7-437b-848a-3cd25bd1f364\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3e28f6778ec0dba6ea3e6b69e5a2c1c73c0d5db1abeead241fb58efdbfaefe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hh5kn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6bkc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:21Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.335826 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.335916 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.335929 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.335954 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.335971 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:21Z","lastTransitionTime":"2025-12-11T02:01:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.385637 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.385737 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.385821 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 02:01:21 crc kubenswrapper[4824]: E1211 02:01:21.386013 4824 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.386076 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 02:01:21 crc kubenswrapper[4824]: E1211 02:01:21.386143 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-11 02:01:25.386093998 +0000 UTC m=+27.075131367 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 11 02:01:21 crc kubenswrapper[4824]: E1211 02:01:21.386208 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 11 02:01:21 crc kubenswrapper[4824]: E1211 02:01:21.386255 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 11 02:01:21 crc kubenswrapper[4824]: E1211 02:01:21.386276 4824 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 02:01:21 crc kubenswrapper[4824]: E1211 02:01:21.386317 4824 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 11 02:01:21 crc kubenswrapper[4824]: E1211 02:01:21.386366 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-11 02:01:25.386338834 +0000 UTC m=+27.075376253 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 02:01:21 crc kubenswrapper[4824]: E1211 02:01:21.386383 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 11 02:01:21 crc kubenswrapper[4824]: E1211 02:01:21.386451 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 11 02:01:21 crc kubenswrapper[4824]: E1211 02:01:21.386473 4824 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 02:01:21 crc kubenswrapper[4824]: E1211 02:01:21.386415 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-11 02:01:25.386386406 +0000 UTC m=+27.075423945 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 11 02:01:21 crc kubenswrapper[4824]: E1211 02:01:21.386590 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-11 02:01:25.3865624 +0000 UTC m=+27.075599819 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.399961 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b193bd2-b536-4056-92f1-94c9836ab2eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f6pgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:21Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.427942 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:21Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.438665 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.438712 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.438724 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.438744 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.438756 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:21Z","lastTransitionTime":"2025-12-11T02:01:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.448766 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:21Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.477998 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44b156e4-64a4-4d45-aa5e-9b10a862faed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7368d92144b188e95a7796fabf497b0f8d9c0b240293bdc7b4aec1ff75497bc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w62dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcf289ac27dbda4f24aabbde3fb5119329e7b4fc7d6e018168da9a69661a4c57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w62dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gx6xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:21Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.517742 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"253d4291-2b2f-487c-8862-3d305f0bdc55\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44b29bb7a6eda7cb890d36f86375b5f3f197ae5cf390f16e71139b0904873e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://922af7f67b99a0ce08776f3f75c0c377e6ea69af1720b6178a6d1e2f9b4b8ba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://136c68020325db8e6930d0f40b0d0d7e013c14c4dd5e7d1449d01d6d458104d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c89afaf56b4a9597f6543ec2e3b176c7a6dab7dc341941fc3bc8ec85cc4edaf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:00:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:21Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.540872 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.540928 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.540944 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.540972 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.540990 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:21Z","lastTransitionTime":"2025-12-11T02:01:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.558847 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:21Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.602404 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kptws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68568dba-83a5-46c1-8d01-337359679d91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f965a6d6db619995132f61b5d51b66f0b2b3fdbef3fbec63e231b9a62ccd0a9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd7nf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kptws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:21Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.631738 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.631842 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.631738 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 02:01:21 crc kubenswrapper[4824]: E1211 02:01:21.631993 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 02:01:21 crc kubenswrapper[4824]: E1211 02:01:21.632126 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 02:01:21 crc kubenswrapper[4824]: E1211 02:01:21.632291 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.644213 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.644278 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.644303 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.644332 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.644356 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:21Z","lastTransitionTime":"2025-12-11T02:01:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.650255 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"beab386b-6c65-4bc3-9ef0-dc3f68b4403f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c26f06d2ed5e6e98b61a51f073130675f2c49acda106b8aedcbb2a6d4f7580c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20af9b10a67c8b21a7abcd23e42bc412798d159c8422e7c8ff58f5b997ca19a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9bbfc6edfe050b3fa9dbeee5f98d416bcb2a1533a38608a881b75396fd0b059\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7c766f8de73ca762ffde0c18b2d1c46bb152b2b37535a611c5a0e454a21b7d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d45819f0611c91b20c83eeee1092031a3441fb69f3015241632ad419880227d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 02:01:11.070250 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 02:01:11.071673 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2752882565/tls.crt::/tmp/serving-cert-2752882565/tls.key\\\\\\\"\\\\nI1211 02:01:17.042610 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 02:01:17.048046 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 02:01:17.048080 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 02:01:17.048172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 02:01:17.048188 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 02:01:17.058248 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 02:01:17.058294 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 02:01:17.058303 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 02:01:17.058315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 02:01:17.058322 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 02:01:17.058328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 02:01:17.058335 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 02:01:17.058808 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 02:01:17.060225 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14a940a72b2404a6cbd52a69e150b03266bcd7f5c1d3f4d07b51c0c69e51786\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:00:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:21Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.686239 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ded146c045085be2b58821749052413a36850ede00ae1024fe2dae9fa5ac84b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:21Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.725764 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76b10bbd1e891b8c682f4b2e17d023fd787a8ed44ed94fbf7d97314d641ccf36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f32838f33832b9b67e7dd9c53340680e17721cf1c7bdf0170230c5f8feafe64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:21Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.746931 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.746983 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.746997 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.747022 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.747036 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:21Z","lastTransitionTime":"2025-12-11T02:01:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.761807 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-br4j9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce125194-9c9d-4ea4-8a72-2c30d5722295\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:19Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:19Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fc5s8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-br4j9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:21Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.810141 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:21Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.849140 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kh8g9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"50c69d41-a65b-463e-a829-ea349bd2fe46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69af1f33fd0bafc929b94e702b4f46c6c7d7a362d5552b121729c477892d91d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69af1f33fd0bafc929b94e702b4f46c6c7d7a362d5552b121729c477892d91d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd9675d0afded587bc660a3862fd0c5dd6885961afacda957aa9ee8501d68a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd9675d0afded587bc660a3862fd0c5dd6885961afacda957aa9ee8501d68a69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aac94149458bd6321be90b26b48ee26428d273e7f61407d95f9c9d6b3aeae4b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aac94149458bd6321be90b26b48ee26428d273e7f61407d95f9c9d6b3aeae4b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kh8g9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:21Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.850659 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.850711 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.850730 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.850755 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.850774 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:21Z","lastTransitionTime":"2025-12-11T02:01:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.875929 4824 generic.go:334] "Generic (PLEG): container finished" podID="50c69d41-a65b-463e-a829-ea349bd2fe46" containerID="dbee11bd7cb65e031415990869b52d2b8f81bf8b56e6cc935d38e09695db7b48" exitCode=0 Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.876076 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-kh8g9" event={"ID":"50c69d41-a65b-463e-a829-ea349bd2fe46","Type":"ContainerDied","Data":"dbee11bd7cb65e031415990869b52d2b8f81bf8b56e6cc935d38e09695db7b48"} Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.884944 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6bkc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8022f696-d0e7-437b-848a-3cd25bd1f364\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3e28f6778ec0dba6ea3e6b69e5a2c1c73c0d5db1abeead241fb58efdbfaefe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hh5kn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6bkc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:21Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.890481 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-br4j9" event={"ID":"ce125194-9c9d-4ea4-8a72-2c30d5722295","Type":"ContainerStarted","Data":"9798f904052e08196a0db1745060dab11e2710308bca01ff160475910f2665df"} Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.897955 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"d14986666febd651039afb52ddf4db8969aca55e88eda9bbb20a67e35ba722ab"} Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.929293 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b193bd2-b536-4056-92f1-94c9836ab2eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f6pgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:21Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.954877 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.954927 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.954945 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.954966 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.954977 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:21Z","lastTransitionTime":"2025-12-11T02:01:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:21 crc kubenswrapper[4824]: I1211 02:01:21.965847 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d14986666febd651039afb52ddf4db8969aca55e88eda9bbb20a67e35ba722ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:21Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:22 crc kubenswrapper[4824]: I1211 02:01:22.003474 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:22Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:22 crc kubenswrapper[4824]: I1211 02:01:22.041556 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44b156e4-64a4-4d45-aa5e-9b10a862faed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7368d92144b188e95a7796fabf497b0f8d9c0b240293bdc7b4aec1ff75497bc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w62dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcf289ac27dbda4f24aabbde3fb5119329e7b4fc7d6e018168da9a69661a4c57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w62dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gx6xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:22Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:22 crc kubenswrapper[4824]: I1211 02:01:22.059818 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:22 crc kubenswrapper[4824]: I1211 02:01:22.059870 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:22 crc kubenswrapper[4824]: I1211 02:01:22.059883 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:22 crc kubenswrapper[4824]: I1211 02:01:22.059904 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:22 crc kubenswrapper[4824]: I1211 02:01:22.059916 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:22Z","lastTransitionTime":"2025-12-11T02:01:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:22 crc kubenswrapper[4824]: I1211 02:01:22.082763 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"253d4291-2b2f-487c-8862-3d305f0bdc55\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44b29bb7a6eda7cb890d36f86375b5f3f197ae5cf390f16e71139b0904873e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://922af7f67b99a0ce08776f3f75c0c377e6ea69af1720b6178a6d1e2f9b4b8ba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://136c68020325db8e6930d0f40b0d0d7e013c14c4dd5e7d1449d01d6d458104d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c89afaf56b4a9597f6543ec2e3b176c7a6dab7dc341941fc3bc8ec85cc4edaf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:00:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:22Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:22 crc kubenswrapper[4824]: I1211 02:01:22.122757 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:22Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:22 crc kubenswrapper[4824]: I1211 02:01:22.163317 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:22 crc kubenswrapper[4824]: I1211 02:01:22.163371 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:22 crc kubenswrapper[4824]: I1211 02:01:22.163384 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:22 crc kubenswrapper[4824]: I1211 02:01:22.163404 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:22 crc kubenswrapper[4824]: I1211 02:01:22.163416 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:22Z","lastTransitionTime":"2025-12-11T02:01:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:22 crc kubenswrapper[4824]: I1211 02:01:22.167683 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"beab386b-6c65-4bc3-9ef0-dc3f68b4403f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c26f06d2ed5e6e98b61a51f073130675f2c49acda106b8aedcbb2a6d4f7580c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20af9b10a67c8b21a7abcd23e42bc412798d159c8422e7c8ff58f5b997ca19a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9bbfc6edfe050b3fa9dbeee5f98d416bcb2a1533a38608a881b75396fd0b059\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7c766f8de73ca762ffde0c18b2d1c46bb152b2b37535a611c5a0e454a21b7d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d45819f0611c91b20c83eeee1092031a3441fb69f3015241632ad419880227d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 02:01:11.070250 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 02:01:11.071673 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2752882565/tls.crt::/tmp/serving-cert-2752882565/tls.key\\\\\\\"\\\\nI1211 02:01:17.042610 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 02:01:17.048046 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 02:01:17.048080 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 02:01:17.048172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 02:01:17.048188 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 02:01:17.058248 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 02:01:17.058294 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 02:01:17.058303 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 02:01:17.058315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 02:01:17.058322 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 02:01:17.058328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 02:01:17.058335 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 02:01:17.058808 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 02:01:17.060225 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14a940a72b2404a6cbd52a69e150b03266bcd7f5c1d3f4d07b51c0c69e51786\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:00:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:22Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:22 crc kubenswrapper[4824]: I1211 02:01:22.199475 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ded146c045085be2b58821749052413a36850ede00ae1024fe2dae9fa5ac84b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:22Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:22 crc kubenswrapper[4824]: I1211 02:01:22.245823 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76b10bbd1e891b8c682f4b2e17d023fd787a8ed44ed94fbf7d97314d641ccf36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f32838f33832b9b67e7dd9c53340680e17721cf1c7bdf0170230c5f8feafe64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:22Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:22 crc kubenswrapper[4824]: I1211 02:01:22.266883 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:22 crc kubenswrapper[4824]: I1211 02:01:22.266949 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:22 crc kubenswrapper[4824]: I1211 02:01:22.266969 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:22 crc kubenswrapper[4824]: I1211 02:01:22.266995 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:22 crc kubenswrapper[4824]: I1211 02:01:22.267013 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:22Z","lastTransitionTime":"2025-12-11T02:01:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:22 crc kubenswrapper[4824]: I1211 02:01:22.282094 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kptws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68568dba-83a5-46c1-8d01-337359679d91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f965a6d6db619995132f61b5d51b66f0b2b3fdbef3fbec63e231b9a62ccd0a9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd7nf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kptws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:22Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:22 crc kubenswrapper[4824]: I1211 02:01:22.325627 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:22Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:22 crc kubenswrapper[4824]: I1211 02:01:22.371030 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:22 crc kubenswrapper[4824]: I1211 02:01:22.371085 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:22 crc kubenswrapper[4824]: I1211 02:01:22.371102 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:22 crc kubenswrapper[4824]: I1211 02:01:22.371183 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:22 crc kubenswrapper[4824]: I1211 02:01:22.371206 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:22Z","lastTransitionTime":"2025-12-11T02:01:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:22 crc kubenswrapper[4824]: I1211 02:01:22.372281 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kh8g9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"50c69d41-a65b-463e-a829-ea349bd2fe46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69af1f33fd0bafc929b94e702b4f46c6c7d7a362d5552b121729c477892d91d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69af1f33fd0bafc929b94e702b4f46c6c7d7a362d5552b121729c477892d91d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd9675d0afded587bc660a3862fd0c5dd6885961afacda957aa9ee8501d68a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd9675d0afded587bc660a3862fd0c5dd6885961afacda957aa9ee8501d68a69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aac94149458bd6321be90b26b48ee26428d273e7f61407d95f9c9d6b3aeae4b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aac94149458bd6321be90b26b48ee26428d273e7f61407d95f9c9d6b3aeae4b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbee11bd7cb65e031415990869b52d2b8f81bf8b56e6cc935d38e09695db7b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbee11bd7cb65e031415990869b52d2b8f81bf8b56e6cc935d38e09695db7b48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kh8g9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:22Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:22 crc kubenswrapper[4824]: I1211 02:01:22.406577 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6bkc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8022f696-d0e7-437b-848a-3cd25bd1f364\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3e28f6778ec0dba6ea3e6b69e5a2c1c73c0d5db1abeead241fb58efdbfaefe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hh5kn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6bkc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:22Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:22 crc kubenswrapper[4824]: I1211 02:01:22.457393 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b193bd2-b536-4056-92f1-94c9836ab2eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f6pgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:22Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:22 crc kubenswrapper[4824]: I1211 02:01:22.473682 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:22 crc kubenswrapper[4824]: I1211 02:01:22.473731 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:22 crc kubenswrapper[4824]: I1211 02:01:22.473747 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:22 crc kubenswrapper[4824]: I1211 02:01:22.473771 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:22 crc kubenswrapper[4824]: I1211 02:01:22.473788 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:22Z","lastTransitionTime":"2025-12-11T02:01:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:22 crc kubenswrapper[4824]: I1211 02:01:22.487155 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-br4j9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce125194-9c9d-4ea4-8a72-2c30d5722295\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9798f904052e08196a0db1745060dab11e2710308bca01ff160475910f2665df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fc5s8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-br4j9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:22Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:22 crc kubenswrapper[4824]: I1211 02:01:22.577904 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:22 crc kubenswrapper[4824]: I1211 02:01:22.577961 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:22 crc kubenswrapper[4824]: I1211 02:01:22.577981 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:22 crc kubenswrapper[4824]: I1211 02:01:22.578007 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:22 crc kubenswrapper[4824]: I1211 02:01:22.578026 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:22Z","lastTransitionTime":"2025-12-11T02:01:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:22 crc kubenswrapper[4824]: I1211 02:01:22.680890 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:22 crc kubenswrapper[4824]: I1211 02:01:22.680952 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:22 crc kubenswrapper[4824]: I1211 02:01:22.681470 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:22 crc kubenswrapper[4824]: I1211 02:01:22.681569 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:22 crc kubenswrapper[4824]: I1211 02:01:22.681625 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:22Z","lastTransitionTime":"2025-12-11T02:01:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:22 crc kubenswrapper[4824]: I1211 02:01:22.784628 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:22 crc kubenswrapper[4824]: I1211 02:01:22.784685 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:22 crc kubenswrapper[4824]: I1211 02:01:22.784703 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:22 crc kubenswrapper[4824]: I1211 02:01:22.784727 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:22 crc kubenswrapper[4824]: I1211 02:01:22.784746 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:22Z","lastTransitionTime":"2025-12-11T02:01:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:22 crc kubenswrapper[4824]: I1211 02:01:22.888713 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:22 crc kubenswrapper[4824]: I1211 02:01:22.888774 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:22 crc kubenswrapper[4824]: I1211 02:01:22.888791 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:22 crc kubenswrapper[4824]: I1211 02:01:22.888814 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:22 crc kubenswrapper[4824]: I1211 02:01:22.888832 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:22Z","lastTransitionTime":"2025-12-11T02:01:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:22 crc kubenswrapper[4824]: I1211 02:01:22.908150 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" event={"ID":"5b193bd2-b536-4056-92f1-94c9836ab2eb","Type":"ContainerStarted","Data":"9ead58694813806d307bbbc6a7b3cf41730d1c059b231a3002684bf09c10c7f4"} Dec 11 02:01:22 crc kubenswrapper[4824]: I1211 02:01:22.912824 4824 generic.go:334] "Generic (PLEG): container finished" podID="50c69d41-a65b-463e-a829-ea349bd2fe46" containerID="86f693ade72bd593bc16021fe4b7564f1914ec2f1cda7039474ac2fb13de10f5" exitCode=0 Dec 11 02:01:22 crc kubenswrapper[4824]: I1211 02:01:22.912926 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-kh8g9" event={"ID":"50c69d41-a65b-463e-a829-ea349bd2fe46","Type":"ContainerDied","Data":"86f693ade72bd593bc16021fe4b7564f1914ec2f1cda7039474ac2fb13de10f5"} Dec 11 02:01:22 crc kubenswrapper[4824]: I1211 02:01:22.936029 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:22Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:22 crc kubenswrapper[4824]: I1211 02:01:22.955202 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"253d4291-2b2f-487c-8862-3d305f0bdc55\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44b29bb7a6eda7cb890d36f86375b5f3f197ae5cf390f16e71139b0904873e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://922af7f67b99a0ce08776f3f75c0c377e6ea69af1720b6178a6d1e2f9b4b8ba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://136c68020325db8e6930d0f40b0d0d7e013c14c4dd5e7d1449d01d6d458104d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c89afaf56b4a9597f6543ec2e3b176c7a6dab7dc341941fc3bc8ec85cc4edaf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:00:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:22Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:22 crc kubenswrapper[4824]: I1211 02:01:22.970294 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76b10bbd1e891b8c682f4b2e17d023fd787a8ed44ed94fbf7d97314d641ccf36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f32838f33832b9b67e7dd9c53340680e17721cf1c7bdf0170230c5f8feafe64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:22Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:22 crc kubenswrapper[4824]: I1211 02:01:22.984642 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kptws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68568dba-83a5-46c1-8d01-337359679d91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f965a6d6db619995132f61b5d51b66f0b2b3fdbef3fbec63e231b9a62ccd0a9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd7nf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kptws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:22Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:22 crc kubenswrapper[4824]: I1211 02:01:22.998696 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:22 crc kubenswrapper[4824]: I1211 02:01:22.998756 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:22 crc kubenswrapper[4824]: I1211 02:01:22.998775 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:22 crc kubenswrapper[4824]: I1211 02:01:22.998798 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:22 crc kubenswrapper[4824]: I1211 02:01:22.998818 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:22Z","lastTransitionTime":"2025-12-11T02:01:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:23 crc kubenswrapper[4824]: I1211 02:01:23.006146 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"beab386b-6c65-4bc3-9ef0-dc3f68b4403f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c26f06d2ed5e6e98b61a51f073130675f2c49acda106b8aedcbb2a6d4f7580c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20af9b10a67c8b21a7abcd23e42bc412798d159c8422e7c8ff58f5b997ca19a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9bbfc6edfe050b3fa9dbeee5f98d416bcb2a1533a38608a881b75396fd0b059\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7c766f8de73ca762ffde0c18b2d1c46bb152b2b37535a611c5a0e454a21b7d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d45819f0611c91b20c83eeee1092031a3441fb69f3015241632ad419880227d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 02:01:11.070250 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 02:01:11.071673 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2752882565/tls.crt::/tmp/serving-cert-2752882565/tls.key\\\\\\\"\\\\nI1211 02:01:17.042610 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 02:01:17.048046 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 02:01:17.048080 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 02:01:17.048172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 02:01:17.048188 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 02:01:17.058248 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 02:01:17.058294 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 02:01:17.058303 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 02:01:17.058315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 02:01:17.058322 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 02:01:17.058328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 02:01:17.058335 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 02:01:17.058808 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 02:01:17.060225 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14a940a72b2404a6cbd52a69e150b03266bcd7f5c1d3f4d07b51c0c69e51786\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:00:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:23Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:23 crc kubenswrapper[4824]: I1211 02:01:23.025152 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ded146c045085be2b58821749052413a36850ede00ae1024fe2dae9fa5ac84b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:23Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:23 crc kubenswrapper[4824]: I1211 02:01:23.058059 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b193bd2-b536-4056-92f1-94c9836ab2eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f6pgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:23Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:23 crc kubenswrapper[4824]: I1211 02:01:23.075209 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-br4j9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce125194-9c9d-4ea4-8a72-2c30d5722295\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9798f904052e08196a0db1745060dab11e2710308bca01ff160475910f2665df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fc5s8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-br4j9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:23Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:23 crc kubenswrapper[4824]: I1211 02:01:23.093804 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:23Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:23 crc kubenswrapper[4824]: I1211 02:01:23.101558 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:23 crc kubenswrapper[4824]: I1211 02:01:23.101612 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:23 crc kubenswrapper[4824]: I1211 02:01:23.101629 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:23 crc kubenswrapper[4824]: I1211 02:01:23.101651 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:23 crc kubenswrapper[4824]: I1211 02:01:23.101666 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:23Z","lastTransitionTime":"2025-12-11T02:01:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:23 crc kubenswrapper[4824]: I1211 02:01:23.111016 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kh8g9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"50c69d41-a65b-463e-a829-ea349bd2fe46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69af1f33fd0bafc929b94e702b4f46c6c7d7a362d5552b121729c477892d91d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69af1f33fd0bafc929b94e702b4f46c6c7d7a362d5552b121729c477892d91d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd9675d0afded587bc660a3862fd0c5dd6885961afacda957aa9ee8501d68a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd9675d0afded587bc660a3862fd0c5dd6885961afacda957aa9ee8501d68a69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aac94149458bd6321be90b26b48ee26428d273e7f61407d95f9c9d6b3aeae4b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aac94149458bd6321be90b26b48ee26428d273e7f61407d95f9c9d6b3aeae4b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbee11bd7cb65e031415990869b52d2b8f81bf8b56e6cc935d38e09695db7b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbee11bd7cb65e031415990869b52d2b8f81bf8b56e6cc935d38e09695db7b48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86f693ade72bd593bc16021fe4b7564f1914ec2f1cda7039474ac2fb13de10f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86f693ade72bd593bc16021fe4b7564f1914ec2f1cda7039474ac2fb13de10f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kh8g9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:23Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:23 crc kubenswrapper[4824]: I1211 02:01:23.129537 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6bkc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8022f696-d0e7-437b-848a-3cd25bd1f364\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3e28f6778ec0dba6ea3e6b69e5a2c1c73c0d5db1abeead241fb58efdbfaefe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hh5kn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6bkc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:23Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:23 crc kubenswrapper[4824]: I1211 02:01:23.144318 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d14986666febd651039afb52ddf4db8969aca55e88eda9bbb20a67e35ba722ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:23Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:23 crc kubenswrapper[4824]: I1211 02:01:23.161824 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:23Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:23 crc kubenswrapper[4824]: I1211 02:01:23.180036 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44b156e4-64a4-4d45-aa5e-9b10a862faed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7368d92144b188e95a7796fabf497b0f8d9c0b240293bdc7b4aec1ff75497bc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w62dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcf289ac27dbda4f24aabbde3fb5119329e7b4fc7d6e018168da9a69661a4c57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w62dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gx6xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:23Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:23 crc kubenswrapper[4824]: I1211 02:01:23.203283 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:23 crc kubenswrapper[4824]: I1211 02:01:23.203307 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:23 crc kubenswrapper[4824]: I1211 02:01:23.203315 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:23 crc kubenswrapper[4824]: I1211 02:01:23.203328 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:23 crc kubenswrapper[4824]: I1211 02:01:23.203337 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:23Z","lastTransitionTime":"2025-12-11T02:01:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:23 crc kubenswrapper[4824]: I1211 02:01:23.306586 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:23 crc kubenswrapper[4824]: I1211 02:01:23.306632 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:23 crc kubenswrapper[4824]: I1211 02:01:23.306648 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:23 crc kubenswrapper[4824]: I1211 02:01:23.306665 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:23 crc kubenswrapper[4824]: I1211 02:01:23.306677 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:23Z","lastTransitionTime":"2025-12-11T02:01:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:23 crc kubenswrapper[4824]: I1211 02:01:23.408586 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:23 crc kubenswrapper[4824]: I1211 02:01:23.408619 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:23 crc kubenswrapper[4824]: I1211 02:01:23.408627 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:23 crc kubenswrapper[4824]: I1211 02:01:23.408639 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:23 crc kubenswrapper[4824]: I1211 02:01:23.408657 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:23Z","lastTransitionTime":"2025-12-11T02:01:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:23 crc kubenswrapper[4824]: I1211 02:01:23.512229 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:23 crc kubenswrapper[4824]: I1211 02:01:23.512279 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:23 crc kubenswrapper[4824]: I1211 02:01:23.512290 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:23 crc kubenswrapper[4824]: I1211 02:01:23.512308 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:23 crc kubenswrapper[4824]: I1211 02:01:23.512321 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:23Z","lastTransitionTime":"2025-12-11T02:01:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:23 crc kubenswrapper[4824]: I1211 02:01:23.614509 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:23 crc kubenswrapper[4824]: I1211 02:01:23.614534 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:23 crc kubenswrapper[4824]: I1211 02:01:23.614541 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:23 crc kubenswrapper[4824]: I1211 02:01:23.614554 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:23 crc kubenswrapper[4824]: I1211 02:01:23.614563 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:23Z","lastTransitionTime":"2025-12-11T02:01:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:23 crc kubenswrapper[4824]: I1211 02:01:23.631479 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 02:01:23 crc kubenswrapper[4824]: I1211 02:01:23.631520 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 02:01:23 crc kubenswrapper[4824]: I1211 02:01:23.631524 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 02:01:23 crc kubenswrapper[4824]: E1211 02:01:23.631634 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 02:01:23 crc kubenswrapper[4824]: E1211 02:01:23.631829 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 02:01:23 crc kubenswrapper[4824]: E1211 02:01:23.631973 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 02:01:23 crc kubenswrapper[4824]: I1211 02:01:23.720188 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:23 crc kubenswrapper[4824]: I1211 02:01:23.720275 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:23 crc kubenswrapper[4824]: I1211 02:01:23.720301 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:23 crc kubenswrapper[4824]: I1211 02:01:23.720375 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:23 crc kubenswrapper[4824]: I1211 02:01:23.720440 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:23Z","lastTransitionTime":"2025-12-11T02:01:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:23 crc kubenswrapper[4824]: I1211 02:01:23.823225 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:23 crc kubenswrapper[4824]: I1211 02:01:23.823592 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:23 crc kubenswrapper[4824]: I1211 02:01:23.823802 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:23 crc kubenswrapper[4824]: I1211 02:01:23.823980 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:23 crc kubenswrapper[4824]: I1211 02:01:23.824171 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:23Z","lastTransitionTime":"2025-12-11T02:01:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:23 crc kubenswrapper[4824]: I1211 02:01:23.927241 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:23 crc kubenswrapper[4824]: I1211 02:01:23.927306 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:23 crc kubenswrapper[4824]: I1211 02:01:23.927328 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:23 crc kubenswrapper[4824]: I1211 02:01:23.927358 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:23 crc kubenswrapper[4824]: I1211 02:01:23.927378 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:23Z","lastTransitionTime":"2025-12-11T02:01:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:23 crc kubenswrapper[4824]: I1211 02:01:23.928050 4824 generic.go:334] "Generic (PLEG): container finished" podID="50c69d41-a65b-463e-a829-ea349bd2fe46" containerID="f7d7674252f67bde0ded2ec8ad1d925a92a26d3e3f741dee0061b735cb0b7a02" exitCode=0 Dec 11 02:01:23 crc kubenswrapper[4824]: I1211 02:01:23.928101 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-kh8g9" event={"ID":"50c69d41-a65b-463e-a829-ea349bd2fe46","Type":"ContainerDied","Data":"f7d7674252f67bde0ded2ec8ad1d925a92a26d3e3f741dee0061b735cb0b7a02"} Dec 11 02:01:23 crc kubenswrapper[4824]: I1211 02:01:23.946914 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"253d4291-2b2f-487c-8862-3d305f0bdc55\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44b29bb7a6eda7cb890d36f86375b5f3f197ae5cf390f16e71139b0904873e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://922af7f67b99a0ce08776f3f75c0c377e6ea69af1720b6178a6d1e2f9b4b8ba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://136c68020325db8e6930d0f40b0d0d7e013c14c4dd5e7d1449d01d6d458104d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c89afaf56b4a9597f6543ec2e3b176c7a6dab7dc341941fc3bc8ec85cc4edaf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:00:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:23Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:23 crc kubenswrapper[4824]: I1211 02:01:23.970670 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:23Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:23 crc kubenswrapper[4824]: I1211 02:01:23.992092 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"beab386b-6c65-4bc3-9ef0-dc3f68b4403f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c26f06d2ed5e6e98b61a51f073130675f2c49acda106b8aedcbb2a6d4f7580c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20af9b10a67c8b21a7abcd23e42bc412798d159c8422e7c8ff58f5b997ca19a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9bbfc6edfe050b3fa9dbeee5f98d416bcb2a1533a38608a881b75396fd0b059\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7c766f8de73ca762ffde0c18b2d1c46bb152b2b37535a611c5a0e454a21b7d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d45819f0611c91b20c83eeee1092031a3441fb69f3015241632ad419880227d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 02:01:11.070250 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 02:01:11.071673 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2752882565/tls.crt::/tmp/serving-cert-2752882565/tls.key\\\\\\\"\\\\nI1211 02:01:17.042610 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 02:01:17.048046 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 02:01:17.048080 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 02:01:17.048172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 02:01:17.048188 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 02:01:17.058248 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 02:01:17.058294 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 02:01:17.058303 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 02:01:17.058315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 02:01:17.058322 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 02:01:17.058328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 02:01:17.058335 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 02:01:17.058808 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 02:01:17.060225 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14a940a72b2404a6cbd52a69e150b03266bcd7f5c1d3f4d07b51c0c69e51786\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:00:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:23Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:24 crc kubenswrapper[4824]: I1211 02:01:24.018634 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ded146c045085be2b58821749052413a36850ede00ae1024fe2dae9fa5ac84b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:24Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:24 crc kubenswrapper[4824]: I1211 02:01:24.030151 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:24 crc kubenswrapper[4824]: I1211 02:01:24.030198 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:24 crc kubenswrapper[4824]: I1211 02:01:24.030215 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:24 crc kubenswrapper[4824]: I1211 02:01:24.030241 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:24 crc kubenswrapper[4824]: I1211 02:01:24.030259 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:24Z","lastTransitionTime":"2025-12-11T02:01:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:24 crc kubenswrapper[4824]: I1211 02:01:24.042094 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76b10bbd1e891b8c682f4b2e17d023fd787a8ed44ed94fbf7d97314d641ccf36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f32838f33832b9b67e7dd9c53340680e17721cf1c7bdf0170230c5f8feafe64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:24Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:24 crc kubenswrapper[4824]: I1211 02:01:24.058412 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kptws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68568dba-83a5-46c1-8d01-337359679d91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f965a6d6db619995132f61b5d51b66f0b2b3fdbef3fbec63e231b9a62ccd0a9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd7nf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kptws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:24Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:24 crc kubenswrapper[4824]: I1211 02:01:24.074626 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:24Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:24 crc kubenswrapper[4824]: I1211 02:01:24.097433 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kh8g9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"50c69d41-a65b-463e-a829-ea349bd2fe46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69af1f33fd0bafc929b94e702b4f46c6c7d7a362d5552b121729c477892d91d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69af1f33fd0bafc929b94e702b4f46c6c7d7a362d5552b121729c477892d91d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd9675d0afded587bc660a3862fd0c5dd6885961afacda957aa9ee8501d68a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd9675d0afded587bc660a3862fd0c5dd6885961afacda957aa9ee8501d68a69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aac94149458bd6321be90b26b48ee26428d273e7f61407d95f9c9d6b3aeae4b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aac94149458bd6321be90b26b48ee26428d273e7f61407d95f9c9d6b3aeae4b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbee11bd7cb65e031415990869b52d2b8f81bf8b56e6cc935d38e09695db7b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbee11bd7cb65e031415990869b52d2b8f81bf8b56e6cc935d38e09695db7b48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86f693ade72bd593bc16021fe4b7564f1914ec2f1cda7039474ac2fb13de10f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86f693ade72bd593bc16021fe4b7564f1914ec2f1cda7039474ac2fb13de10f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7d7674252f67bde0ded2ec8ad1d925a92a26d3e3f741dee0061b735cb0b7a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7d7674252f67bde0ded2ec8ad1d925a92a26d3e3f741dee0061b735cb0b7a02\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kh8g9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:24Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:24 crc kubenswrapper[4824]: I1211 02:01:24.113515 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6bkc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8022f696-d0e7-437b-848a-3cd25bd1f364\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3e28f6778ec0dba6ea3e6b69e5a2c1c73c0d5db1abeead241fb58efdbfaefe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hh5kn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6bkc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:24Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:24 crc kubenswrapper[4824]: I1211 02:01:24.132786 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:24 crc kubenswrapper[4824]: I1211 02:01:24.132818 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:24 crc kubenswrapper[4824]: I1211 02:01:24.132826 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:24 crc kubenswrapper[4824]: I1211 02:01:24.132840 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:24 crc kubenswrapper[4824]: I1211 02:01:24.132849 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:24Z","lastTransitionTime":"2025-12-11T02:01:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:24 crc kubenswrapper[4824]: I1211 02:01:24.138319 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b193bd2-b536-4056-92f1-94c9836ab2eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f6pgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:24Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:24 crc kubenswrapper[4824]: I1211 02:01:24.151500 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-br4j9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce125194-9c9d-4ea4-8a72-2c30d5722295\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9798f904052e08196a0db1745060dab11e2710308bca01ff160475910f2665df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fc5s8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-br4j9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:24Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:24 crc kubenswrapper[4824]: I1211 02:01:24.166131 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d14986666febd651039afb52ddf4db8969aca55e88eda9bbb20a67e35ba722ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:24Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:24 crc kubenswrapper[4824]: I1211 02:01:24.178131 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:24Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:24 crc kubenswrapper[4824]: I1211 02:01:24.192036 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44b156e4-64a4-4d45-aa5e-9b10a862faed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7368d92144b188e95a7796fabf497b0f8d9c0b240293bdc7b4aec1ff75497bc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w62dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcf289ac27dbda4f24aabbde3fb5119329e7b4fc7d6e018168da9a69661a4c57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w62dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gx6xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:24Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:24 crc kubenswrapper[4824]: I1211 02:01:24.235568 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:24 crc kubenswrapper[4824]: I1211 02:01:24.235624 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:24 crc kubenswrapper[4824]: I1211 02:01:24.235641 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:24 crc kubenswrapper[4824]: I1211 02:01:24.235664 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:24 crc kubenswrapper[4824]: I1211 02:01:24.235683 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:24Z","lastTransitionTime":"2025-12-11T02:01:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:24 crc kubenswrapper[4824]: I1211 02:01:24.338614 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:24 crc kubenswrapper[4824]: I1211 02:01:24.338698 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:24 crc kubenswrapper[4824]: I1211 02:01:24.338723 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:24 crc kubenswrapper[4824]: I1211 02:01:24.338752 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:24 crc kubenswrapper[4824]: I1211 02:01:24.338774 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:24Z","lastTransitionTime":"2025-12-11T02:01:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:24 crc kubenswrapper[4824]: I1211 02:01:24.441529 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:24 crc kubenswrapper[4824]: I1211 02:01:24.441570 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:24 crc kubenswrapper[4824]: I1211 02:01:24.441578 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:24 crc kubenswrapper[4824]: I1211 02:01:24.441593 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:24 crc kubenswrapper[4824]: I1211 02:01:24.441602 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:24Z","lastTransitionTime":"2025-12-11T02:01:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:24 crc kubenswrapper[4824]: I1211 02:01:24.543333 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:24 crc kubenswrapper[4824]: I1211 02:01:24.543377 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:24 crc kubenswrapper[4824]: I1211 02:01:24.543392 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:24 crc kubenswrapper[4824]: I1211 02:01:24.543412 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:24 crc kubenswrapper[4824]: I1211 02:01:24.543428 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:24Z","lastTransitionTime":"2025-12-11T02:01:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:24 crc kubenswrapper[4824]: I1211 02:01:24.645714 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:24 crc kubenswrapper[4824]: I1211 02:01:24.645758 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:24 crc kubenswrapper[4824]: I1211 02:01:24.645772 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:24 crc kubenswrapper[4824]: I1211 02:01:24.645791 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:24 crc kubenswrapper[4824]: I1211 02:01:24.645806 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:24Z","lastTransitionTime":"2025-12-11T02:01:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:24 crc kubenswrapper[4824]: I1211 02:01:24.748562 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:24 crc kubenswrapper[4824]: I1211 02:01:24.748641 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:24 crc kubenswrapper[4824]: I1211 02:01:24.748665 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:24 crc kubenswrapper[4824]: I1211 02:01:24.748694 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:24 crc kubenswrapper[4824]: I1211 02:01:24.748716 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:24Z","lastTransitionTime":"2025-12-11T02:01:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:24 crc kubenswrapper[4824]: I1211 02:01:24.851634 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:24 crc kubenswrapper[4824]: I1211 02:01:24.851720 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:24 crc kubenswrapper[4824]: I1211 02:01:24.851742 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:24 crc kubenswrapper[4824]: I1211 02:01:24.851785 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:24 crc kubenswrapper[4824]: I1211 02:01:24.851820 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:24Z","lastTransitionTime":"2025-12-11T02:01:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:24 crc kubenswrapper[4824]: I1211 02:01:24.939473 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-kh8g9" event={"ID":"50c69d41-a65b-463e-a829-ea349bd2fe46","Type":"ContainerStarted","Data":"3dca6be41bd7a3ce12ecd4567c868f696b65e8eb1bad436feda6e4f9bfb8e525"} Dec 11 02:01:24 crc kubenswrapper[4824]: I1211 02:01:24.953542 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" event={"ID":"5b193bd2-b536-4056-92f1-94c9836ab2eb","Type":"ContainerStarted","Data":"a49b085068cac81453c9619839fab81c33d4d6788bb92cb7f36ba450e8659508"} Dec 11 02:01:24 crc kubenswrapper[4824]: I1211 02:01:24.954315 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" Dec 11 02:01:24 crc kubenswrapper[4824]: I1211 02:01:24.955864 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:24 crc kubenswrapper[4824]: I1211 02:01:24.955906 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:24 crc kubenswrapper[4824]: I1211 02:01:24.955921 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:24 crc kubenswrapper[4824]: I1211 02:01:24.955943 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:24 crc kubenswrapper[4824]: I1211 02:01:24.955961 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:24Z","lastTransitionTime":"2025-12-11T02:01:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:24 crc kubenswrapper[4824]: I1211 02:01:24.964812 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"253d4291-2b2f-487c-8862-3d305f0bdc55\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44b29bb7a6eda7cb890d36f86375b5f3f197ae5cf390f16e71139b0904873e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://922af7f67b99a0ce08776f3f75c0c377e6ea69af1720b6178a6d1e2f9b4b8ba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://136c68020325db8e6930d0f40b0d0d7e013c14c4dd5e7d1449d01d6d458104d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c89afaf56b4a9597f6543ec2e3b176c7a6dab7dc341941fc3bc8ec85cc4edaf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:00:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:24Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:24 crc kubenswrapper[4824]: I1211 02:01:24.982270 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:24Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:24 crc kubenswrapper[4824]: I1211 02:01:24.991640 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" Dec 11 02:01:24 crc kubenswrapper[4824]: I1211 02:01:24.997610 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"beab386b-6c65-4bc3-9ef0-dc3f68b4403f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c26f06d2ed5e6e98b61a51f073130675f2c49acda106b8aedcbb2a6d4f7580c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20af9b10a67c8b21a7abcd23e42bc412798d159c8422e7c8ff58f5b997ca19a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9bbfc6edfe050b3fa9dbeee5f98d416bcb2a1533a38608a881b75396fd0b059\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7c766f8de73ca762ffde0c18b2d1c46bb152b2b37535a611c5a0e454a21b7d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d45819f0611c91b20c83eeee1092031a3441fb69f3015241632ad419880227d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 02:01:11.070250 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 02:01:11.071673 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2752882565/tls.crt::/tmp/serving-cert-2752882565/tls.key\\\\\\\"\\\\nI1211 02:01:17.042610 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 02:01:17.048046 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 02:01:17.048080 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 02:01:17.048172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 02:01:17.048188 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 02:01:17.058248 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 02:01:17.058294 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 02:01:17.058303 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 02:01:17.058315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 02:01:17.058322 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 02:01:17.058328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 02:01:17.058335 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 02:01:17.058808 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 02:01:17.060225 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14a940a72b2404a6cbd52a69e150b03266bcd7f5c1d3f4d07b51c0c69e51786\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:00:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:24Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.013330 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ded146c045085be2b58821749052413a36850ede00ae1024fe2dae9fa5ac84b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:25Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.032307 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76b10bbd1e891b8c682f4b2e17d023fd787a8ed44ed94fbf7d97314d641ccf36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f32838f33832b9b67e7dd9c53340680e17721cf1c7bdf0170230c5f8feafe64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:25Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.046249 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kptws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68568dba-83a5-46c1-8d01-337359679d91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f965a6d6db619995132f61b5d51b66f0b2b3fdbef3fbec63e231b9a62ccd0a9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd7nf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kptws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:25Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.058195 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.058247 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.058263 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.058284 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.058302 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:25Z","lastTransitionTime":"2025-12-11T02:01:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.062002 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:25Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.082717 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kh8g9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"50c69d41-a65b-463e-a829-ea349bd2fe46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dca6be41bd7a3ce12ecd4567c868f696b65e8eb1bad436feda6e4f9bfb8e525\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69af1f33fd0bafc929b94e702b4f46c6c7d7a362d5552b121729c477892d91d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69af1f33fd0bafc929b94e702b4f46c6c7d7a362d5552b121729c477892d91d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd9675d0afded587bc660a3862fd0c5dd6885961afacda957aa9ee8501d68a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd9675d0afded587bc660a3862fd0c5dd6885961afacda957aa9ee8501d68a69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aac94149458bd6321be90b26b48ee26428d273e7f61407d95f9c9d6b3aeae4b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aac94149458bd6321be90b26b48ee26428d273e7f61407d95f9c9d6b3aeae4b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbee11bd7cb65e031415990869b52d2b8f81bf8b56e6cc935d38e09695db7b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbee11bd7cb65e031415990869b52d2b8f81bf8b56e6cc935d38e09695db7b48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86f693ade72bd593bc16021fe4b7564f1914ec2f1cda7039474ac2fb13de10f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86f693ade72bd593bc16021fe4b7564f1914ec2f1cda7039474ac2fb13de10f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7d7674252f67bde0ded2ec8ad1d925a92a26d3e3f741dee0061b735cb0b7a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7d7674252f67bde0ded2ec8ad1d925a92a26d3e3f741dee0061b735cb0b7a02\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kh8g9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:25Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.101743 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6bkc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8022f696-d0e7-437b-848a-3cd25bd1f364\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3e28f6778ec0dba6ea3e6b69e5a2c1c73c0d5db1abeead241fb58efdbfaefe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hh5kn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6bkc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:25Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.124479 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b193bd2-b536-4056-92f1-94c9836ab2eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f6pgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:25Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.136354 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-br4j9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce125194-9c9d-4ea4-8a72-2c30d5722295\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9798f904052e08196a0db1745060dab11e2710308bca01ff160475910f2665df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fc5s8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-br4j9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:25Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.151922 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d14986666febd651039afb52ddf4db8969aca55e88eda9bbb20a67e35ba722ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:25Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.160186 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.160229 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.160239 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.160254 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.160264 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:25Z","lastTransitionTime":"2025-12-11T02:01:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.166602 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:25Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.182746 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44b156e4-64a4-4d45-aa5e-9b10a862faed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7368d92144b188e95a7796fabf497b0f8d9c0b240293bdc7b4aec1ff75497bc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w62dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcf289ac27dbda4f24aabbde3fb5119329e7b4fc7d6e018168da9a69661a4c57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w62dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gx6xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:25Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.202052 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:25Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.220621 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kh8g9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"50c69d41-a65b-463e-a829-ea349bd2fe46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dca6be41bd7a3ce12ecd4567c868f696b65e8eb1bad436feda6e4f9bfb8e525\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69af1f33fd0bafc929b94e702b4f46c6c7d7a362d5552b121729c477892d91d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69af1f33fd0bafc929b94e702b4f46c6c7d7a362d5552b121729c477892d91d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd9675d0afded587bc660a3862fd0c5dd6885961afacda957aa9ee8501d68a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd9675d0afded587bc660a3862fd0c5dd6885961afacda957aa9ee8501d68a69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aac94149458bd6321be90b26b48ee26428d273e7f61407d95f9c9d6b3aeae4b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aac94149458bd6321be90b26b48ee26428d273e7f61407d95f9c9d6b3aeae4b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbee11bd7cb65e031415990869b52d2b8f81bf8b56e6cc935d38e09695db7b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbee11bd7cb65e031415990869b52d2b8f81bf8b56e6cc935d38e09695db7b48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86f693ade72bd593bc16021fe4b7564f1914ec2f1cda7039474ac2fb13de10f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86f693ade72bd593bc16021fe4b7564f1914ec2f1cda7039474ac2fb13de10f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7d7674252f67bde0ded2ec8ad1d925a92a26d3e3f741dee0061b735cb0b7a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7d7674252f67bde0ded2ec8ad1d925a92a26d3e3f741dee0061b735cb0b7a02\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kh8g9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:25Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.240837 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6bkc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8022f696-d0e7-437b-848a-3cd25bd1f364\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3e28f6778ec0dba6ea3e6b69e5a2c1c73c0d5db1abeead241fb58efdbfaefe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hh5kn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6bkc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:25Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.263659 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.263705 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.263717 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.263733 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.263744 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:25Z","lastTransitionTime":"2025-12-11T02:01:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.269661 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b193bd2-b536-4056-92f1-94c9836ab2eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51f3c41d86ddc44f7eca118029876ad353317f753d0ad173ed4d30c2a15f10e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c345850891744a0240781277e9a89a59f6899c4c944995f62e44650f807440a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed2ae40b1fe28111efc76d9d72345edecb5f3ea34522b238f3730812b29f2bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce7aa078800e6f299cfe03770fa7e3b64fa3742fe0c0f383c87d84cf48569ddf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6aa9fc1f1bff192c69bd59d32a05d4a72d09641d9ce4b265bc4517147a7e15dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b4a3bc50eb80777188206adab774e8acad8b2e4952a8415f91b1c29ba9a0ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a49b085068cac81453c9619839fab81c33d4d6788bb92cb7f36ba450e8659508\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ead58694813806d307bbbc6a7b3cf41730d1c059b231a3002684bf09c10c7f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f6pgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:25Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.282879 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-br4j9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce125194-9c9d-4ea4-8a72-2c30d5722295\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9798f904052e08196a0db1745060dab11e2710308bca01ff160475910f2665df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fc5s8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-br4j9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:25Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.300925 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d14986666febd651039afb52ddf4db8969aca55e88eda9bbb20a67e35ba722ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:25Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.316581 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:25Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.332517 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 02:01:25 crc kubenswrapper[4824]: E1211 02:01:25.332671 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 02:01:33.332647623 +0000 UTC m=+35.021685012 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.333801 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44b156e4-64a4-4d45-aa5e-9b10a862faed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7368d92144b188e95a7796fabf497b0f8d9c0b240293bdc7b4aec1ff75497bc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w62dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcf289ac27dbda4f24aabbde3fb5119329e7b4fc7d6e018168da9a69661a4c57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w62dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gx6xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:25Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.354008 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"253d4291-2b2f-487c-8862-3d305f0bdc55\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44b29bb7a6eda7cb890d36f86375b5f3f197ae5cf390f16e71139b0904873e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://922af7f67b99a0ce08776f3f75c0c377e6ea69af1720b6178a6d1e2f9b4b8ba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://136c68020325db8e6930d0f40b0d0d7e013c14c4dd5e7d1449d01d6d458104d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c89afaf56b4a9597f6543ec2e3b176c7a6dab7dc341941fc3bc8ec85cc4edaf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:00:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:25Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.366290 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.366332 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.366344 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.366362 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.366382 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:25Z","lastTransitionTime":"2025-12-11T02:01:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.374862 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:25Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.395826 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"beab386b-6c65-4bc3-9ef0-dc3f68b4403f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c26f06d2ed5e6e98b61a51f073130675f2c49acda106b8aedcbb2a6d4f7580c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20af9b10a67c8b21a7abcd23e42bc412798d159c8422e7c8ff58f5b997ca19a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9bbfc6edfe050b3fa9dbeee5f98d416bcb2a1533a38608a881b75396fd0b059\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7c766f8de73ca762ffde0c18b2d1c46bb152b2b37535a611c5a0e454a21b7d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d45819f0611c91b20c83eeee1092031a3441fb69f3015241632ad419880227d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 02:01:11.070250 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 02:01:11.071673 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2752882565/tls.crt::/tmp/serving-cert-2752882565/tls.key\\\\\\\"\\\\nI1211 02:01:17.042610 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 02:01:17.048046 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 02:01:17.048080 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 02:01:17.048172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 02:01:17.048188 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 02:01:17.058248 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 02:01:17.058294 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 02:01:17.058303 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 02:01:17.058315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 02:01:17.058322 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 02:01:17.058328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 02:01:17.058335 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 02:01:17.058808 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 02:01:17.060225 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14a940a72b2404a6cbd52a69e150b03266bcd7f5c1d3f4d07b51c0c69e51786\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:00:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:25Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.412674 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ded146c045085be2b58821749052413a36850ede00ae1024fe2dae9fa5ac84b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:25Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.430906 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76b10bbd1e891b8c682f4b2e17d023fd787a8ed44ed94fbf7d97314d641ccf36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f32838f33832b9b67e7dd9c53340680e17721cf1c7bdf0170230c5f8feafe64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:25Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.433478 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.433541 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.433583 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.433637 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 02:01:25 crc kubenswrapper[4824]: E1211 02:01:25.433631 4824 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 11 02:01:25 crc kubenswrapper[4824]: E1211 02:01:25.433747 4824 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 11 02:01:25 crc kubenswrapper[4824]: E1211 02:01:25.433762 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 11 02:01:25 crc kubenswrapper[4824]: E1211 02:01:25.433823 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 11 02:01:25 crc kubenswrapper[4824]: E1211 02:01:25.433761 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-11 02:01:33.433738252 +0000 UTC m=+35.122775671 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 11 02:01:25 crc kubenswrapper[4824]: E1211 02:01:25.433847 4824 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 02:01:25 crc kubenswrapper[4824]: E1211 02:01:25.433882 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-11 02:01:33.433850005 +0000 UTC m=+35.122887424 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 11 02:01:25 crc kubenswrapper[4824]: E1211 02:01:25.433913 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 11 02:01:25 crc kubenswrapper[4824]: E1211 02:01:25.433981 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 11 02:01:25 crc kubenswrapper[4824]: E1211 02:01:25.433929 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-11 02:01:33.433904026 +0000 UTC m=+35.122941435 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 02:01:25 crc kubenswrapper[4824]: E1211 02:01:25.434009 4824 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 02:01:25 crc kubenswrapper[4824]: E1211 02:01:25.434162 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-11 02:01:33.434089611 +0000 UTC m=+35.123127030 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.450068 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kptws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68568dba-83a5-46c1-8d01-337359679d91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f965a6d6db619995132f61b5d51b66f0b2b3fdbef3fbec63e231b9a62ccd0a9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd7nf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kptws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:25Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.469640 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.469696 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.469711 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.469735 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.469752 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:25Z","lastTransitionTime":"2025-12-11T02:01:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.572852 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.572912 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.572935 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.572960 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.572980 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:25Z","lastTransitionTime":"2025-12-11T02:01:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.631603 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.631697 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.631614 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 02:01:25 crc kubenswrapper[4824]: E1211 02:01:25.631877 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 02:01:25 crc kubenswrapper[4824]: E1211 02:01:25.631779 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 02:01:25 crc kubenswrapper[4824]: E1211 02:01:25.632068 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.675632 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.675684 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.675700 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.675724 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.675741 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:25Z","lastTransitionTime":"2025-12-11T02:01:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.779216 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.779272 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.779291 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.779315 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.779333 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:25Z","lastTransitionTime":"2025-12-11T02:01:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.882228 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.882309 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.882332 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.882366 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.882388 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:25Z","lastTransitionTime":"2025-12-11T02:01:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.956637 4824 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.957084 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.984546 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.984603 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.984621 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.984647 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.984668 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:25Z","lastTransitionTime":"2025-12-11T02:01:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:25 crc kubenswrapper[4824]: I1211 02:01:25.992552 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" Dec 11 02:01:26 crc kubenswrapper[4824]: I1211 02:01:26.011409 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"253d4291-2b2f-487c-8862-3d305f0bdc55\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44b29bb7a6eda7cb890d36f86375b5f3f197ae5cf390f16e71139b0904873e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://922af7f67b99a0ce08776f3f75c0c377e6ea69af1720b6178a6d1e2f9b4b8ba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://136c68020325db8e6930d0f40b0d0d7e013c14c4dd5e7d1449d01d6d458104d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c89afaf56b4a9597f6543ec2e3b176c7a6dab7dc341941fc3bc8ec85cc4edaf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:00:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:26Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:26 crc kubenswrapper[4824]: I1211 02:01:26.031254 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:26Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:26 crc kubenswrapper[4824]: I1211 02:01:26.053225 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"beab386b-6c65-4bc3-9ef0-dc3f68b4403f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c26f06d2ed5e6e98b61a51f073130675f2c49acda106b8aedcbb2a6d4f7580c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20af9b10a67c8b21a7abcd23e42bc412798d159c8422e7c8ff58f5b997ca19a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9bbfc6edfe050b3fa9dbeee5f98d416bcb2a1533a38608a881b75396fd0b059\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7c766f8de73ca762ffde0c18b2d1c46bb152b2b37535a611c5a0e454a21b7d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d45819f0611c91b20c83eeee1092031a3441fb69f3015241632ad419880227d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 02:01:11.070250 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 02:01:11.071673 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2752882565/tls.crt::/tmp/serving-cert-2752882565/tls.key\\\\\\\"\\\\nI1211 02:01:17.042610 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 02:01:17.048046 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 02:01:17.048080 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 02:01:17.048172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 02:01:17.048188 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 02:01:17.058248 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 02:01:17.058294 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 02:01:17.058303 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 02:01:17.058315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 02:01:17.058322 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 02:01:17.058328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 02:01:17.058335 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 02:01:17.058808 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 02:01:17.060225 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14a940a72b2404a6cbd52a69e150b03266bcd7f5c1d3f4d07b51c0c69e51786\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:00:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:26Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:26 crc kubenswrapper[4824]: I1211 02:01:26.073822 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ded146c045085be2b58821749052413a36850ede00ae1024fe2dae9fa5ac84b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:26Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:26 crc kubenswrapper[4824]: I1211 02:01:26.087355 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:26 crc kubenswrapper[4824]: I1211 02:01:26.087407 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:26 crc kubenswrapper[4824]: I1211 02:01:26.087427 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:26 crc kubenswrapper[4824]: I1211 02:01:26.087451 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:26 crc kubenswrapper[4824]: I1211 02:01:26.087469 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:26Z","lastTransitionTime":"2025-12-11T02:01:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:26 crc kubenswrapper[4824]: I1211 02:01:26.094931 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76b10bbd1e891b8c682f4b2e17d023fd787a8ed44ed94fbf7d97314d641ccf36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f32838f33832b9b67e7dd9c53340680e17721cf1c7bdf0170230c5f8feafe64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:26Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:26 crc kubenswrapper[4824]: I1211 02:01:26.113390 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kptws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68568dba-83a5-46c1-8d01-337359679d91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f965a6d6db619995132f61b5d51b66f0b2b3fdbef3fbec63e231b9a62ccd0a9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd7nf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kptws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:26Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:26 crc kubenswrapper[4824]: I1211 02:01:26.130995 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:26Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:26 crc kubenswrapper[4824]: I1211 02:01:26.152093 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kh8g9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"50c69d41-a65b-463e-a829-ea349bd2fe46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dca6be41bd7a3ce12ecd4567c868f696b65e8eb1bad436feda6e4f9bfb8e525\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69af1f33fd0bafc929b94e702b4f46c6c7d7a362d5552b121729c477892d91d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69af1f33fd0bafc929b94e702b4f46c6c7d7a362d5552b121729c477892d91d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd9675d0afded587bc660a3862fd0c5dd6885961afacda957aa9ee8501d68a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd9675d0afded587bc660a3862fd0c5dd6885961afacda957aa9ee8501d68a69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aac94149458bd6321be90b26b48ee26428d273e7f61407d95f9c9d6b3aeae4b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aac94149458bd6321be90b26b48ee26428d273e7f61407d95f9c9d6b3aeae4b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbee11bd7cb65e031415990869b52d2b8f81bf8b56e6cc935d38e09695db7b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbee11bd7cb65e031415990869b52d2b8f81bf8b56e6cc935d38e09695db7b48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86f693ade72bd593bc16021fe4b7564f1914ec2f1cda7039474ac2fb13de10f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86f693ade72bd593bc16021fe4b7564f1914ec2f1cda7039474ac2fb13de10f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7d7674252f67bde0ded2ec8ad1d925a92a26d3e3f741dee0061b735cb0b7a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7d7674252f67bde0ded2ec8ad1d925a92a26d3e3f741dee0061b735cb0b7a02\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kh8g9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:26Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:26 crc kubenswrapper[4824]: I1211 02:01:26.170158 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6bkc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8022f696-d0e7-437b-848a-3cd25bd1f364\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3e28f6778ec0dba6ea3e6b69e5a2c1c73c0d5db1abeead241fb58efdbfaefe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hh5kn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6bkc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:26Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:26 crc kubenswrapper[4824]: I1211 02:01:26.189800 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:26 crc kubenswrapper[4824]: I1211 02:01:26.189854 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:26 crc kubenswrapper[4824]: I1211 02:01:26.189870 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:26 crc kubenswrapper[4824]: I1211 02:01:26.189888 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:26 crc kubenswrapper[4824]: I1211 02:01:26.189901 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:26Z","lastTransitionTime":"2025-12-11T02:01:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:26 crc kubenswrapper[4824]: I1211 02:01:26.193547 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b193bd2-b536-4056-92f1-94c9836ab2eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51f3c41d86ddc44f7eca118029876ad353317f753d0ad173ed4d30c2a15f10e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c345850891744a0240781277e9a89a59f6899c4c944995f62e44650f807440a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed2ae40b1fe28111efc76d9d72345edecb5f3ea34522b238f3730812b29f2bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce7aa078800e6f299cfe03770fa7e3b64fa3742fe0c0f383c87d84cf48569ddf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6aa9fc1f1bff192c69bd59d32a05d4a72d09641d9ce4b265bc4517147a7e15dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b4a3bc50eb80777188206adab774e8acad8b2e4952a8415f91b1c29ba9a0ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a49b085068cac81453c9619839fab81c33d4d6788bb92cb7f36ba450e8659508\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ead58694813806d307bbbc6a7b3cf41730d1c059b231a3002684bf09c10c7f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f6pgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:26Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:26 crc kubenswrapper[4824]: I1211 02:01:26.212642 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-br4j9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce125194-9c9d-4ea4-8a72-2c30d5722295\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9798f904052e08196a0db1745060dab11e2710308bca01ff160475910f2665df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fc5s8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-br4j9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:26Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:26 crc kubenswrapper[4824]: I1211 02:01:26.239975 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d14986666febd651039afb52ddf4db8969aca55e88eda9bbb20a67e35ba722ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:26Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:26 crc kubenswrapper[4824]: I1211 02:01:26.260522 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:26Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:26 crc kubenswrapper[4824]: I1211 02:01:26.272904 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44b156e4-64a4-4d45-aa5e-9b10a862faed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7368d92144b188e95a7796fabf497b0f8d9c0b240293bdc7b4aec1ff75497bc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w62dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcf289ac27dbda4f24aabbde3fb5119329e7b4fc7d6e018168da9a69661a4c57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w62dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gx6xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:26Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:26 crc kubenswrapper[4824]: I1211 02:01:26.292514 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:26 crc kubenswrapper[4824]: I1211 02:01:26.292553 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:26 crc kubenswrapper[4824]: I1211 02:01:26.292562 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:26 crc kubenswrapper[4824]: I1211 02:01:26.292579 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:26 crc kubenswrapper[4824]: I1211 02:01:26.292588 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:26Z","lastTransitionTime":"2025-12-11T02:01:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:26 crc kubenswrapper[4824]: I1211 02:01:26.395307 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:26 crc kubenswrapper[4824]: I1211 02:01:26.395373 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:26 crc kubenswrapper[4824]: I1211 02:01:26.395392 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:26 crc kubenswrapper[4824]: I1211 02:01:26.395441 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:26 crc kubenswrapper[4824]: I1211 02:01:26.395465 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:26Z","lastTransitionTime":"2025-12-11T02:01:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:26 crc kubenswrapper[4824]: I1211 02:01:26.498266 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:26 crc kubenswrapper[4824]: I1211 02:01:26.498344 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:26 crc kubenswrapper[4824]: I1211 02:01:26.498362 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:26 crc kubenswrapper[4824]: I1211 02:01:26.498386 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:26 crc kubenswrapper[4824]: I1211 02:01:26.498403 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:26Z","lastTransitionTime":"2025-12-11T02:01:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:26 crc kubenswrapper[4824]: I1211 02:01:26.601581 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:26 crc kubenswrapper[4824]: I1211 02:01:26.601644 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:26 crc kubenswrapper[4824]: I1211 02:01:26.601663 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:26 crc kubenswrapper[4824]: I1211 02:01:26.601688 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:26 crc kubenswrapper[4824]: I1211 02:01:26.601705 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:26Z","lastTransitionTime":"2025-12-11T02:01:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:26 crc kubenswrapper[4824]: I1211 02:01:26.705102 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:26 crc kubenswrapper[4824]: I1211 02:01:26.705186 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:26 crc kubenswrapper[4824]: I1211 02:01:26.705205 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:26 crc kubenswrapper[4824]: I1211 02:01:26.705231 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:26 crc kubenswrapper[4824]: I1211 02:01:26.705248 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:26Z","lastTransitionTime":"2025-12-11T02:01:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:26 crc kubenswrapper[4824]: I1211 02:01:26.807393 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:26 crc kubenswrapper[4824]: I1211 02:01:26.807426 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:26 crc kubenswrapper[4824]: I1211 02:01:26.807438 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:26 crc kubenswrapper[4824]: I1211 02:01:26.807456 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:26 crc kubenswrapper[4824]: I1211 02:01:26.807465 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:26Z","lastTransitionTime":"2025-12-11T02:01:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:26 crc kubenswrapper[4824]: I1211 02:01:26.910331 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:26 crc kubenswrapper[4824]: I1211 02:01:26.910376 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:26 crc kubenswrapper[4824]: I1211 02:01:26.910389 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:26 crc kubenswrapper[4824]: I1211 02:01:26.910406 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:26 crc kubenswrapper[4824]: I1211 02:01:26.910419 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:26Z","lastTransitionTime":"2025-12-11T02:01:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:26 crc kubenswrapper[4824]: I1211 02:01:26.958933 4824 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 11 02:01:27 crc kubenswrapper[4824]: I1211 02:01:27.012739 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:27 crc kubenswrapper[4824]: I1211 02:01:27.012785 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:27 crc kubenswrapper[4824]: I1211 02:01:27.012801 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:27 crc kubenswrapper[4824]: I1211 02:01:27.012824 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:27 crc kubenswrapper[4824]: I1211 02:01:27.012841 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:27Z","lastTransitionTime":"2025-12-11T02:01:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:27 crc kubenswrapper[4824]: I1211 02:01:27.118532 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:27 crc kubenswrapper[4824]: I1211 02:01:27.118594 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:27 crc kubenswrapper[4824]: I1211 02:01:27.118631 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:27 crc kubenswrapper[4824]: I1211 02:01:27.118666 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:27 crc kubenswrapper[4824]: I1211 02:01:27.118689 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:27Z","lastTransitionTime":"2025-12-11T02:01:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:27 crc kubenswrapper[4824]: I1211 02:01:27.221588 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:27 crc kubenswrapper[4824]: I1211 02:01:27.221655 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:27 crc kubenswrapper[4824]: I1211 02:01:27.221674 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:27 crc kubenswrapper[4824]: I1211 02:01:27.221698 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:27 crc kubenswrapper[4824]: I1211 02:01:27.221717 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:27Z","lastTransitionTime":"2025-12-11T02:01:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:27 crc kubenswrapper[4824]: I1211 02:01:27.325386 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:27 crc kubenswrapper[4824]: I1211 02:01:27.325451 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:27 crc kubenswrapper[4824]: I1211 02:01:27.325469 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:27 crc kubenswrapper[4824]: I1211 02:01:27.325495 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:27 crc kubenswrapper[4824]: I1211 02:01:27.325513 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:27Z","lastTransitionTime":"2025-12-11T02:01:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:27 crc kubenswrapper[4824]: I1211 02:01:27.428614 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:27 crc kubenswrapper[4824]: I1211 02:01:27.428673 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:27 crc kubenswrapper[4824]: I1211 02:01:27.428691 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:27 crc kubenswrapper[4824]: I1211 02:01:27.428715 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:27 crc kubenswrapper[4824]: I1211 02:01:27.428732 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:27Z","lastTransitionTime":"2025-12-11T02:01:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:27 crc kubenswrapper[4824]: I1211 02:01:27.530905 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:27 crc kubenswrapper[4824]: I1211 02:01:27.531001 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:27 crc kubenswrapper[4824]: I1211 02:01:27.531020 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:27 crc kubenswrapper[4824]: I1211 02:01:27.531044 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:27 crc kubenswrapper[4824]: I1211 02:01:27.531066 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:27Z","lastTransitionTime":"2025-12-11T02:01:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:27 crc kubenswrapper[4824]: I1211 02:01:27.631700 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 02:01:27 crc kubenswrapper[4824]: E1211 02:01:27.631883 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 02:01:27 crc kubenswrapper[4824]: I1211 02:01:27.632005 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 02:01:27 crc kubenswrapper[4824]: E1211 02:01:27.632238 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 02:01:27 crc kubenswrapper[4824]: I1211 02:01:27.632394 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 02:01:27 crc kubenswrapper[4824]: E1211 02:01:27.632544 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 02:01:27 crc kubenswrapper[4824]: I1211 02:01:27.634670 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:27 crc kubenswrapper[4824]: I1211 02:01:27.634713 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:27 crc kubenswrapper[4824]: I1211 02:01:27.634726 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:27 crc kubenswrapper[4824]: I1211 02:01:27.634787 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:27 crc kubenswrapper[4824]: I1211 02:01:27.634884 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:27Z","lastTransitionTime":"2025-12-11T02:01:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:27 crc kubenswrapper[4824]: I1211 02:01:27.738005 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:27 crc kubenswrapper[4824]: I1211 02:01:27.738055 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:27 crc kubenswrapper[4824]: I1211 02:01:27.738065 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:27 crc kubenswrapper[4824]: I1211 02:01:27.738082 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:27 crc kubenswrapper[4824]: I1211 02:01:27.738093 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:27Z","lastTransitionTime":"2025-12-11T02:01:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:27 crc kubenswrapper[4824]: I1211 02:01:27.841962 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:27 crc kubenswrapper[4824]: I1211 02:01:27.842026 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:27 crc kubenswrapper[4824]: I1211 02:01:27.842044 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:27 crc kubenswrapper[4824]: I1211 02:01:27.842075 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:27 crc kubenswrapper[4824]: I1211 02:01:27.842097 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:27Z","lastTransitionTime":"2025-12-11T02:01:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:27 crc kubenswrapper[4824]: I1211 02:01:27.945775 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:27 crc kubenswrapper[4824]: I1211 02:01:27.945851 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:27 crc kubenswrapper[4824]: I1211 02:01:27.945866 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:27 crc kubenswrapper[4824]: I1211 02:01:27.945888 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:27 crc kubenswrapper[4824]: I1211 02:01:27.945901 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:27Z","lastTransitionTime":"2025-12-11T02:01:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:27 crc kubenswrapper[4824]: I1211 02:01:27.964667 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f6pgc_5b193bd2-b536-4056-92f1-94c9836ab2eb/ovnkube-controller/0.log" Dec 11 02:01:27 crc kubenswrapper[4824]: I1211 02:01:27.968644 4824 generic.go:334] "Generic (PLEG): container finished" podID="5b193bd2-b536-4056-92f1-94c9836ab2eb" containerID="a49b085068cac81453c9619839fab81c33d4d6788bb92cb7f36ba450e8659508" exitCode=1 Dec 11 02:01:27 crc kubenswrapper[4824]: I1211 02:01:27.968712 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" event={"ID":"5b193bd2-b536-4056-92f1-94c9836ab2eb","Type":"ContainerDied","Data":"a49b085068cac81453c9619839fab81c33d4d6788bb92cb7f36ba450e8659508"} Dec 11 02:01:27 crc kubenswrapper[4824]: I1211 02:01:27.969750 4824 scope.go:117] "RemoveContainer" containerID="a49b085068cac81453c9619839fab81c33d4d6788bb92cb7f36ba450e8659508" Dec 11 02:01:27 crc kubenswrapper[4824]: I1211 02:01:27.985560 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-br4j9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce125194-9c9d-4ea4-8a72-2c30d5722295\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9798f904052e08196a0db1745060dab11e2710308bca01ff160475910f2665df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fc5s8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-br4j9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:27Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:28 crc kubenswrapper[4824]: I1211 02:01:28.005734 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:28Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:28 crc kubenswrapper[4824]: I1211 02:01:28.029882 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kh8g9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"50c69d41-a65b-463e-a829-ea349bd2fe46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dca6be41bd7a3ce12ecd4567c868f696b65e8eb1bad436feda6e4f9bfb8e525\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69af1f33fd0bafc929b94e702b4f46c6c7d7a362d5552b121729c477892d91d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69af1f33fd0bafc929b94e702b4f46c6c7d7a362d5552b121729c477892d91d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd9675d0afded587bc660a3862fd0c5dd6885961afacda957aa9ee8501d68a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd9675d0afded587bc660a3862fd0c5dd6885961afacda957aa9ee8501d68a69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aac94149458bd6321be90b26b48ee26428d273e7f61407d95f9c9d6b3aeae4b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aac94149458bd6321be90b26b48ee26428d273e7f61407d95f9c9d6b3aeae4b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbee11bd7cb65e031415990869b52d2b8f81bf8b56e6cc935d38e09695db7b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbee11bd7cb65e031415990869b52d2b8f81bf8b56e6cc935d38e09695db7b48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86f693ade72bd593bc16021fe4b7564f1914ec2f1cda7039474ac2fb13de10f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86f693ade72bd593bc16021fe4b7564f1914ec2f1cda7039474ac2fb13de10f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7d7674252f67bde0ded2ec8ad1d925a92a26d3e3f741dee0061b735cb0b7a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7d7674252f67bde0ded2ec8ad1d925a92a26d3e3f741dee0061b735cb0b7a02\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kh8g9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:28Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:28 crc kubenswrapper[4824]: I1211 02:01:28.050786 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:28 crc kubenswrapper[4824]: I1211 02:01:28.050869 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:28 crc kubenswrapper[4824]: I1211 02:01:28.050889 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:28 crc kubenswrapper[4824]: I1211 02:01:28.050917 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:28 crc kubenswrapper[4824]: I1211 02:01:28.050932 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:28Z","lastTransitionTime":"2025-12-11T02:01:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:28 crc kubenswrapper[4824]: I1211 02:01:28.053089 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6bkc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8022f696-d0e7-437b-848a-3cd25bd1f364\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3e28f6778ec0dba6ea3e6b69e5a2c1c73c0d5db1abeead241fb58efdbfaefe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hh5kn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6bkc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:28Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:28 crc kubenswrapper[4824]: I1211 02:01:28.085845 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b193bd2-b536-4056-92f1-94c9836ab2eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51f3c41d86ddc44f7eca118029876ad353317f753d0ad173ed4d30c2a15f10e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c345850891744a0240781277e9a89a59f6899c4c944995f62e44650f807440a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed2ae40b1fe28111efc76d9d72345edecb5f3ea34522b238f3730812b29f2bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce7aa078800e6f299cfe03770fa7e3b64fa3742fe0c0f383c87d84cf48569ddf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6aa9fc1f1bff192c69bd59d32a05d4a72d09641d9ce4b265bc4517147a7e15dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b4a3bc50eb80777188206adab774e8acad8b2e4952a8415f91b1c29ba9a0ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a49b085068cac81453c9619839fab81c33d4d6788bb92cb7f36ba450e8659508\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a49b085068cac81453c9619839fab81c33d4d6788bb92cb7f36ba450e8659508\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T02:01:27Z\\\",\\\"message\\\":\\\" 6078 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1211 02:01:26.910738 6078 handler.go:208] Removed *v1.Node event handler 2\\\\nI1211 02:01:26.910745 6078 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1211 02:01:26.910827 6078 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1211 02:01:26.911133 6078 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1211 02:01:26.911294 6078 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1211 02:01:26.911525 6078 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1211 02:01:26.911572 6078 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1211 02:01:26.911601 6078 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1211 02:01:26.911644 6078 factory.go:656] Stopping watch factory\\\\nI1211 02:01:26.911667 6078 ovnkube.go:599] Stopped ovnkube\\\\nI1211 02:01:26.911699 6078 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1211 02:01:26.911719 6078 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1211 02:01:2\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ead58694813806d307bbbc6a7b3cf41730d1c059b231a3002684bf09c10c7f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f6pgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:28Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:28 crc kubenswrapper[4824]: I1211 02:01:28.104631 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d14986666febd651039afb52ddf4db8969aca55e88eda9bbb20a67e35ba722ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:28Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:28 crc kubenswrapper[4824]: I1211 02:01:28.122711 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:28Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:28 crc kubenswrapper[4824]: I1211 02:01:28.140308 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44b156e4-64a4-4d45-aa5e-9b10a862faed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7368d92144b188e95a7796fabf497b0f8d9c0b240293bdc7b4aec1ff75497bc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w62dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcf289ac27dbda4f24aabbde3fb5119329e7b4fc7d6e018168da9a69661a4c57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w62dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gx6xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:28Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:28 crc kubenswrapper[4824]: I1211 02:01:28.153577 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:28 crc kubenswrapper[4824]: I1211 02:01:28.153602 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:28 crc kubenswrapper[4824]: I1211 02:01:28.153613 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:28 crc kubenswrapper[4824]: I1211 02:01:28.153636 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:28 crc kubenswrapper[4824]: I1211 02:01:28.153648 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:28Z","lastTransitionTime":"2025-12-11T02:01:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:28 crc kubenswrapper[4824]: I1211 02:01:28.186459 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"253d4291-2b2f-487c-8862-3d305f0bdc55\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44b29bb7a6eda7cb890d36f86375b5f3f197ae5cf390f16e71139b0904873e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://922af7f67b99a0ce08776f3f75c0c377e6ea69af1720b6178a6d1e2f9b4b8ba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://136c68020325db8e6930d0f40b0d0d7e013c14c4dd5e7d1449d01d6d458104d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c89afaf56b4a9597f6543ec2e3b176c7a6dab7dc341941fc3bc8ec85cc4edaf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:00:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:28Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:28 crc kubenswrapper[4824]: I1211 02:01:28.207450 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:28Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:28 crc kubenswrapper[4824]: I1211 02:01:28.222711 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kptws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68568dba-83a5-46c1-8d01-337359679d91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f965a6d6db619995132f61b5d51b66f0b2b3fdbef3fbec63e231b9a62ccd0a9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd7nf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kptws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:28Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:28 crc kubenswrapper[4824]: I1211 02:01:28.239541 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"beab386b-6c65-4bc3-9ef0-dc3f68b4403f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c26f06d2ed5e6e98b61a51f073130675f2c49acda106b8aedcbb2a6d4f7580c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20af9b10a67c8b21a7abcd23e42bc412798d159c8422e7c8ff58f5b997ca19a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9bbfc6edfe050b3fa9dbeee5f98d416bcb2a1533a38608a881b75396fd0b059\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7c766f8de73ca762ffde0c18b2d1c46bb152b2b37535a611c5a0e454a21b7d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d45819f0611c91b20c83eeee1092031a3441fb69f3015241632ad419880227d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 02:01:11.070250 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 02:01:11.071673 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2752882565/tls.crt::/tmp/serving-cert-2752882565/tls.key\\\\\\\"\\\\nI1211 02:01:17.042610 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 02:01:17.048046 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 02:01:17.048080 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 02:01:17.048172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 02:01:17.048188 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 02:01:17.058248 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 02:01:17.058294 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 02:01:17.058303 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 02:01:17.058315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 02:01:17.058322 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 02:01:17.058328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 02:01:17.058335 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 02:01:17.058808 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 02:01:17.060225 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14a940a72b2404a6cbd52a69e150b03266bcd7f5c1d3f4d07b51c0c69e51786\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:00:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:28Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:28 crc kubenswrapper[4824]: I1211 02:01:28.257055 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:28 crc kubenswrapper[4824]: I1211 02:01:28.257081 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:28 crc kubenswrapper[4824]: I1211 02:01:28.257089 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:28 crc kubenswrapper[4824]: I1211 02:01:28.257103 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:28 crc kubenswrapper[4824]: I1211 02:01:28.257129 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:28Z","lastTransitionTime":"2025-12-11T02:01:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:28 crc kubenswrapper[4824]: I1211 02:01:28.263159 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ded146c045085be2b58821749052413a36850ede00ae1024fe2dae9fa5ac84b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:28Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:28 crc kubenswrapper[4824]: I1211 02:01:28.280732 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76b10bbd1e891b8c682f4b2e17d023fd787a8ed44ed94fbf7d97314d641ccf36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f32838f33832b9b67e7dd9c53340680e17721cf1c7bdf0170230c5f8feafe64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:28Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:28 crc kubenswrapper[4824]: I1211 02:01:28.359967 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:28 crc kubenswrapper[4824]: I1211 02:01:28.360023 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:28 crc kubenswrapper[4824]: I1211 02:01:28.360044 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:28 crc kubenswrapper[4824]: I1211 02:01:28.360066 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:28 crc kubenswrapper[4824]: I1211 02:01:28.360081 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:28Z","lastTransitionTime":"2025-12-11T02:01:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:28 crc kubenswrapper[4824]: I1211 02:01:28.462375 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:28 crc kubenswrapper[4824]: I1211 02:01:28.462427 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:28 crc kubenswrapper[4824]: I1211 02:01:28.462437 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:28 crc kubenswrapper[4824]: I1211 02:01:28.462452 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:28 crc kubenswrapper[4824]: I1211 02:01:28.462463 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:28Z","lastTransitionTime":"2025-12-11T02:01:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:28 crc kubenswrapper[4824]: I1211 02:01:28.565276 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:28 crc kubenswrapper[4824]: I1211 02:01:28.565355 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:28 crc kubenswrapper[4824]: I1211 02:01:28.565390 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:28 crc kubenswrapper[4824]: I1211 02:01:28.565419 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:28 crc kubenswrapper[4824]: I1211 02:01:28.565436 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:28Z","lastTransitionTime":"2025-12-11T02:01:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:28 crc kubenswrapper[4824]: I1211 02:01:28.643510 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"253d4291-2b2f-487c-8862-3d305f0bdc55\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44b29bb7a6eda7cb890d36f86375b5f3f197ae5cf390f16e71139b0904873e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://922af7f67b99a0ce08776f3f75c0c377e6ea69af1720b6178a6d1e2f9b4b8ba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://136c68020325db8e6930d0f40b0d0d7e013c14c4dd5e7d1449d01d6d458104d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c89afaf56b4a9597f6543ec2e3b176c7a6dab7dc341941fc3bc8ec85cc4edaf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:00:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:28Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:28 crc kubenswrapper[4824]: I1211 02:01:28.657579 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:28Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:28 crc kubenswrapper[4824]: I1211 02:01:28.668134 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:28 crc kubenswrapper[4824]: I1211 02:01:28.668177 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:28 crc kubenswrapper[4824]: I1211 02:01:28.668189 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:28 crc kubenswrapper[4824]: I1211 02:01:28.668207 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:28 crc kubenswrapper[4824]: I1211 02:01:28.668219 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:28Z","lastTransitionTime":"2025-12-11T02:01:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:28 crc kubenswrapper[4824]: I1211 02:01:28.671703 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"beab386b-6c65-4bc3-9ef0-dc3f68b4403f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c26f06d2ed5e6e98b61a51f073130675f2c49acda106b8aedcbb2a6d4f7580c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20af9b10a67c8b21a7abcd23e42bc412798d159c8422e7c8ff58f5b997ca19a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9bbfc6edfe050b3fa9dbeee5f98d416bcb2a1533a38608a881b75396fd0b059\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7c766f8de73ca762ffde0c18b2d1c46bb152b2b37535a611c5a0e454a21b7d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d45819f0611c91b20c83eeee1092031a3441fb69f3015241632ad419880227d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 02:01:11.070250 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 02:01:11.071673 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2752882565/tls.crt::/tmp/serving-cert-2752882565/tls.key\\\\\\\"\\\\nI1211 02:01:17.042610 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 02:01:17.048046 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 02:01:17.048080 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 02:01:17.048172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 02:01:17.048188 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 02:01:17.058248 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 02:01:17.058294 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 02:01:17.058303 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 02:01:17.058315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 02:01:17.058322 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 02:01:17.058328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 02:01:17.058335 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 02:01:17.058808 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 02:01:17.060225 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14a940a72b2404a6cbd52a69e150b03266bcd7f5c1d3f4d07b51c0c69e51786\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:00:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:28Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:28 crc kubenswrapper[4824]: I1211 02:01:28.683172 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ded146c045085be2b58821749052413a36850ede00ae1024fe2dae9fa5ac84b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:28Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:28 crc kubenswrapper[4824]: I1211 02:01:28.722924 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76b10bbd1e891b8c682f4b2e17d023fd787a8ed44ed94fbf7d97314d641ccf36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f32838f33832b9b67e7dd9c53340680e17721cf1c7bdf0170230c5f8feafe64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:28Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:28 crc kubenswrapper[4824]: I1211 02:01:28.739715 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kptws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68568dba-83a5-46c1-8d01-337359679d91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f965a6d6db619995132f61b5d51b66f0b2b3fdbef3fbec63e231b9a62ccd0a9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd7nf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kptws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:28Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:28 crc kubenswrapper[4824]: I1211 02:01:28.753860 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:28Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:28 crc kubenswrapper[4824]: I1211 02:01:28.770881 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:28 crc kubenswrapper[4824]: I1211 02:01:28.770915 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:28 crc kubenswrapper[4824]: I1211 02:01:28.770926 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:28 crc kubenswrapper[4824]: I1211 02:01:28.770944 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:28 crc kubenswrapper[4824]: I1211 02:01:28.770959 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:28Z","lastTransitionTime":"2025-12-11T02:01:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:28 crc kubenswrapper[4824]: I1211 02:01:28.776399 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kh8g9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"50c69d41-a65b-463e-a829-ea349bd2fe46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dca6be41bd7a3ce12ecd4567c868f696b65e8eb1bad436feda6e4f9bfb8e525\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69af1f33fd0bafc929b94e702b4f46c6c7d7a362d5552b121729c477892d91d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69af1f33fd0bafc929b94e702b4f46c6c7d7a362d5552b121729c477892d91d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd9675d0afded587bc660a3862fd0c5dd6885961afacda957aa9ee8501d68a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd9675d0afded587bc660a3862fd0c5dd6885961afacda957aa9ee8501d68a69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aac94149458bd6321be90b26b48ee26428d273e7f61407d95f9c9d6b3aeae4b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aac94149458bd6321be90b26b48ee26428d273e7f61407d95f9c9d6b3aeae4b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbee11bd7cb65e031415990869b52d2b8f81bf8b56e6cc935d38e09695db7b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbee11bd7cb65e031415990869b52d2b8f81bf8b56e6cc935d38e09695db7b48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86f693ade72bd593bc16021fe4b7564f1914ec2f1cda7039474ac2fb13de10f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86f693ade72bd593bc16021fe4b7564f1914ec2f1cda7039474ac2fb13de10f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7d7674252f67bde0ded2ec8ad1d925a92a26d3e3f741dee0061b735cb0b7a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7d7674252f67bde0ded2ec8ad1d925a92a26d3e3f741dee0061b735cb0b7a02\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kh8g9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:28Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:28 crc kubenswrapper[4824]: I1211 02:01:28.791917 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6bkc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8022f696-d0e7-437b-848a-3cd25bd1f364\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3e28f6778ec0dba6ea3e6b69e5a2c1c73c0d5db1abeead241fb58efdbfaefe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hh5kn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6bkc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:28Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:28 crc kubenswrapper[4824]: I1211 02:01:28.812433 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b193bd2-b536-4056-92f1-94c9836ab2eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51f3c41d86ddc44f7eca118029876ad353317f753d0ad173ed4d30c2a15f10e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c345850891744a0240781277e9a89a59f6899c4c944995f62e44650f807440a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed2ae40b1fe28111efc76d9d72345edecb5f3ea34522b238f3730812b29f2bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce7aa078800e6f299cfe03770fa7e3b64fa3742fe0c0f383c87d84cf48569ddf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6aa9fc1f1bff192c69bd59d32a05d4a72d09641d9ce4b265bc4517147a7e15dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b4a3bc50eb80777188206adab774e8acad8b2e4952a8415f91b1c29ba9a0ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a49b085068cac81453c9619839fab81c33d4d6788bb92cb7f36ba450e8659508\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a49b085068cac81453c9619839fab81c33d4d6788bb92cb7f36ba450e8659508\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T02:01:27Z\\\",\\\"message\\\":\\\" 6078 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1211 02:01:26.910738 6078 handler.go:208] Removed *v1.Node event handler 2\\\\nI1211 02:01:26.910745 6078 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1211 02:01:26.910827 6078 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1211 02:01:26.911133 6078 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1211 02:01:26.911294 6078 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1211 02:01:26.911525 6078 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1211 02:01:26.911572 6078 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1211 02:01:26.911601 6078 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1211 02:01:26.911644 6078 factory.go:656] Stopping watch factory\\\\nI1211 02:01:26.911667 6078 ovnkube.go:599] Stopped ovnkube\\\\nI1211 02:01:26.911699 6078 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1211 02:01:26.911719 6078 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1211 02:01:2\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ead58694813806d307bbbc6a7b3cf41730d1c059b231a3002684bf09c10c7f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f6pgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:28Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:28 crc kubenswrapper[4824]: I1211 02:01:28.823796 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-br4j9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce125194-9c9d-4ea4-8a72-2c30d5722295\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9798f904052e08196a0db1745060dab11e2710308bca01ff160475910f2665df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fc5s8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-br4j9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:28Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:28 crc kubenswrapper[4824]: I1211 02:01:28.835593 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d14986666febd651039afb52ddf4db8969aca55e88eda9bbb20a67e35ba722ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:28Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:28 crc kubenswrapper[4824]: I1211 02:01:28.846641 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:28Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:28 crc kubenswrapper[4824]: I1211 02:01:28.859044 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44b156e4-64a4-4d45-aa5e-9b10a862faed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7368d92144b188e95a7796fabf497b0f8d9c0b240293bdc7b4aec1ff75497bc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w62dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcf289ac27dbda4f24aabbde3fb5119329e7b4fc7d6e018168da9a69661a4c57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w62dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gx6xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:28Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:28 crc kubenswrapper[4824]: I1211 02:01:28.872966 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:28 crc kubenswrapper[4824]: I1211 02:01:28.872993 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:28 crc kubenswrapper[4824]: I1211 02:01:28.873001 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:28 crc kubenswrapper[4824]: I1211 02:01:28.873015 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:28 crc kubenswrapper[4824]: I1211 02:01:28.873025 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:28Z","lastTransitionTime":"2025-12-11T02:01:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:28 crc kubenswrapper[4824]: I1211 02:01:28.973996 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f6pgc_5b193bd2-b536-4056-92f1-94c9836ab2eb/ovnkube-controller/0.log" Dec 11 02:01:28 crc kubenswrapper[4824]: I1211 02:01:28.974574 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:28 crc kubenswrapper[4824]: I1211 02:01:28.974626 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:28 crc kubenswrapper[4824]: I1211 02:01:28.974641 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:28 crc kubenswrapper[4824]: I1211 02:01:28.974656 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:28 crc kubenswrapper[4824]: I1211 02:01:28.974668 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:28Z","lastTransitionTime":"2025-12-11T02:01:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:28 crc kubenswrapper[4824]: I1211 02:01:28.982471 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" event={"ID":"5b193bd2-b536-4056-92f1-94c9836ab2eb","Type":"ContainerStarted","Data":"c803a25bdafcd4b9b8a4dfbcf8b67d6ade84b490f9d57f46873c32f9c763401d"} Dec 11 02:01:28 crc kubenswrapper[4824]: I1211 02:01:28.982626 4824 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 11 02:01:29 crc kubenswrapper[4824]: I1211 02:01:29.002093 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kh8g9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"50c69d41-a65b-463e-a829-ea349bd2fe46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dca6be41bd7a3ce12ecd4567c868f696b65e8eb1bad436feda6e4f9bfb8e525\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69af1f33fd0bafc929b94e702b4f46c6c7d7a362d5552b121729c477892d91d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69af1f33fd0bafc929b94e702b4f46c6c7d7a362d5552b121729c477892d91d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd9675d0afded587bc660a3862fd0c5dd6885961afacda957aa9ee8501d68a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd9675d0afded587bc660a3862fd0c5dd6885961afacda957aa9ee8501d68a69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aac94149458bd6321be90b26b48ee26428d273e7f61407d95f9c9d6b3aeae4b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aac94149458bd6321be90b26b48ee26428d273e7f61407d95f9c9d6b3aeae4b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbee11bd7cb65e031415990869b52d2b8f81bf8b56e6cc935d38e09695db7b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbee11bd7cb65e031415990869b52d2b8f81bf8b56e6cc935d38e09695db7b48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86f693ade72bd593bc16021fe4b7564f1914ec2f1cda7039474ac2fb13de10f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86f693ade72bd593bc16021fe4b7564f1914ec2f1cda7039474ac2fb13de10f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7d7674252f67bde0ded2ec8ad1d925a92a26d3e3f741dee0061b735cb0b7a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7d7674252f67bde0ded2ec8ad1d925a92a26d3e3f741dee0061b735cb0b7a02\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kh8g9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:28Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:29 crc kubenswrapper[4824]: I1211 02:01:29.023994 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6bkc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8022f696-d0e7-437b-848a-3cd25bd1f364\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3e28f6778ec0dba6ea3e6b69e5a2c1c73c0d5db1abeead241fb58efdbfaefe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hh5kn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6bkc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:29Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:29 crc kubenswrapper[4824]: I1211 02:01:29.053295 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b193bd2-b536-4056-92f1-94c9836ab2eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51f3c41d86ddc44f7eca118029876ad353317f753d0ad173ed4d30c2a15f10e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c345850891744a0240781277e9a89a59f6899c4c944995f62e44650f807440a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed2ae40b1fe28111efc76d9d72345edecb5f3ea34522b238f3730812b29f2bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce7aa078800e6f299cfe03770fa7e3b64fa3742fe0c0f383c87d84cf48569ddf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6aa9fc1f1bff192c69bd59d32a05d4a72d09641d9ce4b265bc4517147a7e15dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b4a3bc50eb80777188206adab774e8acad8b2e4952a8415f91b1c29ba9a0ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c803a25bdafcd4b9b8a4dfbcf8b67d6ade84b490f9d57f46873c32f9c763401d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a49b085068cac81453c9619839fab81c33d4d6788bb92cb7f36ba450e8659508\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T02:01:27Z\\\",\\\"message\\\":\\\" 6078 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1211 02:01:26.910738 6078 handler.go:208] Removed *v1.Node event handler 2\\\\nI1211 02:01:26.910745 6078 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1211 02:01:26.910827 6078 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1211 02:01:26.911133 6078 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1211 02:01:26.911294 6078 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1211 02:01:26.911525 6078 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1211 02:01:26.911572 6078 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1211 02:01:26.911601 6078 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1211 02:01:26.911644 6078 factory.go:656] Stopping watch factory\\\\nI1211 02:01:26.911667 6078 ovnkube.go:599] Stopped ovnkube\\\\nI1211 02:01:26.911699 6078 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1211 02:01:26.911719 6078 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1211 02:01:2\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:24Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ead58694813806d307bbbc6a7b3cf41730d1c059b231a3002684bf09c10c7f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f6pgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:29Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:29 crc kubenswrapper[4824]: I1211 02:01:29.069272 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-br4j9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce125194-9c9d-4ea4-8a72-2c30d5722295\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9798f904052e08196a0db1745060dab11e2710308bca01ff160475910f2665df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fc5s8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-br4j9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:29Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:29 crc kubenswrapper[4824]: I1211 02:01:29.077674 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:29 crc kubenswrapper[4824]: I1211 02:01:29.077721 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:29 crc kubenswrapper[4824]: I1211 02:01:29.077731 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:29 crc kubenswrapper[4824]: I1211 02:01:29.077746 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:29 crc kubenswrapper[4824]: I1211 02:01:29.077757 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:29Z","lastTransitionTime":"2025-12-11T02:01:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:29 crc kubenswrapper[4824]: I1211 02:01:29.093754 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:29Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:29 crc kubenswrapper[4824]: I1211 02:01:29.112029 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:29Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:29 crc kubenswrapper[4824]: I1211 02:01:29.126186 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44b156e4-64a4-4d45-aa5e-9b10a862faed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7368d92144b188e95a7796fabf497b0f8d9c0b240293bdc7b4aec1ff75497bc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w62dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcf289ac27dbda4f24aabbde3fb5119329e7b4fc7d6e018168da9a69661a4c57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w62dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gx6xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:29Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:29 crc kubenswrapper[4824]: I1211 02:01:29.139319 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d14986666febd651039afb52ddf4db8969aca55e88eda9bbb20a67e35ba722ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:29Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:29 crc kubenswrapper[4824]: I1211 02:01:29.158276 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"253d4291-2b2f-487c-8862-3d305f0bdc55\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44b29bb7a6eda7cb890d36f86375b5f3f197ae5cf390f16e71139b0904873e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://922af7f67b99a0ce08776f3f75c0c377e6ea69af1720b6178a6d1e2f9b4b8ba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://136c68020325db8e6930d0f40b0d0d7e013c14c4dd5e7d1449d01d6d458104d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c89afaf56b4a9597f6543ec2e3b176c7a6dab7dc341941fc3bc8ec85cc4edaf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:00:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:29Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:29 crc kubenswrapper[4824]: I1211 02:01:29.176956 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:29Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:29 crc kubenswrapper[4824]: I1211 02:01:29.179740 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:29 crc kubenswrapper[4824]: I1211 02:01:29.179805 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:29 crc kubenswrapper[4824]: I1211 02:01:29.179825 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:29 crc kubenswrapper[4824]: I1211 02:01:29.179854 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:29 crc kubenswrapper[4824]: I1211 02:01:29.179872 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:29Z","lastTransitionTime":"2025-12-11T02:01:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:29 crc kubenswrapper[4824]: I1211 02:01:29.198815 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"beab386b-6c65-4bc3-9ef0-dc3f68b4403f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c26f06d2ed5e6e98b61a51f073130675f2c49acda106b8aedcbb2a6d4f7580c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20af9b10a67c8b21a7abcd23e42bc412798d159c8422e7c8ff58f5b997ca19a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9bbfc6edfe050b3fa9dbeee5f98d416bcb2a1533a38608a881b75396fd0b059\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7c766f8de73ca762ffde0c18b2d1c46bb152b2b37535a611c5a0e454a21b7d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d45819f0611c91b20c83eeee1092031a3441fb69f3015241632ad419880227d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 02:01:11.070250 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 02:01:11.071673 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2752882565/tls.crt::/tmp/serving-cert-2752882565/tls.key\\\\\\\"\\\\nI1211 02:01:17.042610 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 02:01:17.048046 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 02:01:17.048080 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 02:01:17.048172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 02:01:17.048188 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 02:01:17.058248 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 02:01:17.058294 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 02:01:17.058303 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 02:01:17.058315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 02:01:17.058322 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 02:01:17.058328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 02:01:17.058335 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 02:01:17.058808 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 02:01:17.060225 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14a940a72b2404a6cbd52a69e150b03266bcd7f5c1d3f4d07b51c0c69e51786\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:00:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:29Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:29 crc kubenswrapper[4824]: I1211 02:01:29.221454 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ded146c045085be2b58821749052413a36850ede00ae1024fe2dae9fa5ac84b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:29Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:29 crc kubenswrapper[4824]: I1211 02:01:29.237648 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76b10bbd1e891b8c682f4b2e17d023fd787a8ed44ed94fbf7d97314d641ccf36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f32838f33832b9b67e7dd9c53340680e17721cf1c7bdf0170230c5f8feafe64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:29Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:29 crc kubenswrapper[4824]: I1211 02:01:29.249415 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kptws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68568dba-83a5-46c1-8d01-337359679d91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f965a6d6db619995132f61b5d51b66f0b2b3fdbef3fbec63e231b9a62ccd0a9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd7nf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kptws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:29Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:29 crc kubenswrapper[4824]: I1211 02:01:29.282887 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:29 crc kubenswrapper[4824]: I1211 02:01:29.282951 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:29 crc kubenswrapper[4824]: I1211 02:01:29.282969 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:29 crc kubenswrapper[4824]: I1211 02:01:29.282997 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:29 crc kubenswrapper[4824]: I1211 02:01:29.283014 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:29Z","lastTransitionTime":"2025-12-11T02:01:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:29 crc kubenswrapper[4824]: I1211 02:01:29.386029 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:29 crc kubenswrapper[4824]: I1211 02:01:29.386092 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:29 crc kubenswrapper[4824]: I1211 02:01:29.386141 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:29 crc kubenswrapper[4824]: I1211 02:01:29.386191 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:29 crc kubenswrapper[4824]: I1211 02:01:29.386208 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:29Z","lastTransitionTime":"2025-12-11T02:01:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:29 crc kubenswrapper[4824]: I1211 02:01:29.488936 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:29 crc kubenswrapper[4824]: I1211 02:01:29.488974 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:29 crc kubenswrapper[4824]: I1211 02:01:29.488985 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:29 crc kubenswrapper[4824]: I1211 02:01:29.488999 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:29 crc kubenswrapper[4824]: I1211 02:01:29.489008 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:29Z","lastTransitionTime":"2025-12-11T02:01:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:29 crc kubenswrapper[4824]: I1211 02:01:29.592264 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:29 crc kubenswrapper[4824]: I1211 02:01:29.592317 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:29 crc kubenswrapper[4824]: I1211 02:01:29.592339 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:29 crc kubenswrapper[4824]: I1211 02:01:29.592365 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:29 crc kubenswrapper[4824]: I1211 02:01:29.592386 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:29Z","lastTransitionTime":"2025-12-11T02:01:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:29 crc kubenswrapper[4824]: I1211 02:01:29.632039 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 02:01:29 crc kubenswrapper[4824]: I1211 02:01:29.632065 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 02:01:29 crc kubenswrapper[4824]: E1211 02:01:29.632272 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 02:01:29 crc kubenswrapper[4824]: E1211 02:01:29.632463 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 02:01:29 crc kubenswrapper[4824]: I1211 02:01:29.632074 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 02:01:29 crc kubenswrapper[4824]: E1211 02:01:29.632918 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 02:01:29 crc kubenswrapper[4824]: I1211 02:01:29.695586 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:29 crc kubenswrapper[4824]: I1211 02:01:29.695648 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:29 crc kubenswrapper[4824]: I1211 02:01:29.695668 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:29 crc kubenswrapper[4824]: I1211 02:01:29.695695 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:29 crc kubenswrapper[4824]: I1211 02:01:29.695715 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:29Z","lastTransitionTime":"2025-12-11T02:01:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:29 crc kubenswrapper[4824]: I1211 02:01:29.799057 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:29 crc kubenswrapper[4824]: I1211 02:01:29.799190 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:29 crc kubenswrapper[4824]: I1211 02:01:29.799219 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:29 crc kubenswrapper[4824]: I1211 02:01:29.799246 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:29 crc kubenswrapper[4824]: I1211 02:01:29.799264 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:29Z","lastTransitionTime":"2025-12-11T02:01:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:29 crc kubenswrapper[4824]: I1211 02:01:29.901854 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:29 crc kubenswrapper[4824]: I1211 02:01:29.901917 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:29 crc kubenswrapper[4824]: I1211 02:01:29.901940 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:29 crc kubenswrapper[4824]: I1211 02:01:29.901967 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:29 crc kubenswrapper[4824]: I1211 02:01:29.901987 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:29Z","lastTransitionTime":"2025-12-11T02:01:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:29 crc kubenswrapper[4824]: I1211 02:01:29.989389 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f6pgc_5b193bd2-b536-4056-92f1-94c9836ab2eb/ovnkube-controller/1.log" Dec 11 02:01:29 crc kubenswrapper[4824]: I1211 02:01:29.990430 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f6pgc_5b193bd2-b536-4056-92f1-94c9836ab2eb/ovnkube-controller/0.log" Dec 11 02:01:29 crc kubenswrapper[4824]: I1211 02:01:29.995775 4824 generic.go:334] "Generic (PLEG): container finished" podID="5b193bd2-b536-4056-92f1-94c9836ab2eb" containerID="c803a25bdafcd4b9b8a4dfbcf8b67d6ade84b490f9d57f46873c32f9c763401d" exitCode=1 Dec 11 02:01:29 crc kubenswrapper[4824]: I1211 02:01:29.995843 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" event={"ID":"5b193bd2-b536-4056-92f1-94c9836ab2eb","Type":"ContainerDied","Data":"c803a25bdafcd4b9b8a4dfbcf8b67d6ade84b490f9d57f46873c32f9c763401d"} Dec 11 02:01:29 crc kubenswrapper[4824]: I1211 02:01:29.995922 4824 scope.go:117] "RemoveContainer" containerID="a49b085068cac81453c9619839fab81c33d4d6788bb92cb7f36ba450e8659508" Dec 11 02:01:29 crc kubenswrapper[4824]: I1211 02:01:29.997684 4824 scope.go:117] "RemoveContainer" containerID="c803a25bdafcd4b9b8a4dfbcf8b67d6ade84b490f9d57f46873c32f9c763401d" Dec 11 02:01:29 crc kubenswrapper[4824]: E1211 02:01:29.997967 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-f6pgc_openshift-ovn-kubernetes(5b193bd2-b536-4056-92f1-94c9836ab2eb)\"" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" podUID="5b193bd2-b536-4056-92f1-94c9836ab2eb" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.005701 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.005732 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.005741 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.005753 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.005763 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:30Z","lastTransitionTime":"2025-12-11T02:01:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.022574 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"beab386b-6c65-4bc3-9ef0-dc3f68b4403f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c26f06d2ed5e6e98b61a51f073130675f2c49acda106b8aedcbb2a6d4f7580c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20af9b10a67c8b21a7abcd23e42bc412798d159c8422e7c8ff58f5b997ca19a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9bbfc6edfe050b3fa9dbeee5f98d416bcb2a1533a38608a881b75396fd0b059\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7c766f8de73ca762ffde0c18b2d1c46bb152b2b37535a611c5a0e454a21b7d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d45819f0611c91b20c83eeee1092031a3441fb69f3015241632ad419880227d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 02:01:11.070250 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 02:01:11.071673 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2752882565/tls.crt::/tmp/serving-cert-2752882565/tls.key\\\\\\\"\\\\nI1211 02:01:17.042610 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 02:01:17.048046 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 02:01:17.048080 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 02:01:17.048172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 02:01:17.048188 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 02:01:17.058248 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 02:01:17.058294 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 02:01:17.058303 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 02:01:17.058315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 02:01:17.058322 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 02:01:17.058328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 02:01:17.058335 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 02:01:17.058808 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 02:01:17.060225 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14a940a72b2404a6cbd52a69e150b03266bcd7f5c1d3f4d07b51c0c69e51786\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:00:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:30Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.048418 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ded146c045085be2b58821749052413a36850ede00ae1024fe2dae9fa5ac84b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:30Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.072898 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76b10bbd1e891b8c682f4b2e17d023fd787a8ed44ed94fbf7d97314d641ccf36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f32838f33832b9b67e7dd9c53340680e17721cf1c7bdf0170230c5f8feafe64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:30Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.092936 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kptws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68568dba-83a5-46c1-8d01-337359679d91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f965a6d6db619995132f61b5d51b66f0b2b3fdbef3fbec63e231b9a62ccd0a9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd7nf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kptws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:30Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.108797 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.108860 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.108883 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.108908 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.108926 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:30Z","lastTransitionTime":"2025-12-11T02:01:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.111030 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:30Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.131891 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kh8g9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"50c69d41-a65b-463e-a829-ea349bd2fe46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dca6be41bd7a3ce12ecd4567c868f696b65e8eb1bad436feda6e4f9bfb8e525\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69af1f33fd0bafc929b94e702b4f46c6c7d7a362d5552b121729c477892d91d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69af1f33fd0bafc929b94e702b4f46c6c7d7a362d5552b121729c477892d91d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd9675d0afded587bc660a3862fd0c5dd6885961afacda957aa9ee8501d68a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd9675d0afded587bc660a3862fd0c5dd6885961afacda957aa9ee8501d68a69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aac94149458bd6321be90b26b48ee26428d273e7f61407d95f9c9d6b3aeae4b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aac94149458bd6321be90b26b48ee26428d273e7f61407d95f9c9d6b3aeae4b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbee11bd7cb65e031415990869b52d2b8f81bf8b56e6cc935d38e09695db7b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbee11bd7cb65e031415990869b52d2b8f81bf8b56e6cc935d38e09695db7b48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86f693ade72bd593bc16021fe4b7564f1914ec2f1cda7039474ac2fb13de10f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86f693ade72bd593bc16021fe4b7564f1914ec2f1cda7039474ac2fb13de10f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7d7674252f67bde0ded2ec8ad1d925a92a26d3e3f741dee0061b735cb0b7a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7d7674252f67bde0ded2ec8ad1d925a92a26d3e3f741dee0061b735cb0b7a02\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kh8g9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:30Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.150058 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6bkc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8022f696-d0e7-437b-848a-3cd25bd1f364\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3e28f6778ec0dba6ea3e6b69e5a2c1c73c0d5db1abeead241fb58efdbfaefe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hh5kn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6bkc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:30Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.172870 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b193bd2-b536-4056-92f1-94c9836ab2eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51f3c41d86ddc44f7eca118029876ad353317f753d0ad173ed4d30c2a15f10e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c345850891744a0240781277e9a89a59f6899c4c944995f62e44650f807440a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed2ae40b1fe28111efc76d9d72345edecb5f3ea34522b238f3730812b29f2bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce7aa078800e6f299cfe03770fa7e3b64fa3742fe0c0f383c87d84cf48569ddf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6aa9fc1f1bff192c69bd59d32a05d4a72d09641d9ce4b265bc4517147a7e15dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b4a3bc50eb80777188206adab774e8acad8b2e4952a8415f91b1c29ba9a0ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c803a25bdafcd4b9b8a4dfbcf8b67d6ade84b490f9d57f46873c32f9c763401d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a49b085068cac81453c9619839fab81c33d4d6788bb92cb7f36ba450e8659508\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T02:01:27Z\\\",\\\"message\\\":\\\" 6078 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1211 02:01:26.910738 6078 handler.go:208] Removed *v1.Node event handler 2\\\\nI1211 02:01:26.910745 6078 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1211 02:01:26.910827 6078 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1211 02:01:26.911133 6078 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1211 02:01:26.911294 6078 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1211 02:01:26.911525 6078 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1211 02:01:26.911572 6078 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1211 02:01:26.911601 6078 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1211 02:01:26.911644 6078 factory.go:656] Stopping watch factory\\\\nI1211 02:01:26.911667 6078 ovnkube.go:599] Stopped ovnkube\\\\nI1211 02:01:26.911699 6078 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1211 02:01:26.911719 6078 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1211 02:01:2\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:24Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c803a25bdafcd4b9b8a4dfbcf8b67d6ade84b490f9d57f46873c32f9c763401d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T02:01:29Z\\\",\\\"message\\\":\\\"ode_controller event handlers\\\\nI1211 02:01:29.024367 6208 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1211 02:01:29.024404 6208 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1211 02:01:29.024418 6208 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1211 02:01:29.024439 6208 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1211 02:01:29.024462 6208 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1211 02:01:29.024470 6208 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1211 02:01:29.024492 6208 factory.go:656] Stopping watch factory\\\\nI1211 02:01:29.024509 6208 ovnkube.go:599] Stopped ovnkube\\\\nI1211 02:01:29.024540 6208 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1211 02:01:29.024559 6208 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1211 02:01:29.024571 6208 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1211 02:01:29.024582 6208 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1211 02:01:29.024593 6208 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1211 02:01:29.024604 6208 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1211 02:01:29.024620 6208 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1211 02:01:29.024705 6208 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ead58694813806d307bbbc6a7b3cf41730d1c059b231a3002684bf09c10c7f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f6pgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:30Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.188947 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-br4j9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce125194-9c9d-4ea4-8a72-2c30d5722295\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9798f904052e08196a0db1745060dab11e2710308bca01ff160475910f2665df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fc5s8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-br4j9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:30Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.207191 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d14986666febd651039afb52ddf4db8969aca55e88eda9bbb20a67e35ba722ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:30Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.211514 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.211641 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.211670 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.211702 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.211727 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:30Z","lastTransitionTime":"2025-12-11T02:01:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.224773 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:30Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.241048 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44b156e4-64a4-4d45-aa5e-9b10a862faed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7368d92144b188e95a7796fabf497b0f8d9c0b240293bdc7b4aec1ff75497bc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w62dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcf289ac27dbda4f24aabbde3fb5119329e7b4fc7d6e018168da9a69661a4c57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w62dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gx6xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:30Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.257734 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"253d4291-2b2f-487c-8862-3d305f0bdc55\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44b29bb7a6eda7cb890d36f86375b5f3f197ae5cf390f16e71139b0904873e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://922af7f67b99a0ce08776f3f75c0c377e6ea69af1720b6178a6d1e2f9b4b8ba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://136c68020325db8e6930d0f40b0d0d7e013c14c4dd5e7d1449d01d6d458104d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c89afaf56b4a9597f6543ec2e3b176c7a6dab7dc341941fc3bc8ec85cc4edaf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:00:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:30Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.275004 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:30Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.314407 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.314507 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.314533 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.314563 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.314587 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:30Z","lastTransitionTime":"2025-12-11T02:01:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.417662 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.417735 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.417761 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.417793 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.417818 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:30Z","lastTransitionTime":"2025-12-11T02:01:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.520702 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.520738 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.520748 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.520763 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.520771 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:30Z","lastTransitionTime":"2025-12-11T02:01:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.579789 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8v4jn"] Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.580827 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8v4jn" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.583574 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f4hk4\" (UniqueName: \"kubernetes.io/projected/99a6bfee-afca-4726-a964-b95a948945a4-kube-api-access-f4hk4\") pod \"ovnkube-control-plane-749d76644c-8v4jn\" (UID: \"99a6bfee-afca-4726-a964-b95a948945a4\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8v4jn" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.583636 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/99a6bfee-afca-4726-a964-b95a948945a4-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-8v4jn\" (UID: \"99a6bfee-afca-4726-a964-b95a948945a4\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8v4jn" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.583671 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/99a6bfee-afca-4726-a964-b95a948945a4-env-overrides\") pod \"ovnkube-control-plane-749d76644c-8v4jn\" (UID: \"99a6bfee-afca-4726-a964-b95a948945a4\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8v4jn" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.583725 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/99a6bfee-afca-4726-a964-b95a948945a4-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-8v4jn\" (UID: \"99a6bfee-afca-4726-a964-b95a948945a4\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8v4jn" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.584097 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.584600 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.600449 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44b156e4-64a4-4d45-aa5e-9b10a862faed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7368d92144b188e95a7796fabf497b0f8d9c0b240293bdc7b4aec1ff75497bc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w62dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcf289ac27dbda4f24aabbde3fb5119329e7b4fc7d6e018168da9a69661a4c57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w62dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gx6xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:30Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.613900 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d14986666febd651039afb52ddf4db8969aca55e88eda9bbb20a67e35ba722ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:30Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.624009 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.624047 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.624062 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.624082 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.624095 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:30Z","lastTransitionTime":"2025-12-11T02:01:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.628464 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.628507 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.628518 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.628571 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.628604 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:30Z","lastTransitionTime":"2025-12-11T02:01:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.630451 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:30Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.646941 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:30Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:30 crc kubenswrapper[4824]: E1211 02:01:30.647411 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:01:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:01:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:01:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:01:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d57bece6-6fc6-4a84-b8a8-e1678f9e50db\\\",\\\"systemUUID\\\":\\\"f76b32e4-2764-4219-bbf4-f7e56d76df3b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:30Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.651379 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.651475 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.651507 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.651533 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.651552 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:30Z","lastTransitionTime":"2025-12-11T02:01:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.666948 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"253d4291-2b2f-487c-8862-3d305f0bdc55\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44b29bb7a6eda7cb890d36f86375b5f3f197ae5cf390f16e71139b0904873e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://922af7f67b99a0ce08776f3f75c0c377e6ea69af1720b6178a6d1e2f9b4b8ba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://136c68020325db8e6930d0f40b0d0d7e013c14c4dd5e7d1449d01d6d458104d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c89afaf56b4a9597f6543ec2e3b176c7a6dab7dc341941fc3bc8ec85cc4edaf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:00:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:30Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:30 crc kubenswrapper[4824]: E1211 02:01:30.672345 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:01:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:01:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:01:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:01:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d57bece6-6fc6-4a84-b8a8-e1678f9e50db\\\",\\\"systemUUID\\\":\\\"f76b32e4-2764-4219-bbf4-f7e56d76df3b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:30Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.676470 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.676541 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.676560 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.676586 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.676606 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:30Z","lastTransitionTime":"2025-12-11T02:01:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.685210 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/99a6bfee-afca-4726-a964-b95a948945a4-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-8v4jn\" (UID: \"99a6bfee-afca-4726-a964-b95a948945a4\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8v4jn" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.685253 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/99a6bfee-afca-4726-a964-b95a948945a4-env-overrides\") pod \"ovnkube-control-plane-749d76644c-8v4jn\" (UID: \"99a6bfee-afca-4726-a964-b95a948945a4\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8v4jn" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.685288 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f4hk4\" (UniqueName: \"kubernetes.io/projected/99a6bfee-afca-4726-a964-b95a948945a4-kube-api-access-f4hk4\") pod \"ovnkube-control-plane-749d76644c-8v4jn\" (UID: \"99a6bfee-afca-4726-a964-b95a948945a4\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8v4jn" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.685354 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/99a6bfee-afca-4726-a964-b95a948945a4-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-8v4jn\" (UID: \"99a6bfee-afca-4726-a964-b95a948945a4\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8v4jn" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.686278 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/99a6bfee-afca-4726-a964-b95a948945a4-env-overrides\") pod \"ovnkube-control-plane-749d76644c-8v4jn\" (UID: \"99a6bfee-afca-4726-a964-b95a948945a4\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8v4jn" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.687068 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/99a6bfee-afca-4726-a964-b95a948945a4-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-8v4jn\" (UID: \"99a6bfee-afca-4726-a964-b95a948945a4\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8v4jn" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.688799 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ded146c045085be2b58821749052413a36850ede00ae1024fe2dae9fa5ac84b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:30Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.694004 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/99a6bfee-afca-4726-a964-b95a948945a4-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-8v4jn\" (UID: \"99a6bfee-afca-4726-a964-b95a948945a4\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8v4jn" Dec 11 02:01:30 crc kubenswrapper[4824]: E1211 02:01:30.702884 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:01:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:01:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:01:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:01:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d57bece6-6fc6-4a84-b8a8-e1678f9e50db\\\",\\\"systemUUID\\\":\\\"f76b32e4-2764-4219-bbf4-f7e56d76df3b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:30Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.707498 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.707616 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.707524 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76b10bbd1e891b8c682f4b2e17d023fd787a8ed44ed94fbf7d97314d641ccf36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f32838f33832b9b67e7dd9c53340680e17721cf1c7bdf0170230c5f8feafe64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:30Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.707692 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.707928 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.707949 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:30Z","lastTransitionTime":"2025-12-11T02:01:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.709178 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f4hk4\" (UniqueName: \"kubernetes.io/projected/99a6bfee-afca-4726-a964-b95a948945a4-kube-api-access-f4hk4\") pod \"ovnkube-control-plane-749d76644c-8v4jn\" (UID: \"99a6bfee-afca-4726-a964-b95a948945a4\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8v4jn" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.722030 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kptws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68568dba-83a5-46c1-8d01-337359679d91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f965a6d6db619995132f61b5d51b66f0b2b3fdbef3fbec63e231b9a62ccd0a9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd7nf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kptws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:30Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:30 crc kubenswrapper[4824]: E1211 02:01:30.725202 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:01:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:01:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:01:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:01:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d57bece6-6fc6-4a84-b8a8-e1678f9e50db\\\",\\\"systemUUID\\\":\\\"f76b32e4-2764-4219-bbf4-f7e56d76df3b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:30Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.731275 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.731328 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.731341 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.731358 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.731370 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:30Z","lastTransitionTime":"2025-12-11T02:01:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.739038 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"beab386b-6c65-4bc3-9ef0-dc3f68b4403f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c26f06d2ed5e6e98b61a51f073130675f2c49acda106b8aedcbb2a6d4f7580c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20af9b10a67c8b21a7abcd23e42bc412798d159c8422e7c8ff58f5b997ca19a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9bbfc6edfe050b3fa9dbeee5f98d416bcb2a1533a38608a881b75396fd0b059\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7c766f8de73ca762ffde0c18b2d1c46bb152b2b37535a611c5a0e454a21b7d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d45819f0611c91b20c83eeee1092031a3441fb69f3015241632ad419880227d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 02:01:11.070250 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 02:01:11.071673 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2752882565/tls.crt::/tmp/serving-cert-2752882565/tls.key\\\\\\\"\\\\nI1211 02:01:17.042610 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 02:01:17.048046 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 02:01:17.048080 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 02:01:17.048172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 02:01:17.048188 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 02:01:17.058248 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 02:01:17.058294 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 02:01:17.058303 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 02:01:17.058315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 02:01:17.058322 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 02:01:17.058328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 02:01:17.058335 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 02:01:17.058808 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 02:01:17.060225 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14a940a72b2404a6cbd52a69e150b03266bcd7f5c1d3f4d07b51c0c69e51786\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:00:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:30Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:30 crc kubenswrapper[4824]: E1211 02:01:30.746449 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:01:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:01:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:01:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:01:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d57bece6-6fc6-4a84-b8a8-e1678f9e50db\\\",\\\"systemUUID\\\":\\\"f76b32e4-2764-4219-bbf4-f7e56d76df3b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:30Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:30 crc kubenswrapper[4824]: E1211 02:01:30.746591 4824 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.748667 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.748709 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.748720 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.748739 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.748751 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:30Z","lastTransitionTime":"2025-12-11T02:01:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.754558 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6bkc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8022f696-d0e7-437b-848a-3cd25bd1f364\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3e28f6778ec0dba6ea3e6b69e5a2c1c73c0d5db1abeead241fb58efdbfaefe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hh5kn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6bkc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:30Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.776447 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b193bd2-b536-4056-92f1-94c9836ab2eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51f3c41d86ddc44f7eca118029876ad353317f753d0ad173ed4d30c2a15f10e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c345850891744a0240781277e9a89a59f6899c4c944995f62e44650f807440a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed2ae40b1fe28111efc76d9d72345edecb5f3ea34522b238f3730812b29f2bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce7aa078800e6f299cfe03770fa7e3b64fa3742fe0c0f383c87d84cf48569ddf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6aa9fc1f1bff192c69bd59d32a05d4a72d09641d9ce4b265bc4517147a7e15dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b4a3bc50eb80777188206adab774e8acad8b2e4952a8415f91b1c29ba9a0ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c803a25bdafcd4b9b8a4dfbcf8b67d6ade84b490f9d57f46873c32f9c763401d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a49b085068cac81453c9619839fab81c33d4d6788bb92cb7f36ba450e8659508\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T02:01:27Z\\\",\\\"message\\\":\\\" 6078 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1211 02:01:26.910738 6078 handler.go:208] Removed *v1.Node event handler 2\\\\nI1211 02:01:26.910745 6078 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1211 02:01:26.910827 6078 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1211 02:01:26.911133 6078 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1211 02:01:26.911294 6078 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1211 02:01:26.911525 6078 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1211 02:01:26.911572 6078 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1211 02:01:26.911601 6078 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1211 02:01:26.911644 6078 factory.go:656] Stopping watch factory\\\\nI1211 02:01:26.911667 6078 ovnkube.go:599] Stopped ovnkube\\\\nI1211 02:01:26.911699 6078 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1211 02:01:26.911719 6078 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1211 02:01:2\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:24Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c803a25bdafcd4b9b8a4dfbcf8b67d6ade84b490f9d57f46873c32f9c763401d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T02:01:29Z\\\",\\\"message\\\":\\\"ode_controller event handlers\\\\nI1211 02:01:29.024367 6208 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1211 02:01:29.024404 6208 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1211 02:01:29.024418 6208 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1211 02:01:29.024439 6208 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1211 02:01:29.024462 6208 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1211 02:01:29.024470 6208 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1211 02:01:29.024492 6208 factory.go:656] Stopping watch factory\\\\nI1211 02:01:29.024509 6208 ovnkube.go:599] Stopped ovnkube\\\\nI1211 02:01:29.024540 6208 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1211 02:01:29.024559 6208 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1211 02:01:29.024571 6208 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1211 02:01:29.024582 6208 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1211 02:01:29.024593 6208 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1211 02:01:29.024604 6208 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1211 02:01:29.024620 6208 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1211 02:01:29.024705 6208 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ead58694813806d307bbbc6a7b3cf41730d1c059b231a3002684bf09c10c7f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f6pgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:30Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.789654 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-br4j9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce125194-9c9d-4ea4-8a72-2c30d5722295\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9798f904052e08196a0db1745060dab11e2710308bca01ff160475910f2665df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fc5s8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-br4j9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:30Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.805458 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8v4jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99a6bfee-afca-4726-a964-b95a948945a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4hk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4hk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-8v4jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:30Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.823421 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:30Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.851758 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.852077 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.852324 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.852531 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.852721 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:30Z","lastTransitionTime":"2025-12-11T02:01:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.857660 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kh8g9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"50c69d41-a65b-463e-a829-ea349bd2fe46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dca6be41bd7a3ce12ecd4567c868f696b65e8eb1bad436feda6e4f9bfb8e525\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69af1f33fd0bafc929b94e702b4f46c6c7d7a362d5552b121729c477892d91d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69af1f33fd0bafc929b94e702b4f46c6c7d7a362d5552b121729c477892d91d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd9675d0afded587bc660a3862fd0c5dd6885961afacda957aa9ee8501d68a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd9675d0afded587bc660a3862fd0c5dd6885961afacda957aa9ee8501d68a69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aac94149458bd6321be90b26b48ee26428d273e7f61407d95f9c9d6b3aeae4b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aac94149458bd6321be90b26b48ee26428d273e7f61407d95f9c9d6b3aeae4b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbee11bd7cb65e031415990869b52d2b8f81bf8b56e6cc935d38e09695db7b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbee11bd7cb65e031415990869b52d2b8f81bf8b56e6cc935d38e09695db7b48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86f693ade72bd593bc16021fe4b7564f1914ec2f1cda7039474ac2fb13de10f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86f693ade72bd593bc16021fe4b7564f1914ec2f1cda7039474ac2fb13de10f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7d7674252f67bde0ded2ec8ad1d925a92a26d3e3f741dee0061b735cb0b7a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7d7674252f67bde0ded2ec8ad1d925a92a26d3e3f741dee0061b735cb0b7a02\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kh8g9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:30Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.903168 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8v4jn" Dec 11 02:01:30 crc kubenswrapper[4824]: W1211 02:01:30.923199 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod99a6bfee_afca_4726_a964_b95a948945a4.slice/crio-79fd5d1f47ee3f6e7df6b24e667ff53b0adb26a5e874fc6a48bdf8cfc554d0dc WatchSource:0}: Error finding container 79fd5d1f47ee3f6e7df6b24e667ff53b0adb26a5e874fc6a48bdf8cfc554d0dc: Status 404 returned error can't find the container with id 79fd5d1f47ee3f6e7df6b24e667ff53b0adb26a5e874fc6a48bdf8cfc554d0dc Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.958442 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.958918 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.958941 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.958970 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:30 crc kubenswrapper[4824]: I1211 02:01:30.959231 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:30Z","lastTransitionTime":"2025-12-11T02:01:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.003030 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f6pgc_5b193bd2-b536-4056-92f1-94c9836ab2eb/ovnkube-controller/1.log" Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.007979 4824 scope.go:117] "RemoveContainer" containerID="c803a25bdafcd4b9b8a4dfbcf8b67d6ade84b490f9d57f46873c32f9c763401d" Dec 11 02:01:31 crc kubenswrapper[4824]: E1211 02:01:31.008245 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-f6pgc_openshift-ovn-kubernetes(5b193bd2-b536-4056-92f1-94c9836ab2eb)\"" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" podUID="5b193bd2-b536-4056-92f1-94c9836ab2eb" Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.008833 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8v4jn" event={"ID":"99a6bfee-afca-4726-a964-b95a948945a4","Type":"ContainerStarted","Data":"79fd5d1f47ee3f6e7df6b24e667ff53b0adb26a5e874fc6a48bdf8cfc554d0dc"} Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.028784 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d14986666febd651039afb52ddf4db8969aca55e88eda9bbb20a67e35ba722ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:31Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.051254 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:31Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.062583 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.062649 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.062666 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.062694 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.062715 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:31Z","lastTransitionTime":"2025-12-11T02:01:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.069475 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44b156e4-64a4-4d45-aa5e-9b10a862faed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7368d92144b188e95a7796fabf497b0f8d9c0b240293bdc7b4aec1ff75497bc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w62dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcf289ac27dbda4f24aabbde3fb5119329e7b4fc7d6e018168da9a69661a4c57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w62dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gx6xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:31Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.093036 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"253d4291-2b2f-487c-8862-3d305f0bdc55\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44b29bb7a6eda7cb890d36f86375b5f3f197ae5cf390f16e71139b0904873e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://922af7f67b99a0ce08776f3f75c0c377e6ea69af1720b6178a6d1e2f9b4b8ba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://136c68020325db8e6930d0f40b0d0d7e013c14c4dd5e7d1449d01d6d458104d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c89afaf56b4a9597f6543ec2e3b176c7a6dab7dc341941fc3bc8ec85cc4edaf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:00:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:31Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.113175 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:31Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.134102 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"beab386b-6c65-4bc3-9ef0-dc3f68b4403f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c26f06d2ed5e6e98b61a51f073130675f2c49acda106b8aedcbb2a6d4f7580c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20af9b10a67c8b21a7abcd23e42bc412798d159c8422e7c8ff58f5b997ca19a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9bbfc6edfe050b3fa9dbeee5f98d416bcb2a1533a38608a881b75396fd0b059\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7c766f8de73ca762ffde0c18b2d1c46bb152b2b37535a611c5a0e454a21b7d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d45819f0611c91b20c83eeee1092031a3441fb69f3015241632ad419880227d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 02:01:11.070250 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 02:01:11.071673 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2752882565/tls.crt::/tmp/serving-cert-2752882565/tls.key\\\\\\\"\\\\nI1211 02:01:17.042610 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 02:01:17.048046 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 02:01:17.048080 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 02:01:17.048172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 02:01:17.048188 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 02:01:17.058248 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 02:01:17.058294 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 02:01:17.058303 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 02:01:17.058315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 02:01:17.058322 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 02:01:17.058328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 02:01:17.058335 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 02:01:17.058808 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 02:01:17.060225 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14a940a72b2404a6cbd52a69e150b03266bcd7f5c1d3f4d07b51c0c69e51786\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:00:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:31Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.154932 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ded146c045085be2b58821749052413a36850ede00ae1024fe2dae9fa5ac84b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:31Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.165828 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.165887 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.165906 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.165933 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.165955 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:31Z","lastTransitionTime":"2025-12-11T02:01:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.172290 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76b10bbd1e891b8c682f4b2e17d023fd787a8ed44ed94fbf7d97314d641ccf36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f32838f33832b9b67e7dd9c53340680e17721cf1c7bdf0170230c5f8feafe64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:31Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.185898 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kptws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68568dba-83a5-46c1-8d01-337359679d91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f965a6d6db619995132f61b5d51b66f0b2b3fdbef3fbec63e231b9a62ccd0a9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd7nf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kptws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:31Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.202599 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8v4jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99a6bfee-afca-4726-a964-b95a948945a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4hk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4hk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-8v4jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:31Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.222455 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:31Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.250423 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kh8g9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"50c69d41-a65b-463e-a829-ea349bd2fe46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dca6be41bd7a3ce12ecd4567c868f696b65e8eb1bad436feda6e4f9bfb8e525\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69af1f33fd0bafc929b94e702b4f46c6c7d7a362d5552b121729c477892d91d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69af1f33fd0bafc929b94e702b4f46c6c7d7a362d5552b121729c477892d91d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd9675d0afded587bc660a3862fd0c5dd6885961afacda957aa9ee8501d68a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd9675d0afded587bc660a3862fd0c5dd6885961afacda957aa9ee8501d68a69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aac94149458bd6321be90b26b48ee26428d273e7f61407d95f9c9d6b3aeae4b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aac94149458bd6321be90b26b48ee26428d273e7f61407d95f9c9d6b3aeae4b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbee11bd7cb65e031415990869b52d2b8f81bf8b56e6cc935d38e09695db7b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbee11bd7cb65e031415990869b52d2b8f81bf8b56e6cc935d38e09695db7b48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86f693ade72bd593bc16021fe4b7564f1914ec2f1cda7039474ac2fb13de10f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86f693ade72bd593bc16021fe4b7564f1914ec2f1cda7039474ac2fb13de10f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7d7674252f67bde0ded2ec8ad1d925a92a26d3e3f741dee0061b735cb0b7a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7d7674252f67bde0ded2ec8ad1d925a92a26d3e3f741dee0061b735cb0b7a02\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kh8g9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:31Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.268994 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.269045 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.269063 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.269087 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.269105 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:31Z","lastTransitionTime":"2025-12-11T02:01:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.273197 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6bkc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8022f696-d0e7-437b-848a-3cd25bd1f364\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3e28f6778ec0dba6ea3e6b69e5a2c1c73c0d5db1abeead241fb58efdbfaefe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hh5kn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6bkc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:31Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.298458 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b193bd2-b536-4056-92f1-94c9836ab2eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51f3c41d86ddc44f7eca118029876ad353317f753d0ad173ed4d30c2a15f10e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c345850891744a0240781277e9a89a59f6899c4c944995f62e44650f807440a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed2ae40b1fe28111efc76d9d72345edecb5f3ea34522b238f3730812b29f2bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce7aa078800e6f299cfe03770fa7e3b64fa3742fe0c0f383c87d84cf48569ddf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6aa9fc1f1bff192c69bd59d32a05d4a72d09641d9ce4b265bc4517147a7e15dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b4a3bc50eb80777188206adab774e8acad8b2e4952a8415f91b1c29ba9a0ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c803a25bdafcd4b9b8a4dfbcf8b67d6ade84b490f9d57f46873c32f9c763401d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c803a25bdafcd4b9b8a4dfbcf8b67d6ade84b490f9d57f46873c32f9c763401d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T02:01:29Z\\\",\\\"message\\\":\\\"ode_controller event handlers\\\\nI1211 02:01:29.024367 6208 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1211 02:01:29.024404 6208 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1211 02:01:29.024418 6208 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1211 02:01:29.024439 6208 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1211 02:01:29.024462 6208 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1211 02:01:29.024470 6208 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1211 02:01:29.024492 6208 factory.go:656] Stopping watch factory\\\\nI1211 02:01:29.024509 6208 ovnkube.go:599] Stopped ovnkube\\\\nI1211 02:01:29.024540 6208 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1211 02:01:29.024559 6208 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1211 02:01:29.024571 6208 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1211 02:01:29.024582 6208 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1211 02:01:29.024593 6208 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1211 02:01:29.024604 6208 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1211 02:01:29.024620 6208 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1211 02:01:29.024705 6208 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-f6pgc_openshift-ovn-kubernetes(5b193bd2-b536-4056-92f1-94c9836ab2eb)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ead58694813806d307bbbc6a7b3cf41730d1c059b231a3002684bf09c10c7f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f6pgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:31Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.310905 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-br4j9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce125194-9c9d-4ea4-8a72-2c30d5722295\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9798f904052e08196a0db1745060dab11e2710308bca01ff160475910f2665df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fc5s8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-br4j9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:31Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.372743 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.372807 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.372825 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.372855 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.372877 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:31Z","lastTransitionTime":"2025-12-11T02:01:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.475771 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.475834 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.475854 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.475880 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.475897 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:31Z","lastTransitionTime":"2025-12-11T02:01:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.578597 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.578672 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.578696 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.578728 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.578753 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:31Z","lastTransitionTime":"2025-12-11T02:01:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.631770 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.631829 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.631857 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 02:01:31 crc kubenswrapper[4824]: E1211 02:01:31.631992 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 02:01:31 crc kubenswrapper[4824]: E1211 02:01:31.632169 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 02:01:31 crc kubenswrapper[4824]: E1211 02:01:31.632338 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.681597 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.681677 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.681702 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.681730 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.681750 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:31Z","lastTransitionTime":"2025-12-11T02:01:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.726847 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-grnpt"] Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.727721 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-grnpt" Dec 11 02:01:31 crc kubenswrapper[4824]: E1211 02:01:31.727867 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-grnpt" podUID="09a52948-0404-45ff-85b1-51479268bf71" Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.750662 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"beab386b-6c65-4bc3-9ef0-dc3f68b4403f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c26f06d2ed5e6e98b61a51f073130675f2c49acda106b8aedcbb2a6d4f7580c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20af9b10a67c8b21a7abcd23e42bc412798d159c8422e7c8ff58f5b997ca19a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9bbfc6edfe050b3fa9dbeee5f98d416bcb2a1533a38608a881b75396fd0b059\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7c766f8de73ca762ffde0c18b2d1c46bb152b2b37535a611c5a0e454a21b7d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d45819f0611c91b20c83eeee1092031a3441fb69f3015241632ad419880227d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 02:01:11.070250 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 02:01:11.071673 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2752882565/tls.crt::/tmp/serving-cert-2752882565/tls.key\\\\\\\"\\\\nI1211 02:01:17.042610 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 02:01:17.048046 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 02:01:17.048080 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 02:01:17.048172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 02:01:17.048188 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 02:01:17.058248 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 02:01:17.058294 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 02:01:17.058303 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 02:01:17.058315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 02:01:17.058322 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 02:01:17.058328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 02:01:17.058335 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 02:01:17.058808 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 02:01:17.060225 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14a940a72b2404a6cbd52a69e150b03266bcd7f5c1d3f4d07b51c0c69e51786\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:00:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:31Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.772023 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ded146c045085be2b58821749052413a36850ede00ae1024fe2dae9fa5ac84b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:31Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.784973 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.785026 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.785050 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.785082 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.785106 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:31Z","lastTransitionTime":"2025-12-11T02:01:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.793163 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76b10bbd1e891b8c682f4b2e17d023fd787a8ed44ed94fbf7d97314d641ccf36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f32838f33832b9b67e7dd9c53340680e17721cf1c7bdf0170230c5f8feafe64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:31Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.801064 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/09a52948-0404-45ff-85b1-51479268bf71-metrics-certs\") pod \"network-metrics-daemon-grnpt\" (UID: \"09a52948-0404-45ff-85b1-51479268bf71\") " pod="openshift-multus/network-metrics-daemon-grnpt" Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.801244 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wmmzm\" (UniqueName: \"kubernetes.io/projected/09a52948-0404-45ff-85b1-51479268bf71-kube-api-access-wmmzm\") pod \"network-metrics-daemon-grnpt\" (UID: \"09a52948-0404-45ff-85b1-51479268bf71\") " pod="openshift-multus/network-metrics-daemon-grnpt" Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.812952 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kptws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68568dba-83a5-46c1-8d01-337359679d91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f965a6d6db619995132f61b5d51b66f0b2b3fdbef3fbec63e231b9a62ccd0a9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd7nf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kptws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:31Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.832891 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kh8g9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"50c69d41-a65b-463e-a829-ea349bd2fe46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dca6be41bd7a3ce12ecd4567c868f696b65e8eb1bad436feda6e4f9bfb8e525\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69af1f33fd0bafc929b94e702b4f46c6c7d7a362d5552b121729c477892d91d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69af1f33fd0bafc929b94e702b4f46c6c7d7a362d5552b121729c477892d91d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd9675d0afded587bc660a3862fd0c5dd6885961afacda957aa9ee8501d68a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd9675d0afded587bc660a3862fd0c5dd6885961afacda957aa9ee8501d68a69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aac94149458bd6321be90b26b48ee26428d273e7f61407d95f9c9d6b3aeae4b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aac94149458bd6321be90b26b48ee26428d273e7f61407d95f9c9d6b3aeae4b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbee11bd7cb65e031415990869b52d2b8f81bf8b56e6cc935d38e09695db7b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbee11bd7cb65e031415990869b52d2b8f81bf8b56e6cc935d38e09695db7b48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86f693ade72bd593bc16021fe4b7564f1914ec2f1cda7039474ac2fb13de10f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86f693ade72bd593bc16021fe4b7564f1914ec2f1cda7039474ac2fb13de10f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7d7674252f67bde0ded2ec8ad1d925a92a26d3e3f741dee0061b735cb0b7a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7d7674252f67bde0ded2ec8ad1d925a92a26d3e3f741dee0061b735cb0b7a02\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kh8g9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:31Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.852813 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6bkc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8022f696-d0e7-437b-848a-3cd25bd1f364\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3e28f6778ec0dba6ea3e6b69e5a2c1c73c0d5db1abeead241fb58efdbfaefe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hh5kn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6bkc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:31Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.882304 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b193bd2-b536-4056-92f1-94c9836ab2eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51f3c41d86ddc44f7eca118029876ad353317f753d0ad173ed4d30c2a15f10e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c345850891744a0240781277e9a89a59f6899c4c944995f62e44650f807440a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed2ae40b1fe28111efc76d9d72345edecb5f3ea34522b238f3730812b29f2bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce7aa078800e6f299cfe03770fa7e3b64fa3742fe0c0f383c87d84cf48569ddf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6aa9fc1f1bff192c69bd59d32a05d4a72d09641d9ce4b265bc4517147a7e15dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b4a3bc50eb80777188206adab774e8acad8b2e4952a8415f91b1c29ba9a0ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c803a25bdafcd4b9b8a4dfbcf8b67d6ade84b490f9d57f46873c32f9c763401d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c803a25bdafcd4b9b8a4dfbcf8b67d6ade84b490f9d57f46873c32f9c763401d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T02:01:29Z\\\",\\\"message\\\":\\\"ode_controller event handlers\\\\nI1211 02:01:29.024367 6208 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1211 02:01:29.024404 6208 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1211 02:01:29.024418 6208 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1211 02:01:29.024439 6208 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1211 02:01:29.024462 6208 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1211 02:01:29.024470 6208 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1211 02:01:29.024492 6208 factory.go:656] Stopping watch factory\\\\nI1211 02:01:29.024509 6208 ovnkube.go:599] Stopped ovnkube\\\\nI1211 02:01:29.024540 6208 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1211 02:01:29.024559 6208 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1211 02:01:29.024571 6208 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1211 02:01:29.024582 6208 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1211 02:01:29.024593 6208 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1211 02:01:29.024604 6208 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1211 02:01:29.024620 6208 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1211 02:01:29.024705 6208 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-f6pgc_openshift-ovn-kubernetes(5b193bd2-b536-4056-92f1-94c9836ab2eb)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ead58694813806d307bbbc6a7b3cf41730d1c059b231a3002684bf09c10c7f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f6pgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:31Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.888315 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.888373 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.888392 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.888416 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.888435 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:31Z","lastTransitionTime":"2025-12-11T02:01:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.902376 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/09a52948-0404-45ff-85b1-51479268bf71-metrics-certs\") pod \"network-metrics-daemon-grnpt\" (UID: \"09a52948-0404-45ff-85b1-51479268bf71\") " pod="openshift-multus/network-metrics-daemon-grnpt" Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.902398 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-br4j9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce125194-9c9d-4ea4-8a72-2c30d5722295\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9798f904052e08196a0db1745060dab11e2710308bca01ff160475910f2665df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fc5s8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-br4j9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:31Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.902460 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wmmzm\" (UniqueName: \"kubernetes.io/projected/09a52948-0404-45ff-85b1-51479268bf71-kube-api-access-wmmzm\") pod \"network-metrics-daemon-grnpt\" (UID: \"09a52948-0404-45ff-85b1-51479268bf71\") " pod="openshift-multus/network-metrics-daemon-grnpt" Dec 11 02:01:31 crc kubenswrapper[4824]: E1211 02:01:31.902624 4824 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 11 02:01:31 crc kubenswrapper[4824]: E1211 02:01:31.902739 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/09a52948-0404-45ff-85b1-51479268bf71-metrics-certs podName:09a52948-0404-45ff-85b1-51479268bf71 nodeName:}" failed. No retries permitted until 2025-12-11 02:01:32.402706172 +0000 UTC m=+34.091743581 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/09a52948-0404-45ff-85b1-51479268bf71-metrics-certs") pod "network-metrics-daemon-grnpt" (UID: "09a52948-0404-45ff-85b1-51479268bf71") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.920060 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8v4jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99a6bfee-afca-4726-a964-b95a948945a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4hk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4hk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-8v4jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:31Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.934405 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wmmzm\" (UniqueName: \"kubernetes.io/projected/09a52948-0404-45ff-85b1-51479268bf71-kube-api-access-wmmzm\") pod \"network-metrics-daemon-grnpt\" (UID: \"09a52948-0404-45ff-85b1-51479268bf71\") " pod="openshift-multus/network-metrics-daemon-grnpt" Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.937468 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-grnpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09a52948-0404-45ff-85b1-51479268bf71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmmzm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmmzm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-grnpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:31Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.956432 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:31Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.977457 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:31Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.992233 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.992297 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.992320 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.992352 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.992374 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:31Z","lastTransitionTime":"2025-12-11T02:01:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:31 crc kubenswrapper[4824]: I1211 02:01:31.997363 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44b156e4-64a4-4d45-aa5e-9b10a862faed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7368d92144b188e95a7796fabf497b0f8d9c0b240293bdc7b4aec1ff75497bc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w62dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcf289ac27dbda4f24aabbde3fb5119329e7b4fc7d6e018168da9a69661a4c57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w62dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gx6xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:31Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:32 crc kubenswrapper[4824]: I1211 02:01:32.015817 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8v4jn" event={"ID":"99a6bfee-afca-4726-a964-b95a948945a4","Type":"ContainerStarted","Data":"e3acc1940a0534ba0e6096175e41834fdd0fbcbb4c3fb6e9a38e76aedae55d2a"} Dec 11 02:01:32 crc kubenswrapper[4824]: I1211 02:01:32.015891 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8v4jn" event={"ID":"99a6bfee-afca-4726-a964-b95a948945a4","Type":"ContainerStarted","Data":"b23b9d8f18db3eeadb60f83b36b05b1778cd0fa31669dad2680bb335ae004acf"} Dec 11 02:01:32 crc kubenswrapper[4824]: I1211 02:01:32.020553 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d14986666febd651039afb52ddf4db8969aca55e88eda9bbb20a67e35ba722ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:32Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:32 crc kubenswrapper[4824]: I1211 02:01:32.043719 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"253d4291-2b2f-487c-8862-3d305f0bdc55\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44b29bb7a6eda7cb890d36f86375b5f3f197ae5cf390f16e71139b0904873e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://922af7f67b99a0ce08776f3f75c0c377e6ea69af1720b6178a6d1e2f9b4b8ba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://136c68020325db8e6930d0f40b0d0d7e013c14c4dd5e7d1449d01d6d458104d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c89afaf56b4a9597f6543ec2e3b176c7a6dab7dc341941fc3bc8ec85cc4edaf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:00:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:32Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:32 crc kubenswrapper[4824]: I1211 02:01:32.064530 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:32Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:32 crc kubenswrapper[4824]: I1211 02:01:32.084475 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d14986666febd651039afb52ddf4db8969aca55e88eda9bbb20a67e35ba722ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:32Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:32 crc kubenswrapper[4824]: I1211 02:01:32.094605 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:32 crc kubenswrapper[4824]: I1211 02:01:32.094661 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:32 crc kubenswrapper[4824]: I1211 02:01:32.094676 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:32 crc kubenswrapper[4824]: I1211 02:01:32.094699 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:32 crc kubenswrapper[4824]: I1211 02:01:32.094715 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:32Z","lastTransitionTime":"2025-12-11T02:01:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:32 crc kubenswrapper[4824]: I1211 02:01:32.103050 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:32Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:32 crc kubenswrapper[4824]: I1211 02:01:32.119342 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44b156e4-64a4-4d45-aa5e-9b10a862faed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7368d92144b188e95a7796fabf497b0f8d9c0b240293bdc7b4aec1ff75497bc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w62dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcf289ac27dbda4f24aabbde3fb5119329e7b4fc7d6e018168da9a69661a4c57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w62dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gx6xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:32Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:32 crc kubenswrapper[4824]: I1211 02:01:32.135423 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"253d4291-2b2f-487c-8862-3d305f0bdc55\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44b29bb7a6eda7cb890d36f86375b5f3f197ae5cf390f16e71139b0904873e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://922af7f67b99a0ce08776f3f75c0c377e6ea69af1720b6178a6d1e2f9b4b8ba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://136c68020325db8e6930d0f40b0d0d7e013c14c4dd5e7d1449d01d6d458104d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c89afaf56b4a9597f6543ec2e3b176c7a6dab7dc341941fc3bc8ec85cc4edaf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:00:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:32Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:32 crc kubenswrapper[4824]: I1211 02:01:32.154025 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:32Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:32 crc kubenswrapper[4824]: I1211 02:01:32.173875 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"beab386b-6c65-4bc3-9ef0-dc3f68b4403f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c26f06d2ed5e6e98b61a51f073130675f2c49acda106b8aedcbb2a6d4f7580c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20af9b10a67c8b21a7abcd23e42bc412798d159c8422e7c8ff58f5b997ca19a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9bbfc6edfe050b3fa9dbeee5f98d416bcb2a1533a38608a881b75396fd0b059\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7c766f8de73ca762ffde0c18b2d1c46bb152b2b37535a611c5a0e454a21b7d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d45819f0611c91b20c83eeee1092031a3441fb69f3015241632ad419880227d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 02:01:11.070250 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 02:01:11.071673 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2752882565/tls.crt::/tmp/serving-cert-2752882565/tls.key\\\\\\\"\\\\nI1211 02:01:17.042610 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 02:01:17.048046 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 02:01:17.048080 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 02:01:17.048172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 02:01:17.048188 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 02:01:17.058248 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 02:01:17.058294 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 02:01:17.058303 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 02:01:17.058315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 02:01:17.058322 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 02:01:17.058328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 02:01:17.058335 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 02:01:17.058808 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 02:01:17.060225 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14a940a72b2404a6cbd52a69e150b03266bcd7f5c1d3f4d07b51c0c69e51786\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:00:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:32Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:32 crc kubenswrapper[4824]: I1211 02:01:32.193720 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ded146c045085be2b58821749052413a36850ede00ae1024fe2dae9fa5ac84b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:32Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:32 crc kubenswrapper[4824]: I1211 02:01:32.197954 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:32 crc kubenswrapper[4824]: I1211 02:01:32.198002 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:32 crc kubenswrapper[4824]: I1211 02:01:32.198020 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:32 crc kubenswrapper[4824]: I1211 02:01:32.198045 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:32 crc kubenswrapper[4824]: I1211 02:01:32.198063 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:32Z","lastTransitionTime":"2025-12-11T02:01:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:32 crc kubenswrapper[4824]: I1211 02:01:32.207855 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76b10bbd1e891b8c682f4b2e17d023fd787a8ed44ed94fbf7d97314d641ccf36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f32838f33832b9b67e7dd9c53340680e17721cf1c7bdf0170230c5f8feafe64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:32Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:32 crc kubenswrapper[4824]: I1211 02:01:32.221647 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kptws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68568dba-83a5-46c1-8d01-337359679d91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f965a6d6db619995132f61b5d51b66f0b2b3fdbef3fbec63e231b9a62ccd0a9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd7nf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kptws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:32Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:32 crc kubenswrapper[4824]: I1211 02:01:32.235385 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8v4jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99a6bfee-afca-4726-a964-b95a948945a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b23b9d8f18db3eeadb60f83b36b05b1778cd0fa31669dad2680bb335ae004acf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4hk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3acc1940a0534ba0e6096175e41834fdd0fbcbb4c3fb6e9a38e76aedae55d2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4hk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-8v4jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:32Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:32 crc kubenswrapper[4824]: I1211 02:01:32.247631 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-grnpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09a52948-0404-45ff-85b1-51479268bf71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmmzm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmmzm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-grnpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:32Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:32 crc kubenswrapper[4824]: I1211 02:01:32.267096 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:32Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:32 crc kubenswrapper[4824]: I1211 02:01:32.288272 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kh8g9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"50c69d41-a65b-463e-a829-ea349bd2fe46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dca6be41bd7a3ce12ecd4567c868f696b65e8eb1bad436feda6e4f9bfb8e525\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69af1f33fd0bafc929b94e702b4f46c6c7d7a362d5552b121729c477892d91d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69af1f33fd0bafc929b94e702b4f46c6c7d7a362d5552b121729c477892d91d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd9675d0afded587bc660a3862fd0c5dd6885961afacda957aa9ee8501d68a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd9675d0afded587bc660a3862fd0c5dd6885961afacda957aa9ee8501d68a69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aac94149458bd6321be90b26b48ee26428d273e7f61407d95f9c9d6b3aeae4b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aac94149458bd6321be90b26b48ee26428d273e7f61407d95f9c9d6b3aeae4b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbee11bd7cb65e031415990869b52d2b8f81bf8b56e6cc935d38e09695db7b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbee11bd7cb65e031415990869b52d2b8f81bf8b56e6cc935d38e09695db7b48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86f693ade72bd593bc16021fe4b7564f1914ec2f1cda7039474ac2fb13de10f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86f693ade72bd593bc16021fe4b7564f1914ec2f1cda7039474ac2fb13de10f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7d7674252f67bde0ded2ec8ad1d925a92a26d3e3f741dee0061b735cb0b7a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7d7674252f67bde0ded2ec8ad1d925a92a26d3e3f741dee0061b735cb0b7a02\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kh8g9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:32Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:32 crc kubenswrapper[4824]: I1211 02:01:32.300792 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:32 crc kubenswrapper[4824]: I1211 02:01:32.301149 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:32 crc kubenswrapper[4824]: I1211 02:01:32.301208 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:32 crc kubenswrapper[4824]: I1211 02:01:32.301245 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:32 crc kubenswrapper[4824]: I1211 02:01:32.301271 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:32Z","lastTransitionTime":"2025-12-11T02:01:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:32 crc kubenswrapper[4824]: I1211 02:01:32.307897 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6bkc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8022f696-d0e7-437b-848a-3cd25bd1f364\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3e28f6778ec0dba6ea3e6b69e5a2c1c73c0d5db1abeead241fb58efdbfaefe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hh5kn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6bkc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:32Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:32 crc kubenswrapper[4824]: I1211 02:01:32.339391 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b193bd2-b536-4056-92f1-94c9836ab2eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51f3c41d86ddc44f7eca118029876ad353317f753d0ad173ed4d30c2a15f10e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c345850891744a0240781277e9a89a59f6899c4c944995f62e44650f807440a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed2ae40b1fe28111efc76d9d72345edecb5f3ea34522b238f3730812b29f2bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce7aa078800e6f299cfe03770fa7e3b64fa3742fe0c0f383c87d84cf48569ddf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6aa9fc1f1bff192c69bd59d32a05d4a72d09641d9ce4b265bc4517147a7e15dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b4a3bc50eb80777188206adab774e8acad8b2e4952a8415f91b1c29ba9a0ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c803a25bdafcd4b9b8a4dfbcf8b67d6ade84b490f9d57f46873c32f9c763401d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c803a25bdafcd4b9b8a4dfbcf8b67d6ade84b490f9d57f46873c32f9c763401d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T02:01:29Z\\\",\\\"message\\\":\\\"ode_controller event handlers\\\\nI1211 02:01:29.024367 6208 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1211 02:01:29.024404 6208 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1211 02:01:29.024418 6208 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1211 02:01:29.024439 6208 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1211 02:01:29.024462 6208 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1211 02:01:29.024470 6208 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1211 02:01:29.024492 6208 factory.go:656] Stopping watch factory\\\\nI1211 02:01:29.024509 6208 ovnkube.go:599] Stopped ovnkube\\\\nI1211 02:01:29.024540 6208 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1211 02:01:29.024559 6208 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1211 02:01:29.024571 6208 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1211 02:01:29.024582 6208 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1211 02:01:29.024593 6208 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1211 02:01:29.024604 6208 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1211 02:01:29.024620 6208 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1211 02:01:29.024705 6208 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-f6pgc_openshift-ovn-kubernetes(5b193bd2-b536-4056-92f1-94c9836ab2eb)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ead58694813806d307bbbc6a7b3cf41730d1c059b231a3002684bf09c10c7f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f6pgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:32Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:32 crc kubenswrapper[4824]: I1211 02:01:32.355799 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-br4j9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce125194-9c9d-4ea4-8a72-2c30d5722295\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9798f904052e08196a0db1745060dab11e2710308bca01ff160475910f2665df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fc5s8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-br4j9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:32Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:32 crc kubenswrapper[4824]: I1211 02:01:32.404213 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:32 crc kubenswrapper[4824]: I1211 02:01:32.404275 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:32 crc kubenswrapper[4824]: I1211 02:01:32.404293 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:32 crc kubenswrapper[4824]: I1211 02:01:32.404317 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:32 crc kubenswrapper[4824]: I1211 02:01:32.404337 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:32Z","lastTransitionTime":"2025-12-11T02:01:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:32 crc kubenswrapper[4824]: I1211 02:01:32.408202 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/09a52948-0404-45ff-85b1-51479268bf71-metrics-certs\") pod \"network-metrics-daemon-grnpt\" (UID: \"09a52948-0404-45ff-85b1-51479268bf71\") " pod="openshift-multus/network-metrics-daemon-grnpt" Dec 11 02:01:32 crc kubenswrapper[4824]: E1211 02:01:32.408393 4824 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 11 02:01:32 crc kubenswrapper[4824]: E1211 02:01:32.408476 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/09a52948-0404-45ff-85b1-51479268bf71-metrics-certs podName:09a52948-0404-45ff-85b1-51479268bf71 nodeName:}" failed. No retries permitted until 2025-12-11 02:01:33.408452426 +0000 UTC m=+35.097489835 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/09a52948-0404-45ff-85b1-51479268bf71-metrics-certs") pod "network-metrics-daemon-grnpt" (UID: "09a52948-0404-45ff-85b1-51479268bf71") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 11 02:01:32 crc kubenswrapper[4824]: I1211 02:01:32.507216 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:32 crc kubenswrapper[4824]: I1211 02:01:32.507275 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:32 crc kubenswrapper[4824]: I1211 02:01:32.507292 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:32 crc kubenswrapper[4824]: I1211 02:01:32.507319 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:32 crc kubenswrapper[4824]: I1211 02:01:32.507342 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:32Z","lastTransitionTime":"2025-12-11T02:01:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:32 crc kubenswrapper[4824]: I1211 02:01:32.609948 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:32 crc kubenswrapper[4824]: I1211 02:01:32.610003 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:32 crc kubenswrapper[4824]: I1211 02:01:32.610020 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:32 crc kubenswrapper[4824]: I1211 02:01:32.610046 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:32 crc kubenswrapper[4824]: I1211 02:01:32.610076 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:32Z","lastTransitionTime":"2025-12-11T02:01:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:32 crc kubenswrapper[4824]: I1211 02:01:32.712781 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:32 crc kubenswrapper[4824]: I1211 02:01:32.712857 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:32 crc kubenswrapper[4824]: I1211 02:01:32.712880 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:32 crc kubenswrapper[4824]: I1211 02:01:32.712914 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:32 crc kubenswrapper[4824]: I1211 02:01:32.712958 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:32Z","lastTransitionTime":"2025-12-11T02:01:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:32 crc kubenswrapper[4824]: I1211 02:01:32.817046 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:32 crc kubenswrapper[4824]: I1211 02:01:32.817153 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:32 crc kubenswrapper[4824]: I1211 02:01:32.817174 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:32 crc kubenswrapper[4824]: I1211 02:01:32.817200 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:32 crc kubenswrapper[4824]: I1211 02:01:32.817221 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:32Z","lastTransitionTime":"2025-12-11T02:01:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:32 crc kubenswrapper[4824]: I1211 02:01:32.920089 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:32 crc kubenswrapper[4824]: I1211 02:01:32.920165 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:32 crc kubenswrapper[4824]: I1211 02:01:32.920180 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:32 crc kubenswrapper[4824]: I1211 02:01:32.920203 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:32 crc kubenswrapper[4824]: I1211 02:01:32.920217 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:32Z","lastTransitionTime":"2025-12-11T02:01:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:33 crc kubenswrapper[4824]: I1211 02:01:33.022316 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:33 crc kubenswrapper[4824]: I1211 02:01:33.022411 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:33 crc kubenswrapper[4824]: I1211 02:01:33.022435 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:33 crc kubenswrapper[4824]: I1211 02:01:33.022464 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:33 crc kubenswrapper[4824]: I1211 02:01:33.022488 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:33Z","lastTransitionTime":"2025-12-11T02:01:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:33 crc kubenswrapper[4824]: I1211 02:01:33.125763 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:33 crc kubenswrapper[4824]: I1211 02:01:33.125839 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:33 crc kubenswrapper[4824]: I1211 02:01:33.125864 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:33 crc kubenswrapper[4824]: I1211 02:01:33.125892 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:33 crc kubenswrapper[4824]: I1211 02:01:33.125914 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:33Z","lastTransitionTime":"2025-12-11T02:01:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:33 crc kubenswrapper[4824]: I1211 02:01:33.228946 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:33 crc kubenswrapper[4824]: I1211 02:01:33.229023 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:33 crc kubenswrapper[4824]: I1211 02:01:33.229042 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:33 crc kubenswrapper[4824]: I1211 02:01:33.229070 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:33 crc kubenswrapper[4824]: I1211 02:01:33.229088 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:33Z","lastTransitionTime":"2025-12-11T02:01:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:33 crc kubenswrapper[4824]: I1211 02:01:33.332331 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:33 crc kubenswrapper[4824]: I1211 02:01:33.332405 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:33 crc kubenswrapper[4824]: I1211 02:01:33.332429 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:33 crc kubenswrapper[4824]: I1211 02:01:33.332459 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:33 crc kubenswrapper[4824]: I1211 02:01:33.332481 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:33Z","lastTransitionTime":"2025-12-11T02:01:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:33 crc kubenswrapper[4824]: I1211 02:01:33.419707 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 02:01:33 crc kubenswrapper[4824]: I1211 02:01:33.419930 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/09a52948-0404-45ff-85b1-51479268bf71-metrics-certs\") pod \"network-metrics-daemon-grnpt\" (UID: \"09a52948-0404-45ff-85b1-51479268bf71\") " pod="openshift-multus/network-metrics-daemon-grnpt" Dec 11 02:01:33 crc kubenswrapper[4824]: E1211 02:01:33.420090 4824 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 11 02:01:33 crc kubenswrapper[4824]: E1211 02:01:33.420210 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/09a52948-0404-45ff-85b1-51479268bf71-metrics-certs podName:09a52948-0404-45ff-85b1-51479268bf71 nodeName:}" failed. No retries permitted until 2025-12-11 02:01:35.420188299 +0000 UTC m=+37.109225718 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/09a52948-0404-45ff-85b1-51479268bf71-metrics-certs") pod "network-metrics-daemon-grnpt" (UID: "09a52948-0404-45ff-85b1-51479268bf71") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 11 02:01:33 crc kubenswrapper[4824]: E1211 02:01:33.420442 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 02:01:49.420404835 +0000 UTC m=+51.109442244 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:01:33 crc kubenswrapper[4824]: I1211 02:01:33.435650 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:33 crc kubenswrapper[4824]: I1211 02:01:33.435700 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:33 crc kubenswrapper[4824]: I1211 02:01:33.435713 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:33 crc kubenswrapper[4824]: I1211 02:01:33.435733 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:33 crc kubenswrapper[4824]: I1211 02:01:33.435746 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:33Z","lastTransitionTime":"2025-12-11T02:01:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:33 crc kubenswrapper[4824]: I1211 02:01:33.521420 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 02:01:33 crc kubenswrapper[4824]: I1211 02:01:33.521488 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 02:01:33 crc kubenswrapper[4824]: I1211 02:01:33.521527 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 02:01:33 crc kubenswrapper[4824]: E1211 02:01:33.521549 4824 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 11 02:01:33 crc kubenswrapper[4824]: E1211 02:01:33.521605 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-11 02:01:49.521590856 +0000 UTC m=+51.210628235 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 11 02:01:33 crc kubenswrapper[4824]: I1211 02:01:33.521622 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 02:01:33 crc kubenswrapper[4824]: E1211 02:01:33.521669 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 11 02:01:33 crc kubenswrapper[4824]: E1211 02:01:33.521703 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 11 02:01:33 crc kubenswrapper[4824]: E1211 02:01:33.521716 4824 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 02:01:33 crc kubenswrapper[4824]: E1211 02:01:33.521727 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 11 02:01:33 crc kubenswrapper[4824]: E1211 02:01:33.521760 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 11 02:01:33 crc kubenswrapper[4824]: E1211 02:01:33.521768 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-11 02:01:49.52175021 +0000 UTC m=+51.210787589 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 02:01:33 crc kubenswrapper[4824]: E1211 02:01:33.521686 4824 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 11 02:01:33 crc kubenswrapper[4824]: E1211 02:01:33.521784 4824 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 02:01:33 crc kubenswrapper[4824]: E1211 02:01:33.521802 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-11 02:01:49.521794892 +0000 UTC m=+51.210832271 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 11 02:01:33 crc kubenswrapper[4824]: E1211 02:01:33.521854 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-11 02:01:49.521831703 +0000 UTC m=+51.210869152 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 02:01:33 crc kubenswrapper[4824]: I1211 02:01:33.538307 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:33 crc kubenswrapper[4824]: I1211 02:01:33.538352 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:33 crc kubenswrapper[4824]: I1211 02:01:33.538367 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:33 crc kubenswrapper[4824]: I1211 02:01:33.538386 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:33 crc kubenswrapper[4824]: I1211 02:01:33.538401 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:33Z","lastTransitionTime":"2025-12-11T02:01:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:33 crc kubenswrapper[4824]: I1211 02:01:33.631521 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 02:01:33 crc kubenswrapper[4824]: I1211 02:01:33.631544 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 02:01:33 crc kubenswrapper[4824]: I1211 02:01:33.631584 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-grnpt" Dec 11 02:01:33 crc kubenswrapper[4824]: I1211 02:01:33.631607 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 02:01:33 crc kubenswrapper[4824]: E1211 02:01:33.631692 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 02:01:33 crc kubenswrapper[4824]: E1211 02:01:33.631832 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-grnpt" podUID="09a52948-0404-45ff-85b1-51479268bf71" Dec 11 02:01:33 crc kubenswrapper[4824]: E1211 02:01:33.631959 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 02:01:33 crc kubenswrapper[4824]: E1211 02:01:33.632046 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 02:01:33 crc kubenswrapper[4824]: I1211 02:01:33.641238 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:33 crc kubenswrapper[4824]: I1211 02:01:33.641288 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:33 crc kubenswrapper[4824]: I1211 02:01:33.641306 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:33 crc kubenswrapper[4824]: I1211 02:01:33.641339 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:33 crc kubenswrapper[4824]: I1211 02:01:33.641358 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:33Z","lastTransitionTime":"2025-12-11T02:01:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:33 crc kubenswrapper[4824]: I1211 02:01:33.743844 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:33 crc kubenswrapper[4824]: I1211 02:01:33.743903 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:33 crc kubenswrapper[4824]: I1211 02:01:33.743919 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:33 crc kubenswrapper[4824]: I1211 02:01:33.743943 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:33 crc kubenswrapper[4824]: I1211 02:01:33.743960 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:33Z","lastTransitionTime":"2025-12-11T02:01:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:33 crc kubenswrapper[4824]: I1211 02:01:33.847280 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:33 crc kubenswrapper[4824]: I1211 02:01:33.847350 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:33 crc kubenswrapper[4824]: I1211 02:01:33.847367 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:33 crc kubenswrapper[4824]: I1211 02:01:33.847392 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:33 crc kubenswrapper[4824]: I1211 02:01:33.847409 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:33Z","lastTransitionTime":"2025-12-11T02:01:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:33 crc kubenswrapper[4824]: I1211 02:01:33.950017 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:33 crc kubenswrapper[4824]: I1211 02:01:33.950065 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:33 crc kubenswrapper[4824]: I1211 02:01:33.950098 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:33 crc kubenswrapper[4824]: I1211 02:01:33.950161 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:33 crc kubenswrapper[4824]: I1211 02:01:33.950179 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:33Z","lastTransitionTime":"2025-12-11T02:01:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:34 crc kubenswrapper[4824]: I1211 02:01:34.053051 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:34 crc kubenswrapper[4824]: I1211 02:01:34.053172 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:34 crc kubenswrapper[4824]: I1211 02:01:34.053192 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:34 crc kubenswrapper[4824]: I1211 02:01:34.053219 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:34 crc kubenswrapper[4824]: I1211 02:01:34.053238 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:34Z","lastTransitionTime":"2025-12-11T02:01:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:34 crc kubenswrapper[4824]: I1211 02:01:34.156610 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:34 crc kubenswrapper[4824]: I1211 02:01:34.156681 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:34 crc kubenswrapper[4824]: I1211 02:01:34.156699 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:34 crc kubenswrapper[4824]: I1211 02:01:34.156722 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:34 crc kubenswrapper[4824]: I1211 02:01:34.156741 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:34Z","lastTransitionTime":"2025-12-11T02:01:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:34 crc kubenswrapper[4824]: I1211 02:01:34.259531 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:34 crc kubenswrapper[4824]: I1211 02:01:34.259612 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:34 crc kubenswrapper[4824]: I1211 02:01:34.259630 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:34 crc kubenswrapper[4824]: I1211 02:01:34.259652 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:34 crc kubenswrapper[4824]: I1211 02:01:34.259669 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:34Z","lastTransitionTime":"2025-12-11T02:01:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:34 crc kubenswrapper[4824]: I1211 02:01:34.362788 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:34 crc kubenswrapper[4824]: I1211 02:01:34.362863 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:34 crc kubenswrapper[4824]: I1211 02:01:34.362887 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:34 crc kubenswrapper[4824]: I1211 02:01:34.362917 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:34 crc kubenswrapper[4824]: I1211 02:01:34.362966 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:34Z","lastTransitionTime":"2025-12-11T02:01:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:34 crc kubenswrapper[4824]: I1211 02:01:34.466161 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:34 crc kubenswrapper[4824]: I1211 02:01:34.466223 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:34 crc kubenswrapper[4824]: I1211 02:01:34.466240 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:34 crc kubenswrapper[4824]: I1211 02:01:34.466265 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:34 crc kubenswrapper[4824]: I1211 02:01:34.466283 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:34Z","lastTransitionTime":"2025-12-11T02:01:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:34 crc kubenswrapper[4824]: I1211 02:01:34.568563 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:34 crc kubenswrapper[4824]: I1211 02:01:34.568618 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:34 crc kubenswrapper[4824]: I1211 02:01:34.568635 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:34 crc kubenswrapper[4824]: I1211 02:01:34.568658 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:34 crc kubenswrapper[4824]: I1211 02:01:34.568723 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:34Z","lastTransitionTime":"2025-12-11T02:01:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:34 crc kubenswrapper[4824]: I1211 02:01:34.671443 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:34 crc kubenswrapper[4824]: I1211 02:01:34.671509 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:34 crc kubenswrapper[4824]: I1211 02:01:34.671529 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:34 crc kubenswrapper[4824]: I1211 02:01:34.671558 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:34 crc kubenswrapper[4824]: I1211 02:01:34.671577 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:34Z","lastTransitionTime":"2025-12-11T02:01:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:34 crc kubenswrapper[4824]: I1211 02:01:34.775583 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:34 crc kubenswrapper[4824]: I1211 02:01:34.775636 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:34 crc kubenswrapper[4824]: I1211 02:01:34.775653 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:34 crc kubenswrapper[4824]: I1211 02:01:34.775678 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:34 crc kubenswrapper[4824]: I1211 02:01:34.775695 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:34Z","lastTransitionTime":"2025-12-11T02:01:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:34 crc kubenswrapper[4824]: I1211 02:01:34.878848 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:34 crc kubenswrapper[4824]: I1211 02:01:34.878957 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:34 crc kubenswrapper[4824]: I1211 02:01:34.878979 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:34 crc kubenswrapper[4824]: I1211 02:01:34.879011 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:34 crc kubenswrapper[4824]: I1211 02:01:34.879035 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:34Z","lastTransitionTime":"2025-12-11T02:01:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:34 crc kubenswrapper[4824]: I1211 02:01:34.985369 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:34 crc kubenswrapper[4824]: I1211 02:01:34.985466 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:34 crc kubenswrapper[4824]: I1211 02:01:34.985487 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:34 crc kubenswrapper[4824]: I1211 02:01:34.985514 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:34 crc kubenswrapper[4824]: I1211 02:01:34.985581 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:34Z","lastTransitionTime":"2025-12-11T02:01:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:35 crc kubenswrapper[4824]: I1211 02:01:35.088476 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:35 crc kubenswrapper[4824]: I1211 02:01:35.088562 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:35 crc kubenswrapper[4824]: I1211 02:01:35.088580 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:35 crc kubenswrapper[4824]: I1211 02:01:35.088605 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:35 crc kubenswrapper[4824]: I1211 02:01:35.088626 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:35Z","lastTransitionTime":"2025-12-11T02:01:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:35 crc kubenswrapper[4824]: I1211 02:01:35.191854 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:35 crc kubenswrapper[4824]: I1211 02:01:35.191912 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:35 crc kubenswrapper[4824]: I1211 02:01:35.191928 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:35 crc kubenswrapper[4824]: I1211 02:01:35.191956 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:35 crc kubenswrapper[4824]: I1211 02:01:35.191974 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:35Z","lastTransitionTime":"2025-12-11T02:01:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:35 crc kubenswrapper[4824]: I1211 02:01:35.294760 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:35 crc kubenswrapper[4824]: I1211 02:01:35.294817 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:35 crc kubenswrapper[4824]: I1211 02:01:35.294829 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:35 crc kubenswrapper[4824]: I1211 02:01:35.294848 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:35 crc kubenswrapper[4824]: I1211 02:01:35.294860 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:35Z","lastTransitionTime":"2025-12-11T02:01:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:35 crc kubenswrapper[4824]: I1211 02:01:35.398252 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:35 crc kubenswrapper[4824]: I1211 02:01:35.398292 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:35 crc kubenswrapper[4824]: I1211 02:01:35.398306 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:35 crc kubenswrapper[4824]: I1211 02:01:35.398326 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:35 crc kubenswrapper[4824]: I1211 02:01:35.398339 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:35Z","lastTransitionTime":"2025-12-11T02:01:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:35 crc kubenswrapper[4824]: I1211 02:01:35.442815 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/09a52948-0404-45ff-85b1-51479268bf71-metrics-certs\") pod \"network-metrics-daemon-grnpt\" (UID: \"09a52948-0404-45ff-85b1-51479268bf71\") " pod="openshift-multus/network-metrics-daemon-grnpt" Dec 11 02:01:35 crc kubenswrapper[4824]: E1211 02:01:35.443103 4824 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 11 02:01:35 crc kubenswrapper[4824]: E1211 02:01:35.443276 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/09a52948-0404-45ff-85b1-51479268bf71-metrics-certs podName:09a52948-0404-45ff-85b1-51479268bf71 nodeName:}" failed. No retries permitted until 2025-12-11 02:01:39.443238314 +0000 UTC m=+41.132275733 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/09a52948-0404-45ff-85b1-51479268bf71-metrics-certs") pod "network-metrics-daemon-grnpt" (UID: "09a52948-0404-45ff-85b1-51479268bf71") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 11 02:01:35 crc kubenswrapper[4824]: I1211 02:01:35.501851 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:35 crc kubenswrapper[4824]: I1211 02:01:35.501902 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:35 crc kubenswrapper[4824]: I1211 02:01:35.501920 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:35 crc kubenswrapper[4824]: I1211 02:01:35.501944 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:35 crc kubenswrapper[4824]: I1211 02:01:35.501961 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:35Z","lastTransitionTime":"2025-12-11T02:01:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:35 crc kubenswrapper[4824]: I1211 02:01:35.604712 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:35 crc kubenswrapper[4824]: I1211 02:01:35.604772 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:35 crc kubenswrapper[4824]: I1211 02:01:35.604790 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:35 crc kubenswrapper[4824]: I1211 02:01:35.604814 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:35 crc kubenswrapper[4824]: I1211 02:01:35.604831 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:35Z","lastTransitionTime":"2025-12-11T02:01:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:35 crc kubenswrapper[4824]: I1211 02:01:35.632098 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 02:01:35 crc kubenswrapper[4824]: I1211 02:01:35.632159 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 02:01:35 crc kubenswrapper[4824]: I1211 02:01:35.632165 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-grnpt" Dec 11 02:01:35 crc kubenswrapper[4824]: I1211 02:01:35.632321 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 02:01:35 crc kubenswrapper[4824]: E1211 02:01:35.632526 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 02:01:35 crc kubenswrapper[4824]: E1211 02:01:35.632667 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-grnpt" podUID="09a52948-0404-45ff-85b1-51479268bf71" Dec 11 02:01:35 crc kubenswrapper[4824]: E1211 02:01:35.632774 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 02:01:35 crc kubenswrapper[4824]: E1211 02:01:35.632855 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 02:01:35 crc kubenswrapper[4824]: I1211 02:01:35.707565 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:35 crc kubenswrapper[4824]: I1211 02:01:35.707640 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:35 crc kubenswrapper[4824]: I1211 02:01:35.707664 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:35 crc kubenswrapper[4824]: I1211 02:01:35.707695 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:35 crc kubenswrapper[4824]: I1211 02:01:35.707716 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:35Z","lastTransitionTime":"2025-12-11T02:01:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:35 crc kubenswrapper[4824]: I1211 02:01:35.811194 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:35 crc kubenswrapper[4824]: I1211 02:01:35.811253 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:35 crc kubenswrapper[4824]: I1211 02:01:35.811270 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:35 crc kubenswrapper[4824]: I1211 02:01:35.811294 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:35 crc kubenswrapper[4824]: I1211 02:01:35.811315 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:35Z","lastTransitionTime":"2025-12-11T02:01:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:35 crc kubenswrapper[4824]: I1211 02:01:35.914067 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:35 crc kubenswrapper[4824]: I1211 02:01:35.914106 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:35 crc kubenswrapper[4824]: I1211 02:01:35.914134 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:35 crc kubenswrapper[4824]: I1211 02:01:35.914148 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:35 crc kubenswrapper[4824]: I1211 02:01:35.914157 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:35Z","lastTransitionTime":"2025-12-11T02:01:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:36 crc kubenswrapper[4824]: I1211 02:01:36.016713 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:36 crc kubenswrapper[4824]: I1211 02:01:36.016771 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:36 crc kubenswrapper[4824]: I1211 02:01:36.016789 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:36 crc kubenswrapper[4824]: I1211 02:01:36.016817 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:36 crc kubenswrapper[4824]: I1211 02:01:36.016833 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:36Z","lastTransitionTime":"2025-12-11T02:01:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:36 crc kubenswrapper[4824]: I1211 02:01:36.120565 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:36 crc kubenswrapper[4824]: I1211 02:01:36.120620 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:36 crc kubenswrapper[4824]: I1211 02:01:36.120637 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:36 crc kubenswrapper[4824]: I1211 02:01:36.120660 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:36 crc kubenswrapper[4824]: I1211 02:01:36.120676 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:36Z","lastTransitionTime":"2025-12-11T02:01:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:36 crc kubenswrapper[4824]: I1211 02:01:36.224013 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:36 crc kubenswrapper[4824]: I1211 02:01:36.224382 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:36 crc kubenswrapper[4824]: I1211 02:01:36.224524 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:36 crc kubenswrapper[4824]: I1211 02:01:36.224659 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:36 crc kubenswrapper[4824]: I1211 02:01:36.224785 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:36Z","lastTransitionTime":"2025-12-11T02:01:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:36 crc kubenswrapper[4824]: I1211 02:01:36.328549 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:36 crc kubenswrapper[4824]: I1211 02:01:36.328672 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:36 crc kubenswrapper[4824]: I1211 02:01:36.328694 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:36 crc kubenswrapper[4824]: I1211 02:01:36.328717 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:36 crc kubenswrapper[4824]: I1211 02:01:36.328735 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:36Z","lastTransitionTime":"2025-12-11T02:01:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:36 crc kubenswrapper[4824]: I1211 02:01:36.431899 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:36 crc kubenswrapper[4824]: I1211 02:01:36.431963 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:36 crc kubenswrapper[4824]: I1211 02:01:36.431976 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:36 crc kubenswrapper[4824]: I1211 02:01:36.431995 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:36 crc kubenswrapper[4824]: I1211 02:01:36.432006 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:36Z","lastTransitionTime":"2025-12-11T02:01:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:36 crc kubenswrapper[4824]: I1211 02:01:36.534468 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:36 crc kubenswrapper[4824]: I1211 02:01:36.534552 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:36 crc kubenswrapper[4824]: I1211 02:01:36.534587 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:36 crc kubenswrapper[4824]: I1211 02:01:36.534619 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:36 crc kubenswrapper[4824]: I1211 02:01:36.534642 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:36Z","lastTransitionTime":"2025-12-11T02:01:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:36 crc kubenswrapper[4824]: I1211 02:01:36.637834 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:36 crc kubenswrapper[4824]: I1211 02:01:36.637886 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:36 crc kubenswrapper[4824]: I1211 02:01:36.637902 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:36 crc kubenswrapper[4824]: I1211 02:01:36.637923 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:36 crc kubenswrapper[4824]: I1211 02:01:36.637940 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:36Z","lastTransitionTime":"2025-12-11T02:01:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:36 crc kubenswrapper[4824]: I1211 02:01:36.741167 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:36 crc kubenswrapper[4824]: I1211 02:01:36.741255 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:36 crc kubenswrapper[4824]: I1211 02:01:36.741274 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:36 crc kubenswrapper[4824]: I1211 02:01:36.741298 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:36 crc kubenswrapper[4824]: I1211 02:01:36.741316 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:36Z","lastTransitionTime":"2025-12-11T02:01:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:36 crc kubenswrapper[4824]: I1211 02:01:36.843605 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:36 crc kubenswrapper[4824]: I1211 02:01:36.843669 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:36 crc kubenswrapper[4824]: I1211 02:01:36.843697 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:36 crc kubenswrapper[4824]: I1211 02:01:36.843727 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:36 crc kubenswrapper[4824]: I1211 02:01:36.843748 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:36Z","lastTransitionTime":"2025-12-11T02:01:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:36 crc kubenswrapper[4824]: I1211 02:01:36.947081 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:36 crc kubenswrapper[4824]: I1211 02:01:36.947178 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:36 crc kubenswrapper[4824]: I1211 02:01:36.947196 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:36 crc kubenswrapper[4824]: I1211 02:01:36.947220 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:36 crc kubenswrapper[4824]: I1211 02:01:36.947241 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:36Z","lastTransitionTime":"2025-12-11T02:01:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:37 crc kubenswrapper[4824]: I1211 02:01:37.049588 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:37 crc kubenswrapper[4824]: I1211 02:01:37.049641 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:37 crc kubenswrapper[4824]: I1211 02:01:37.049669 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:37 crc kubenswrapper[4824]: I1211 02:01:37.049703 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:37 crc kubenswrapper[4824]: I1211 02:01:37.049724 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:37Z","lastTransitionTime":"2025-12-11T02:01:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:37 crc kubenswrapper[4824]: I1211 02:01:37.152753 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:37 crc kubenswrapper[4824]: I1211 02:01:37.152822 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:37 crc kubenswrapper[4824]: I1211 02:01:37.152840 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:37 crc kubenswrapper[4824]: I1211 02:01:37.152865 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:37 crc kubenswrapper[4824]: I1211 02:01:37.152884 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:37Z","lastTransitionTime":"2025-12-11T02:01:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:37 crc kubenswrapper[4824]: I1211 02:01:37.255620 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:37 crc kubenswrapper[4824]: I1211 02:01:37.255669 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:37 crc kubenswrapper[4824]: I1211 02:01:37.255684 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:37 crc kubenswrapper[4824]: I1211 02:01:37.255705 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:37 crc kubenswrapper[4824]: I1211 02:01:37.255722 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:37Z","lastTransitionTime":"2025-12-11T02:01:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:37 crc kubenswrapper[4824]: I1211 02:01:37.359355 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:37 crc kubenswrapper[4824]: I1211 02:01:37.359447 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:37 crc kubenswrapper[4824]: I1211 02:01:37.359466 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:37 crc kubenswrapper[4824]: I1211 02:01:37.359493 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:37 crc kubenswrapper[4824]: I1211 02:01:37.359513 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:37Z","lastTransitionTime":"2025-12-11T02:01:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:37 crc kubenswrapper[4824]: I1211 02:01:37.462923 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:37 crc kubenswrapper[4824]: I1211 02:01:37.462986 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:37 crc kubenswrapper[4824]: I1211 02:01:37.463005 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:37 crc kubenswrapper[4824]: I1211 02:01:37.463029 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:37 crc kubenswrapper[4824]: I1211 02:01:37.463046 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:37Z","lastTransitionTime":"2025-12-11T02:01:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:37 crc kubenswrapper[4824]: I1211 02:01:37.566495 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:37 crc kubenswrapper[4824]: I1211 02:01:37.566551 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:37 crc kubenswrapper[4824]: I1211 02:01:37.566567 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:37 crc kubenswrapper[4824]: I1211 02:01:37.566594 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:37 crc kubenswrapper[4824]: I1211 02:01:37.566612 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:37Z","lastTransitionTime":"2025-12-11T02:01:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:37 crc kubenswrapper[4824]: I1211 02:01:37.632085 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-grnpt" Dec 11 02:01:37 crc kubenswrapper[4824]: I1211 02:01:37.632186 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 02:01:37 crc kubenswrapper[4824]: I1211 02:01:37.632255 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 02:01:37 crc kubenswrapper[4824]: E1211 02:01:37.632412 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-grnpt" podUID="09a52948-0404-45ff-85b1-51479268bf71" Dec 11 02:01:37 crc kubenswrapper[4824]: I1211 02:01:37.632699 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 02:01:37 crc kubenswrapper[4824]: E1211 02:01:37.632842 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 02:01:37 crc kubenswrapper[4824]: E1211 02:01:37.632937 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 02:01:37 crc kubenswrapper[4824]: E1211 02:01:37.633330 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 02:01:37 crc kubenswrapper[4824]: I1211 02:01:37.668962 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:37 crc kubenswrapper[4824]: I1211 02:01:37.669374 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:37 crc kubenswrapper[4824]: I1211 02:01:37.669445 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:37 crc kubenswrapper[4824]: I1211 02:01:37.669475 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:37 crc kubenswrapper[4824]: I1211 02:01:37.669499 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:37Z","lastTransitionTime":"2025-12-11T02:01:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:37 crc kubenswrapper[4824]: I1211 02:01:37.772553 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:37 crc kubenswrapper[4824]: I1211 02:01:37.772633 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:37 crc kubenswrapper[4824]: I1211 02:01:37.772651 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:37 crc kubenswrapper[4824]: I1211 02:01:37.772677 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:37 crc kubenswrapper[4824]: I1211 02:01:37.772700 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:37Z","lastTransitionTime":"2025-12-11T02:01:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:37 crc kubenswrapper[4824]: I1211 02:01:37.876223 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:37 crc kubenswrapper[4824]: I1211 02:01:37.876281 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:37 crc kubenswrapper[4824]: I1211 02:01:37.876297 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:37 crc kubenswrapper[4824]: I1211 02:01:37.876322 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:37 crc kubenswrapper[4824]: I1211 02:01:37.876340 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:37Z","lastTransitionTime":"2025-12-11T02:01:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:37 crc kubenswrapper[4824]: I1211 02:01:37.941669 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 02:01:37 crc kubenswrapper[4824]: I1211 02:01:37.958281 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kptws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68568dba-83a5-46c1-8d01-337359679d91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f965a6d6db619995132f61b5d51b66f0b2b3fdbef3fbec63e231b9a62ccd0a9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd7nf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kptws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:37Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:37 crc kubenswrapper[4824]: I1211 02:01:37.979916 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:37 crc kubenswrapper[4824]: I1211 02:01:37.979992 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:37 crc kubenswrapper[4824]: I1211 02:01:37.980015 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:37 crc kubenswrapper[4824]: I1211 02:01:37.980049 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:37 crc kubenswrapper[4824]: I1211 02:01:37.980075 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:37Z","lastTransitionTime":"2025-12-11T02:01:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:37 crc kubenswrapper[4824]: I1211 02:01:37.983226 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"beab386b-6c65-4bc3-9ef0-dc3f68b4403f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c26f06d2ed5e6e98b61a51f073130675f2c49acda106b8aedcbb2a6d4f7580c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20af9b10a67c8b21a7abcd23e42bc412798d159c8422e7c8ff58f5b997ca19a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9bbfc6edfe050b3fa9dbeee5f98d416bcb2a1533a38608a881b75396fd0b059\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7c766f8de73ca762ffde0c18b2d1c46bb152b2b37535a611c5a0e454a21b7d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d45819f0611c91b20c83eeee1092031a3441fb69f3015241632ad419880227d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 02:01:11.070250 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 02:01:11.071673 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2752882565/tls.crt::/tmp/serving-cert-2752882565/tls.key\\\\\\\"\\\\nI1211 02:01:17.042610 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 02:01:17.048046 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 02:01:17.048080 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 02:01:17.048172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 02:01:17.048188 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 02:01:17.058248 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 02:01:17.058294 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 02:01:17.058303 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 02:01:17.058315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 02:01:17.058322 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 02:01:17.058328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 02:01:17.058335 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 02:01:17.058808 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 02:01:17.060225 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14a940a72b2404a6cbd52a69e150b03266bcd7f5c1d3f4d07b51c0c69e51786\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:00:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:37Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:38 crc kubenswrapper[4824]: I1211 02:01:38.004462 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ded146c045085be2b58821749052413a36850ede00ae1024fe2dae9fa5ac84b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:38Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:38 crc kubenswrapper[4824]: I1211 02:01:38.024316 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76b10bbd1e891b8c682f4b2e17d023fd787a8ed44ed94fbf7d97314d641ccf36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f32838f33832b9b67e7dd9c53340680e17721cf1c7bdf0170230c5f8feafe64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:38Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:38 crc kubenswrapper[4824]: I1211 02:01:38.041586 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-br4j9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce125194-9c9d-4ea4-8a72-2c30d5722295\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9798f904052e08196a0db1745060dab11e2710308bca01ff160475910f2665df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fc5s8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-br4j9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:38Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:38 crc kubenswrapper[4824]: I1211 02:01:38.059317 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8v4jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99a6bfee-afca-4726-a964-b95a948945a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b23b9d8f18db3eeadb60f83b36b05b1778cd0fa31669dad2680bb335ae004acf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4hk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3acc1940a0534ba0e6096175e41834fdd0fbcbb4c3fb6e9a38e76aedae55d2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4hk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-8v4jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:38Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:38 crc kubenswrapper[4824]: I1211 02:01:38.076227 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-grnpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09a52948-0404-45ff-85b1-51479268bf71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmmzm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmmzm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-grnpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:38Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:38 crc kubenswrapper[4824]: I1211 02:01:38.082761 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:38 crc kubenswrapper[4824]: I1211 02:01:38.082832 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:38 crc kubenswrapper[4824]: I1211 02:01:38.082861 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:38 crc kubenswrapper[4824]: I1211 02:01:38.082892 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:38 crc kubenswrapper[4824]: I1211 02:01:38.082913 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:38Z","lastTransitionTime":"2025-12-11T02:01:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:38 crc kubenswrapper[4824]: I1211 02:01:38.096760 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:38Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:38 crc kubenswrapper[4824]: I1211 02:01:38.120860 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kh8g9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"50c69d41-a65b-463e-a829-ea349bd2fe46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dca6be41bd7a3ce12ecd4567c868f696b65e8eb1bad436feda6e4f9bfb8e525\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69af1f33fd0bafc929b94e702b4f46c6c7d7a362d5552b121729c477892d91d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69af1f33fd0bafc929b94e702b4f46c6c7d7a362d5552b121729c477892d91d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd9675d0afded587bc660a3862fd0c5dd6885961afacda957aa9ee8501d68a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd9675d0afded587bc660a3862fd0c5dd6885961afacda957aa9ee8501d68a69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aac94149458bd6321be90b26b48ee26428d273e7f61407d95f9c9d6b3aeae4b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aac94149458bd6321be90b26b48ee26428d273e7f61407d95f9c9d6b3aeae4b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbee11bd7cb65e031415990869b52d2b8f81bf8b56e6cc935d38e09695db7b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbee11bd7cb65e031415990869b52d2b8f81bf8b56e6cc935d38e09695db7b48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86f693ade72bd593bc16021fe4b7564f1914ec2f1cda7039474ac2fb13de10f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86f693ade72bd593bc16021fe4b7564f1914ec2f1cda7039474ac2fb13de10f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7d7674252f67bde0ded2ec8ad1d925a92a26d3e3f741dee0061b735cb0b7a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7d7674252f67bde0ded2ec8ad1d925a92a26d3e3f741dee0061b735cb0b7a02\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kh8g9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:38Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:38 crc kubenswrapper[4824]: I1211 02:01:38.143923 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6bkc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8022f696-d0e7-437b-848a-3cd25bd1f364\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3e28f6778ec0dba6ea3e6b69e5a2c1c73c0d5db1abeead241fb58efdbfaefe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hh5kn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6bkc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:38Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:38 crc kubenswrapper[4824]: I1211 02:01:38.186233 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:38 crc kubenswrapper[4824]: I1211 02:01:38.186296 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:38 crc kubenswrapper[4824]: I1211 02:01:38.186316 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:38 crc kubenswrapper[4824]: I1211 02:01:38.186344 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:38 crc kubenswrapper[4824]: I1211 02:01:38.186367 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:38Z","lastTransitionTime":"2025-12-11T02:01:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:38 crc kubenswrapper[4824]: I1211 02:01:38.187192 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b193bd2-b536-4056-92f1-94c9836ab2eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51f3c41d86ddc44f7eca118029876ad353317f753d0ad173ed4d30c2a15f10e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c345850891744a0240781277e9a89a59f6899c4c944995f62e44650f807440a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed2ae40b1fe28111efc76d9d72345edecb5f3ea34522b238f3730812b29f2bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce7aa078800e6f299cfe03770fa7e3b64fa3742fe0c0f383c87d84cf48569ddf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6aa9fc1f1bff192c69bd59d32a05d4a72d09641d9ce4b265bc4517147a7e15dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b4a3bc50eb80777188206adab774e8acad8b2e4952a8415f91b1c29ba9a0ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c803a25bdafcd4b9b8a4dfbcf8b67d6ade84b490f9d57f46873c32f9c763401d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c803a25bdafcd4b9b8a4dfbcf8b67d6ade84b490f9d57f46873c32f9c763401d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T02:01:29Z\\\",\\\"message\\\":\\\"ode_controller event handlers\\\\nI1211 02:01:29.024367 6208 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1211 02:01:29.024404 6208 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1211 02:01:29.024418 6208 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1211 02:01:29.024439 6208 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1211 02:01:29.024462 6208 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1211 02:01:29.024470 6208 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1211 02:01:29.024492 6208 factory.go:656] Stopping watch factory\\\\nI1211 02:01:29.024509 6208 ovnkube.go:599] Stopped ovnkube\\\\nI1211 02:01:29.024540 6208 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1211 02:01:29.024559 6208 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1211 02:01:29.024571 6208 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1211 02:01:29.024582 6208 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1211 02:01:29.024593 6208 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1211 02:01:29.024604 6208 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1211 02:01:29.024620 6208 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1211 02:01:29.024705 6208 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-f6pgc_openshift-ovn-kubernetes(5b193bd2-b536-4056-92f1-94c9836ab2eb)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ead58694813806d307bbbc6a7b3cf41730d1c059b231a3002684bf09c10c7f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f6pgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:38Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:38 crc kubenswrapper[4824]: I1211 02:01:38.209446 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d14986666febd651039afb52ddf4db8969aca55e88eda9bbb20a67e35ba722ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:38Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:38 crc kubenswrapper[4824]: I1211 02:01:38.229944 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:38Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:38 crc kubenswrapper[4824]: I1211 02:01:38.247412 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44b156e4-64a4-4d45-aa5e-9b10a862faed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7368d92144b188e95a7796fabf497b0f8d9c0b240293bdc7b4aec1ff75497bc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w62dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcf289ac27dbda4f24aabbde3fb5119329e7b4fc7d6e018168da9a69661a4c57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w62dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gx6xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:38Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:38 crc kubenswrapper[4824]: I1211 02:01:38.269316 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"253d4291-2b2f-487c-8862-3d305f0bdc55\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44b29bb7a6eda7cb890d36f86375b5f3f197ae5cf390f16e71139b0904873e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://922af7f67b99a0ce08776f3f75c0c377e6ea69af1720b6178a6d1e2f9b4b8ba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://136c68020325db8e6930d0f40b0d0d7e013c14c4dd5e7d1449d01d6d458104d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c89afaf56b4a9597f6543ec2e3b176c7a6dab7dc341941fc3bc8ec85cc4edaf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:00:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:38Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:38 crc kubenswrapper[4824]: I1211 02:01:38.290056 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:38 crc kubenswrapper[4824]: I1211 02:01:38.290116 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:38 crc kubenswrapper[4824]: I1211 02:01:38.290164 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:38 crc kubenswrapper[4824]: I1211 02:01:38.290192 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:38 crc kubenswrapper[4824]: I1211 02:01:38.290210 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:38Z","lastTransitionTime":"2025-12-11T02:01:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:38 crc kubenswrapper[4824]: I1211 02:01:38.292057 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:38Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:38 crc kubenswrapper[4824]: I1211 02:01:38.393630 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:38 crc kubenswrapper[4824]: I1211 02:01:38.393697 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:38 crc kubenswrapper[4824]: I1211 02:01:38.393714 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:38 crc kubenswrapper[4824]: I1211 02:01:38.393789 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:38 crc kubenswrapper[4824]: I1211 02:01:38.393823 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:38Z","lastTransitionTime":"2025-12-11T02:01:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:38 crc kubenswrapper[4824]: I1211 02:01:38.498531 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:38 crc kubenswrapper[4824]: I1211 02:01:38.498602 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:38 crc kubenswrapper[4824]: I1211 02:01:38.498619 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:38 crc kubenswrapper[4824]: I1211 02:01:38.498648 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:38 crc kubenswrapper[4824]: I1211 02:01:38.498667 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:38Z","lastTransitionTime":"2025-12-11T02:01:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:38 crc kubenswrapper[4824]: I1211 02:01:38.601627 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:38 crc kubenswrapper[4824]: I1211 02:01:38.601703 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:38 crc kubenswrapper[4824]: I1211 02:01:38.601723 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:38 crc kubenswrapper[4824]: I1211 02:01:38.601750 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:38 crc kubenswrapper[4824]: I1211 02:01:38.601770 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:38Z","lastTransitionTime":"2025-12-11T02:01:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:38 crc kubenswrapper[4824]: I1211 02:01:38.654169 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:38Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:38 crc kubenswrapper[4824]: I1211 02:01:38.678910 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kh8g9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"50c69d41-a65b-463e-a829-ea349bd2fe46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dca6be41bd7a3ce12ecd4567c868f696b65e8eb1bad436feda6e4f9bfb8e525\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69af1f33fd0bafc929b94e702b4f46c6c7d7a362d5552b121729c477892d91d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69af1f33fd0bafc929b94e702b4f46c6c7d7a362d5552b121729c477892d91d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd9675d0afded587bc660a3862fd0c5dd6885961afacda957aa9ee8501d68a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd9675d0afded587bc660a3862fd0c5dd6885961afacda957aa9ee8501d68a69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aac94149458bd6321be90b26b48ee26428d273e7f61407d95f9c9d6b3aeae4b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aac94149458bd6321be90b26b48ee26428d273e7f61407d95f9c9d6b3aeae4b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbee11bd7cb65e031415990869b52d2b8f81bf8b56e6cc935d38e09695db7b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbee11bd7cb65e031415990869b52d2b8f81bf8b56e6cc935d38e09695db7b48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86f693ade72bd593bc16021fe4b7564f1914ec2f1cda7039474ac2fb13de10f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86f693ade72bd593bc16021fe4b7564f1914ec2f1cda7039474ac2fb13de10f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7d7674252f67bde0ded2ec8ad1d925a92a26d3e3f741dee0061b735cb0b7a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7d7674252f67bde0ded2ec8ad1d925a92a26d3e3f741dee0061b735cb0b7a02\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kh8g9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:38Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:38 crc kubenswrapper[4824]: I1211 02:01:38.701716 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6bkc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8022f696-d0e7-437b-848a-3cd25bd1f364\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3e28f6778ec0dba6ea3e6b69e5a2c1c73c0d5db1abeead241fb58efdbfaefe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hh5kn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6bkc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:38Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:38 crc kubenswrapper[4824]: I1211 02:01:38.706590 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:38 crc kubenswrapper[4824]: I1211 02:01:38.706659 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:38 crc kubenswrapper[4824]: I1211 02:01:38.706684 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:38 crc kubenswrapper[4824]: I1211 02:01:38.706716 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:38 crc kubenswrapper[4824]: I1211 02:01:38.706741 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:38Z","lastTransitionTime":"2025-12-11T02:01:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:38 crc kubenswrapper[4824]: I1211 02:01:38.737671 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b193bd2-b536-4056-92f1-94c9836ab2eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51f3c41d86ddc44f7eca118029876ad353317f753d0ad173ed4d30c2a15f10e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c345850891744a0240781277e9a89a59f6899c4c944995f62e44650f807440a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed2ae40b1fe28111efc76d9d72345edecb5f3ea34522b238f3730812b29f2bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce7aa078800e6f299cfe03770fa7e3b64fa3742fe0c0f383c87d84cf48569ddf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6aa9fc1f1bff192c69bd59d32a05d4a72d09641d9ce4b265bc4517147a7e15dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b4a3bc50eb80777188206adab774e8acad8b2e4952a8415f91b1c29ba9a0ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c803a25bdafcd4b9b8a4dfbcf8b67d6ade84b490f9d57f46873c32f9c763401d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c803a25bdafcd4b9b8a4dfbcf8b67d6ade84b490f9d57f46873c32f9c763401d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T02:01:29Z\\\",\\\"message\\\":\\\"ode_controller event handlers\\\\nI1211 02:01:29.024367 6208 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1211 02:01:29.024404 6208 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1211 02:01:29.024418 6208 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1211 02:01:29.024439 6208 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1211 02:01:29.024462 6208 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1211 02:01:29.024470 6208 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1211 02:01:29.024492 6208 factory.go:656] Stopping watch factory\\\\nI1211 02:01:29.024509 6208 ovnkube.go:599] Stopped ovnkube\\\\nI1211 02:01:29.024540 6208 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1211 02:01:29.024559 6208 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1211 02:01:29.024571 6208 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1211 02:01:29.024582 6208 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1211 02:01:29.024593 6208 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1211 02:01:29.024604 6208 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1211 02:01:29.024620 6208 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1211 02:01:29.024705 6208 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-f6pgc_openshift-ovn-kubernetes(5b193bd2-b536-4056-92f1-94c9836ab2eb)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ead58694813806d307bbbc6a7b3cf41730d1c059b231a3002684bf09c10c7f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f6pgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:38Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:38 crc kubenswrapper[4824]: I1211 02:01:38.760115 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-br4j9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce125194-9c9d-4ea4-8a72-2c30d5722295\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9798f904052e08196a0db1745060dab11e2710308bca01ff160475910f2665df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fc5s8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-br4j9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:38Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:38 crc kubenswrapper[4824]: I1211 02:01:38.774960 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8v4jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99a6bfee-afca-4726-a964-b95a948945a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b23b9d8f18db3eeadb60f83b36b05b1778cd0fa31669dad2680bb335ae004acf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4hk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3acc1940a0534ba0e6096175e41834fdd0fbcbb4c3fb6e9a38e76aedae55d2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4hk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-8v4jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:38Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:38 crc kubenswrapper[4824]: I1211 02:01:38.793323 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-grnpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09a52948-0404-45ff-85b1-51479268bf71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmmzm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmmzm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-grnpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:38Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:38 crc kubenswrapper[4824]: I1211 02:01:38.807427 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d14986666febd651039afb52ddf4db8969aca55e88eda9bbb20a67e35ba722ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:38Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:38 crc kubenswrapper[4824]: I1211 02:01:38.809236 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:38 crc kubenswrapper[4824]: I1211 02:01:38.809308 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:38 crc kubenswrapper[4824]: I1211 02:01:38.809332 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:38 crc kubenswrapper[4824]: I1211 02:01:38.809363 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:38 crc kubenswrapper[4824]: I1211 02:01:38.809381 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:38Z","lastTransitionTime":"2025-12-11T02:01:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:38 crc kubenswrapper[4824]: I1211 02:01:38.822099 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:38Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:38 crc kubenswrapper[4824]: I1211 02:01:38.835358 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44b156e4-64a4-4d45-aa5e-9b10a862faed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7368d92144b188e95a7796fabf497b0f8d9c0b240293bdc7b4aec1ff75497bc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w62dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcf289ac27dbda4f24aabbde3fb5119329e7b4fc7d6e018168da9a69661a4c57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w62dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gx6xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:38Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:38 crc kubenswrapper[4824]: I1211 02:01:38.852957 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"253d4291-2b2f-487c-8862-3d305f0bdc55\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44b29bb7a6eda7cb890d36f86375b5f3f197ae5cf390f16e71139b0904873e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://922af7f67b99a0ce08776f3f75c0c377e6ea69af1720b6178a6d1e2f9b4b8ba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://136c68020325db8e6930d0f40b0d0d7e013c14c4dd5e7d1449d01d6d458104d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c89afaf56b4a9597f6543ec2e3b176c7a6dab7dc341941fc3bc8ec85cc4edaf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:00:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:38Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:38 crc kubenswrapper[4824]: I1211 02:01:38.869313 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:38Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:38 crc kubenswrapper[4824]: I1211 02:01:38.888216 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"beab386b-6c65-4bc3-9ef0-dc3f68b4403f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c26f06d2ed5e6e98b61a51f073130675f2c49acda106b8aedcbb2a6d4f7580c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20af9b10a67c8b21a7abcd23e42bc412798d159c8422e7c8ff58f5b997ca19a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9bbfc6edfe050b3fa9dbeee5f98d416bcb2a1533a38608a881b75396fd0b059\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7c766f8de73ca762ffde0c18b2d1c46bb152b2b37535a611c5a0e454a21b7d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d45819f0611c91b20c83eeee1092031a3441fb69f3015241632ad419880227d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 02:01:11.070250 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 02:01:11.071673 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2752882565/tls.crt::/tmp/serving-cert-2752882565/tls.key\\\\\\\"\\\\nI1211 02:01:17.042610 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 02:01:17.048046 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 02:01:17.048080 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 02:01:17.048172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 02:01:17.048188 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 02:01:17.058248 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 02:01:17.058294 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 02:01:17.058303 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 02:01:17.058315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 02:01:17.058322 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 02:01:17.058328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 02:01:17.058335 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 02:01:17.058808 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 02:01:17.060225 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14a940a72b2404a6cbd52a69e150b03266bcd7f5c1d3f4d07b51c0c69e51786\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:00:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:38Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:38 crc kubenswrapper[4824]: I1211 02:01:38.902802 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ded146c045085be2b58821749052413a36850ede00ae1024fe2dae9fa5ac84b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:38Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:38 crc kubenswrapper[4824]: I1211 02:01:38.911737 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:38 crc kubenswrapper[4824]: I1211 02:01:38.911845 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:38 crc kubenswrapper[4824]: I1211 02:01:38.911870 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:38 crc kubenswrapper[4824]: I1211 02:01:38.911906 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:38 crc kubenswrapper[4824]: I1211 02:01:38.911926 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:38Z","lastTransitionTime":"2025-12-11T02:01:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:38 crc kubenswrapper[4824]: I1211 02:01:38.914998 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76b10bbd1e891b8c682f4b2e17d023fd787a8ed44ed94fbf7d97314d641ccf36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f32838f33832b9b67e7dd9c53340680e17721cf1c7bdf0170230c5f8feafe64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:38Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:38 crc kubenswrapper[4824]: I1211 02:01:38.926844 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kptws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68568dba-83a5-46c1-8d01-337359679d91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f965a6d6db619995132f61b5d51b66f0b2b3fdbef3fbec63e231b9a62ccd0a9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd7nf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kptws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:38Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:39 crc kubenswrapper[4824]: I1211 02:01:39.014541 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:39 crc kubenswrapper[4824]: I1211 02:01:39.014595 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:39 crc kubenswrapper[4824]: I1211 02:01:39.014612 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:39 crc kubenswrapper[4824]: I1211 02:01:39.014636 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:39 crc kubenswrapper[4824]: I1211 02:01:39.014651 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:39Z","lastTransitionTime":"2025-12-11T02:01:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:39 crc kubenswrapper[4824]: I1211 02:01:39.117152 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:39 crc kubenswrapper[4824]: I1211 02:01:39.117188 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:39 crc kubenswrapper[4824]: I1211 02:01:39.117196 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:39 crc kubenswrapper[4824]: I1211 02:01:39.117212 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:39 crc kubenswrapper[4824]: I1211 02:01:39.117220 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:39Z","lastTransitionTime":"2025-12-11T02:01:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:39 crc kubenswrapper[4824]: I1211 02:01:39.220290 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:39 crc kubenswrapper[4824]: I1211 02:01:39.220351 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:39 crc kubenswrapper[4824]: I1211 02:01:39.220369 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:39 crc kubenswrapper[4824]: I1211 02:01:39.220397 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:39 crc kubenswrapper[4824]: I1211 02:01:39.220415 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:39Z","lastTransitionTime":"2025-12-11T02:01:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:39 crc kubenswrapper[4824]: I1211 02:01:39.323874 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:39 crc kubenswrapper[4824]: I1211 02:01:39.323934 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:39 crc kubenswrapper[4824]: I1211 02:01:39.323952 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:39 crc kubenswrapper[4824]: I1211 02:01:39.323980 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:39 crc kubenswrapper[4824]: I1211 02:01:39.323998 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:39Z","lastTransitionTime":"2025-12-11T02:01:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:39 crc kubenswrapper[4824]: I1211 02:01:39.426698 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:39 crc kubenswrapper[4824]: I1211 02:01:39.426759 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:39 crc kubenswrapper[4824]: I1211 02:01:39.426776 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:39 crc kubenswrapper[4824]: I1211 02:01:39.426806 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:39 crc kubenswrapper[4824]: I1211 02:01:39.426826 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:39Z","lastTransitionTime":"2025-12-11T02:01:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:39 crc kubenswrapper[4824]: I1211 02:01:39.497900 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/09a52948-0404-45ff-85b1-51479268bf71-metrics-certs\") pod \"network-metrics-daemon-grnpt\" (UID: \"09a52948-0404-45ff-85b1-51479268bf71\") " pod="openshift-multus/network-metrics-daemon-grnpt" Dec 11 02:01:39 crc kubenswrapper[4824]: E1211 02:01:39.499711 4824 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 11 02:01:39 crc kubenswrapper[4824]: E1211 02:01:39.500579 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/09a52948-0404-45ff-85b1-51479268bf71-metrics-certs podName:09a52948-0404-45ff-85b1-51479268bf71 nodeName:}" failed. No retries permitted until 2025-12-11 02:01:47.500524099 +0000 UTC m=+49.189561518 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/09a52948-0404-45ff-85b1-51479268bf71-metrics-certs") pod "network-metrics-daemon-grnpt" (UID: "09a52948-0404-45ff-85b1-51479268bf71") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 11 02:01:39 crc kubenswrapper[4824]: I1211 02:01:39.529449 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:39 crc kubenswrapper[4824]: I1211 02:01:39.529509 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:39 crc kubenswrapper[4824]: I1211 02:01:39.529531 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:39 crc kubenswrapper[4824]: I1211 02:01:39.529560 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:39 crc kubenswrapper[4824]: I1211 02:01:39.529582 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:39Z","lastTransitionTime":"2025-12-11T02:01:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:39 crc kubenswrapper[4824]: I1211 02:01:39.631674 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-grnpt" Dec 11 02:01:39 crc kubenswrapper[4824]: I1211 02:01:39.631688 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 02:01:39 crc kubenswrapper[4824]: I1211 02:01:39.631716 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 02:01:39 crc kubenswrapper[4824]: I1211 02:01:39.631760 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 02:01:39 crc kubenswrapper[4824]: E1211 02:01:39.632655 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-grnpt" podUID="09a52948-0404-45ff-85b1-51479268bf71" Dec 11 02:01:39 crc kubenswrapper[4824]: E1211 02:01:39.632807 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 02:01:39 crc kubenswrapper[4824]: E1211 02:01:39.632922 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 02:01:39 crc kubenswrapper[4824]: I1211 02:01:39.633313 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:39 crc kubenswrapper[4824]: E1211 02:01:39.633055 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 02:01:39 crc kubenswrapper[4824]: I1211 02:01:39.633354 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:39 crc kubenswrapper[4824]: I1211 02:01:39.633461 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:39 crc kubenswrapper[4824]: I1211 02:01:39.633499 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:39 crc kubenswrapper[4824]: I1211 02:01:39.633519 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:39Z","lastTransitionTime":"2025-12-11T02:01:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:39 crc kubenswrapper[4824]: I1211 02:01:39.736979 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:39 crc kubenswrapper[4824]: I1211 02:01:39.737041 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:39 crc kubenswrapper[4824]: I1211 02:01:39.737058 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:39 crc kubenswrapper[4824]: I1211 02:01:39.737083 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:39 crc kubenswrapper[4824]: I1211 02:01:39.737101 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:39Z","lastTransitionTime":"2025-12-11T02:01:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:39 crc kubenswrapper[4824]: I1211 02:01:39.840943 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:39 crc kubenswrapper[4824]: I1211 02:01:39.841011 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:39 crc kubenswrapper[4824]: I1211 02:01:39.841034 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:39 crc kubenswrapper[4824]: I1211 02:01:39.841066 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:39 crc kubenswrapper[4824]: I1211 02:01:39.841089 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:39Z","lastTransitionTime":"2025-12-11T02:01:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:39 crc kubenswrapper[4824]: I1211 02:01:39.944380 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:39 crc kubenswrapper[4824]: I1211 02:01:39.944441 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:39 crc kubenswrapper[4824]: I1211 02:01:39.944458 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:39 crc kubenswrapper[4824]: I1211 02:01:39.944482 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:39 crc kubenswrapper[4824]: I1211 02:01:39.944505 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:39Z","lastTransitionTime":"2025-12-11T02:01:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:40 crc kubenswrapper[4824]: I1211 02:01:40.046684 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:40 crc kubenswrapper[4824]: I1211 02:01:40.046737 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:40 crc kubenswrapper[4824]: I1211 02:01:40.046754 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:40 crc kubenswrapper[4824]: I1211 02:01:40.046778 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:40 crc kubenswrapper[4824]: I1211 02:01:40.046795 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:40Z","lastTransitionTime":"2025-12-11T02:01:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:40 crc kubenswrapper[4824]: I1211 02:01:40.150064 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:40 crc kubenswrapper[4824]: I1211 02:01:40.150173 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:40 crc kubenswrapper[4824]: I1211 02:01:40.150200 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:40 crc kubenswrapper[4824]: I1211 02:01:40.150234 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:40 crc kubenswrapper[4824]: I1211 02:01:40.150254 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:40Z","lastTransitionTime":"2025-12-11T02:01:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:40 crc kubenswrapper[4824]: I1211 02:01:40.253566 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:40 crc kubenswrapper[4824]: I1211 02:01:40.253650 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:40 crc kubenswrapper[4824]: I1211 02:01:40.253667 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:40 crc kubenswrapper[4824]: I1211 02:01:40.253694 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:40 crc kubenswrapper[4824]: I1211 02:01:40.253712 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:40Z","lastTransitionTime":"2025-12-11T02:01:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:40 crc kubenswrapper[4824]: I1211 02:01:40.356765 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:40 crc kubenswrapper[4824]: I1211 02:01:40.356820 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:40 crc kubenswrapper[4824]: I1211 02:01:40.356839 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:40 crc kubenswrapper[4824]: I1211 02:01:40.356861 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:40 crc kubenswrapper[4824]: I1211 02:01:40.356878 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:40Z","lastTransitionTime":"2025-12-11T02:01:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:40 crc kubenswrapper[4824]: I1211 02:01:40.460434 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:40 crc kubenswrapper[4824]: I1211 02:01:40.460484 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:40 crc kubenswrapper[4824]: I1211 02:01:40.460500 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:40 crc kubenswrapper[4824]: I1211 02:01:40.460522 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:40 crc kubenswrapper[4824]: I1211 02:01:40.460538 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:40Z","lastTransitionTime":"2025-12-11T02:01:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:40 crc kubenswrapper[4824]: I1211 02:01:40.563346 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:40 crc kubenswrapper[4824]: I1211 02:01:40.563412 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:40 crc kubenswrapper[4824]: I1211 02:01:40.563429 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:40 crc kubenswrapper[4824]: I1211 02:01:40.563452 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:40 crc kubenswrapper[4824]: I1211 02:01:40.563468 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:40Z","lastTransitionTime":"2025-12-11T02:01:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:40 crc kubenswrapper[4824]: I1211 02:01:40.665406 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:40 crc kubenswrapper[4824]: I1211 02:01:40.665497 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:40 crc kubenswrapper[4824]: I1211 02:01:40.665515 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:40 crc kubenswrapper[4824]: I1211 02:01:40.665539 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:40 crc kubenswrapper[4824]: I1211 02:01:40.665560 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:40Z","lastTransitionTime":"2025-12-11T02:01:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:40 crc kubenswrapper[4824]: I1211 02:01:40.767895 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:40 crc kubenswrapper[4824]: I1211 02:01:40.767965 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:40 crc kubenswrapper[4824]: I1211 02:01:40.767985 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:40 crc kubenswrapper[4824]: I1211 02:01:40.768012 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:40 crc kubenswrapper[4824]: I1211 02:01:40.768030 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:40Z","lastTransitionTime":"2025-12-11T02:01:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:40 crc kubenswrapper[4824]: I1211 02:01:40.774935 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:40 crc kubenswrapper[4824]: I1211 02:01:40.775008 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:40 crc kubenswrapper[4824]: I1211 02:01:40.775027 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:40 crc kubenswrapper[4824]: I1211 02:01:40.775051 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:40 crc kubenswrapper[4824]: I1211 02:01:40.775073 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:40Z","lastTransitionTime":"2025-12-11T02:01:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:40 crc kubenswrapper[4824]: E1211 02:01:40.793665 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:01:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:01:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:01:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:01:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d57bece6-6fc6-4a84-b8a8-e1678f9e50db\\\",\\\"systemUUID\\\":\\\"f76b32e4-2764-4219-bbf4-f7e56d76df3b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:40Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:40 crc kubenswrapper[4824]: I1211 02:01:40.799081 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:40 crc kubenswrapper[4824]: I1211 02:01:40.799166 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:40 crc kubenswrapper[4824]: I1211 02:01:40.799183 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:40 crc kubenswrapper[4824]: I1211 02:01:40.799207 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:40 crc kubenswrapper[4824]: I1211 02:01:40.799225 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:40Z","lastTransitionTime":"2025-12-11T02:01:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:40 crc kubenswrapper[4824]: E1211 02:01:40.815033 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:01:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:01:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:01:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:01:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d57bece6-6fc6-4a84-b8a8-e1678f9e50db\\\",\\\"systemUUID\\\":\\\"f76b32e4-2764-4219-bbf4-f7e56d76df3b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:40Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:40 crc kubenswrapper[4824]: I1211 02:01:40.820005 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:40 crc kubenswrapper[4824]: I1211 02:01:40.820066 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:40 crc kubenswrapper[4824]: I1211 02:01:40.820082 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:40 crc kubenswrapper[4824]: I1211 02:01:40.820114 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:40 crc kubenswrapper[4824]: I1211 02:01:40.820161 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:40Z","lastTransitionTime":"2025-12-11T02:01:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:40 crc kubenswrapper[4824]: E1211 02:01:40.839849 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:01:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:01:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:01:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:01:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d57bece6-6fc6-4a84-b8a8-e1678f9e50db\\\",\\\"systemUUID\\\":\\\"f76b32e4-2764-4219-bbf4-f7e56d76df3b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:40Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:40 crc kubenswrapper[4824]: I1211 02:01:40.844980 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:40 crc kubenswrapper[4824]: I1211 02:01:40.845043 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:40 crc kubenswrapper[4824]: I1211 02:01:40.845069 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:40 crc kubenswrapper[4824]: I1211 02:01:40.845096 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:40 crc kubenswrapper[4824]: I1211 02:01:40.845149 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:40Z","lastTransitionTime":"2025-12-11T02:01:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:40 crc kubenswrapper[4824]: E1211 02:01:40.860402 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:01:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:01:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:01:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:01:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d57bece6-6fc6-4a84-b8a8-e1678f9e50db\\\",\\\"systemUUID\\\":\\\"f76b32e4-2764-4219-bbf4-f7e56d76df3b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:40Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:40 crc kubenswrapper[4824]: I1211 02:01:40.864764 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:40 crc kubenswrapper[4824]: I1211 02:01:40.864825 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:40 crc kubenswrapper[4824]: I1211 02:01:40.864843 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:40 crc kubenswrapper[4824]: I1211 02:01:40.864869 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:40 crc kubenswrapper[4824]: I1211 02:01:40.864889 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:40Z","lastTransitionTime":"2025-12-11T02:01:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:40 crc kubenswrapper[4824]: E1211 02:01:40.883776 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:01:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:01:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:01:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:01:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d57bece6-6fc6-4a84-b8a8-e1678f9e50db\\\",\\\"systemUUID\\\":\\\"f76b32e4-2764-4219-bbf4-f7e56d76df3b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:40Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:40 crc kubenswrapper[4824]: E1211 02:01:40.884068 4824 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 11 02:01:40 crc kubenswrapper[4824]: I1211 02:01:40.886232 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:40 crc kubenswrapper[4824]: I1211 02:01:40.886299 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:40 crc kubenswrapper[4824]: I1211 02:01:40.886312 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:40 crc kubenswrapper[4824]: I1211 02:01:40.886330 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:40 crc kubenswrapper[4824]: I1211 02:01:40.886361 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:40Z","lastTransitionTime":"2025-12-11T02:01:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:40 crc kubenswrapper[4824]: I1211 02:01:40.989301 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:40 crc kubenswrapper[4824]: I1211 02:01:40.989378 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:40 crc kubenswrapper[4824]: I1211 02:01:40.989398 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:40 crc kubenswrapper[4824]: I1211 02:01:40.989744 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:40 crc kubenswrapper[4824]: I1211 02:01:40.989786 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:40Z","lastTransitionTime":"2025-12-11T02:01:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:41 crc kubenswrapper[4824]: I1211 02:01:41.092622 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:41 crc kubenswrapper[4824]: I1211 02:01:41.092702 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:41 crc kubenswrapper[4824]: I1211 02:01:41.092720 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:41 crc kubenswrapper[4824]: I1211 02:01:41.093323 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:41 crc kubenswrapper[4824]: I1211 02:01:41.093510 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:41Z","lastTransitionTime":"2025-12-11T02:01:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:41 crc kubenswrapper[4824]: I1211 02:01:41.197075 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:41 crc kubenswrapper[4824]: I1211 02:01:41.197215 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:41 crc kubenswrapper[4824]: I1211 02:01:41.197237 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:41 crc kubenswrapper[4824]: I1211 02:01:41.197263 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:41 crc kubenswrapper[4824]: I1211 02:01:41.197284 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:41Z","lastTransitionTime":"2025-12-11T02:01:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:41 crc kubenswrapper[4824]: I1211 02:01:41.300601 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:41 crc kubenswrapper[4824]: I1211 02:01:41.300986 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:41 crc kubenswrapper[4824]: I1211 02:01:41.301175 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:41 crc kubenswrapper[4824]: I1211 02:01:41.301388 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:41 crc kubenswrapper[4824]: I1211 02:01:41.301676 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:41Z","lastTransitionTime":"2025-12-11T02:01:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:41 crc kubenswrapper[4824]: I1211 02:01:41.404656 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:41 crc kubenswrapper[4824]: I1211 02:01:41.404711 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:41 crc kubenswrapper[4824]: I1211 02:01:41.404728 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:41 crc kubenswrapper[4824]: I1211 02:01:41.404752 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:41 crc kubenswrapper[4824]: I1211 02:01:41.404769 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:41Z","lastTransitionTime":"2025-12-11T02:01:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:41 crc kubenswrapper[4824]: I1211 02:01:41.508327 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:41 crc kubenswrapper[4824]: I1211 02:01:41.508427 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:41 crc kubenswrapper[4824]: I1211 02:01:41.508444 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:41 crc kubenswrapper[4824]: I1211 02:01:41.508537 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:41 crc kubenswrapper[4824]: I1211 02:01:41.508555 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:41Z","lastTransitionTime":"2025-12-11T02:01:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:41 crc kubenswrapper[4824]: I1211 02:01:41.611704 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:41 crc kubenswrapper[4824]: I1211 02:01:41.611776 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:41 crc kubenswrapper[4824]: I1211 02:01:41.611794 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:41 crc kubenswrapper[4824]: I1211 02:01:41.611816 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:41 crc kubenswrapper[4824]: I1211 02:01:41.611833 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:41Z","lastTransitionTime":"2025-12-11T02:01:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:41 crc kubenswrapper[4824]: I1211 02:01:41.632599 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 02:01:41 crc kubenswrapper[4824]: I1211 02:01:41.632633 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 02:01:41 crc kubenswrapper[4824]: I1211 02:01:41.632628 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 02:01:41 crc kubenswrapper[4824]: I1211 02:01:41.632775 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-grnpt" Dec 11 02:01:41 crc kubenswrapper[4824]: E1211 02:01:41.632905 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 02:01:41 crc kubenswrapper[4824]: E1211 02:01:41.633048 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 02:01:41 crc kubenswrapper[4824]: E1211 02:01:41.633211 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 02:01:41 crc kubenswrapper[4824]: E1211 02:01:41.633341 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-grnpt" podUID="09a52948-0404-45ff-85b1-51479268bf71" Dec 11 02:01:41 crc kubenswrapper[4824]: I1211 02:01:41.714611 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:41 crc kubenswrapper[4824]: I1211 02:01:41.714677 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:41 crc kubenswrapper[4824]: I1211 02:01:41.714738 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:41 crc kubenswrapper[4824]: I1211 02:01:41.714768 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:41 crc kubenswrapper[4824]: I1211 02:01:41.714790 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:41Z","lastTransitionTime":"2025-12-11T02:01:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:41 crc kubenswrapper[4824]: I1211 02:01:41.818539 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:41 crc kubenswrapper[4824]: I1211 02:01:41.818601 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:41 crc kubenswrapper[4824]: I1211 02:01:41.818623 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:41 crc kubenswrapper[4824]: I1211 02:01:41.818651 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:41 crc kubenswrapper[4824]: I1211 02:01:41.818673 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:41Z","lastTransitionTime":"2025-12-11T02:01:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:41 crc kubenswrapper[4824]: I1211 02:01:41.921976 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:41 crc kubenswrapper[4824]: I1211 02:01:41.922045 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:41 crc kubenswrapper[4824]: I1211 02:01:41.922064 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:41 crc kubenswrapper[4824]: I1211 02:01:41.922091 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:41 crc kubenswrapper[4824]: I1211 02:01:41.922143 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:41Z","lastTransitionTime":"2025-12-11T02:01:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:42 crc kubenswrapper[4824]: I1211 02:01:42.025386 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:42 crc kubenswrapper[4824]: I1211 02:01:42.025437 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:42 crc kubenswrapper[4824]: I1211 02:01:42.025458 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:42 crc kubenswrapper[4824]: I1211 02:01:42.025486 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:42 crc kubenswrapper[4824]: I1211 02:01:42.025507 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:42Z","lastTransitionTime":"2025-12-11T02:01:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:42 crc kubenswrapper[4824]: I1211 02:01:42.128477 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:42 crc kubenswrapper[4824]: I1211 02:01:42.128546 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:42 crc kubenswrapper[4824]: I1211 02:01:42.128571 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:42 crc kubenswrapper[4824]: I1211 02:01:42.128603 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:42 crc kubenswrapper[4824]: I1211 02:01:42.128625 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:42Z","lastTransitionTime":"2025-12-11T02:01:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:42 crc kubenswrapper[4824]: I1211 02:01:42.232059 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:42 crc kubenswrapper[4824]: I1211 02:01:42.232166 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:42 crc kubenswrapper[4824]: I1211 02:01:42.232190 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:42 crc kubenswrapper[4824]: I1211 02:01:42.232219 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:42 crc kubenswrapper[4824]: I1211 02:01:42.232240 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:42Z","lastTransitionTime":"2025-12-11T02:01:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:42 crc kubenswrapper[4824]: I1211 02:01:42.335427 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:42 crc kubenswrapper[4824]: I1211 02:01:42.335484 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:42 crc kubenswrapper[4824]: I1211 02:01:42.335501 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:42 crc kubenswrapper[4824]: I1211 02:01:42.335521 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:42 crc kubenswrapper[4824]: I1211 02:01:42.335535 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:42Z","lastTransitionTime":"2025-12-11T02:01:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:42 crc kubenswrapper[4824]: I1211 02:01:42.438961 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:42 crc kubenswrapper[4824]: I1211 02:01:42.439010 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:42 crc kubenswrapper[4824]: I1211 02:01:42.439026 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:42 crc kubenswrapper[4824]: I1211 02:01:42.439048 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:42 crc kubenswrapper[4824]: I1211 02:01:42.439066 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:42Z","lastTransitionTime":"2025-12-11T02:01:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:42 crc kubenswrapper[4824]: I1211 02:01:42.541937 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:42 crc kubenswrapper[4824]: I1211 02:01:42.541992 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:42 crc kubenswrapper[4824]: I1211 02:01:42.542014 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:42 crc kubenswrapper[4824]: I1211 02:01:42.542041 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:42 crc kubenswrapper[4824]: I1211 02:01:42.542063 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:42Z","lastTransitionTime":"2025-12-11T02:01:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:42 crc kubenswrapper[4824]: I1211 02:01:42.633085 4824 scope.go:117] "RemoveContainer" containerID="c803a25bdafcd4b9b8a4dfbcf8b67d6ade84b490f9d57f46873c32f9c763401d" Dec 11 02:01:42 crc kubenswrapper[4824]: I1211 02:01:42.645978 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:42 crc kubenswrapper[4824]: I1211 02:01:42.646164 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:42 crc kubenswrapper[4824]: I1211 02:01:42.646200 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:42 crc kubenswrapper[4824]: I1211 02:01:42.646291 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:42 crc kubenswrapper[4824]: I1211 02:01:42.646365 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:42Z","lastTransitionTime":"2025-12-11T02:01:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:42 crc kubenswrapper[4824]: I1211 02:01:42.749079 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:42 crc kubenswrapper[4824]: I1211 02:01:42.749187 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:42 crc kubenswrapper[4824]: I1211 02:01:42.749208 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:42 crc kubenswrapper[4824]: I1211 02:01:42.749231 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:42 crc kubenswrapper[4824]: I1211 02:01:42.749288 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:42Z","lastTransitionTime":"2025-12-11T02:01:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:42 crc kubenswrapper[4824]: I1211 02:01:42.853510 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:42 crc kubenswrapper[4824]: I1211 02:01:42.853580 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:42 crc kubenswrapper[4824]: I1211 02:01:42.853594 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:42 crc kubenswrapper[4824]: I1211 02:01:42.853631 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:42 crc kubenswrapper[4824]: I1211 02:01:42.853646 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:42Z","lastTransitionTime":"2025-12-11T02:01:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:42 crc kubenswrapper[4824]: I1211 02:01:42.957557 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:42 crc kubenswrapper[4824]: I1211 02:01:42.957602 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:42 crc kubenswrapper[4824]: I1211 02:01:42.957614 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:42 crc kubenswrapper[4824]: I1211 02:01:42.957630 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:42 crc kubenswrapper[4824]: I1211 02:01:42.957642 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:42Z","lastTransitionTime":"2025-12-11T02:01:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:43 crc kubenswrapper[4824]: I1211 02:01:43.065319 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:43 crc kubenswrapper[4824]: I1211 02:01:43.065370 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:43 crc kubenswrapper[4824]: I1211 02:01:43.065388 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:43 crc kubenswrapper[4824]: I1211 02:01:43.065413 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:43 crc kubenswrapper[4824]: I1211 02:01:43.065433 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:43Z","lastTransitionTime":"2025-12-11T02:01:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:43 crc kubenswrapper[4824]: I1211 02:01:43.066561 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f6pgc_5b193bd2-b536-4056-92f1-94c9836ab2eb/ovnkube-controller/1.log" Dec 11 02:01:43 crc kubenswrapper[4824]: I1211 02:01:43.072375 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" event={"ID":"5b193bd2-b536-4056-92f1-94c9836ab2eb","Type":"ContainerStarted","Data":"09bb8b4c85e1fd5fa8d62db23ebee77a25a0faa274da3bb22ab679987597f26b"} Dec 11 02:01:43 crc kubenswrapper[4824]: I1211 02:01:43.072572 4824 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 11 02:01:43 crc kubenswrapper[4824]: I1211 02:01:43.096384 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:43Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:43 crc kubenswrapper[4824]: I1211 02:01:43.123636 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"253d4291-2b2f-487c-8862-3d305f0bdc55\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44b29bb7a6eda7cb890d36f86375b5f3f197ae5cf390f16e71139b0904873e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://922af7f67b99a0ce08776f3f75c0c377e6ea69af1720b6178a6d1e2f9b4b8ba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://136c68020325db8e6930d0f40b0d0d7e013c14c4dd5e7d1449d01d6d458104d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c89afaf56b4a9597f6543ec2e3b176c7a6dab7dc341941fc3bc8ec85cc4edaf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:00:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:43Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:43 crc kubenswrapper[4824]: I1211 02:01:43.140253 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76b10bbd1e891b8c682f4b2e17d023fd787a8ed44ed94fbf7d97314d641ccf36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f32838f33832b9b67e7dd9c53340680e17721cf1c7bdf0170230c5f8feafe64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:43Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:43 crc kubenswrapper[4824]: I1211 02:01:43.158841 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kptws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68568dba-83a5-46c1-8d01-337359679d91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f965a6d6db619995132f61b5d51b66f0b2b3fdbef3fbec63e231b9a62ccd0a9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd7nf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kptws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:43Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:43 crc kubenswrapper[4824]: I1211 02:01:43.167687 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:43 crc kubenswrapper[4824]: I1211 02:01:43.167742 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:43 crc kubenswrapper[4824]: I1211 02:01:43.167760 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:43 crc kubenswrapper[4824]: I1211 02:01:43.167784 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:43 crc kubenswrapper[4824]: I1211 02:01:43.167801 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:43Z","lastTransitionTime":"2025-12-11T02:01:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:43 crc kubenswrapper[4824]: I1211 02:01:43.176304 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"beab386b-6c65-4bc3-9ef0-dc3f68b4403f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c26f06d2ed5e6e98b61a51f073130675f2c49acda106b8aedcbb2a6d4f7580c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20af9b10a67c8b21a7abcd23e42bc412798d159c8422e7c8ff58f5b997ca19a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9bbfc6edfe050b3fa9dbeee5f98d416bcb2a1533a38608a881b75396fd0b059\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7c766f8de73ca762ffde0c18b2d1c46bb152b2b37535a611c5a0e454a21b7d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d45819f0611c91b20c83eeee1092031a3441fb69f3015241632ad419880227d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 02:01:11.070250 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 02:01:11.071673 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2752882565/tls.crt::/tmp/serving-cert-2752882565/tls.key\\\\\\\"\\\\nI1211 02:01:17.042610 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 02:01:17.048046 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 02:01:17.048080 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 02:01:17.048172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 02:01:17.048188 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 02:01:17.058248 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 02:01:17.058294 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 02:01:17.058303 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 02:01:17.058315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 02:01:17.058322 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 02:01:17.058328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 02:01:17.058335 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 02:01:17.058808 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 02:01:17.060225 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14a940a72b2404a6cbd52a69e150b03266bcd7f5c1d3f4d07b51c0c69e51786\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:00:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:43Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:43 crc kubenswrapper[4824]: I1211 02:01:43.190934 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ded146c045085be2b58821749052413a36850ede00ae1024fe2dae9fa5ac84b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:43Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:43 crc kubenswrapper[4824]: I1211 02:01:43.213532 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b193bd2-b536-4056-92f1-94c9836ab2eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51f3c41d86ddc44f7eca118029876ad353317f753d0ad173ed4d30c2a15f10e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c345850891744a0240781277e9a89a59f6899c4c944995f62e44650f807440a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed2ae40b1fe28111efc76d9d72345edecb5f3ea34522b238f3730812b29f2bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce7aa078800e6f299cfe03770fa7e3b64fa3742fe0c0f383c87d84cf48569ddf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6aa9fc1f1bff192c69bd59d32a05d4a72d09641d9ce4b265bc4517147a7e15dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b4a3bc50eb80777188206adab774e8acad8b2e4952a8415f91b1c29ba9a0ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09bb8b4c85e1fd5fa8d62db23ebee77a25a0faa274da3bb22ab679987597f26b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c803a25bdafcd4b9b8a4dfbcf8b67d6ade84b490f9d57f46873c32f9c763401d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T02:01:29Z\\\",\\\"message\\\":\\\"ode_controller event handlers\\\\nI1211 02:01:29.024367 6208 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1211 02:01:29.024404 6208 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1211 02:01:29.024418 6208 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1211 02:01:29.024439 6208 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1211 02:01:29.024462 6208 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1211 02:01:29.024470 6208 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1211 02:01:29.024492 6208 factory.go:656] Stopping watch factory\\\\nI1211 02:01:29.024509 6208 ovnkube.go:599] Stopped ovnkube\\\\nI1211 02:01:29.024540 6208 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1211 02:01:29.024559 6208 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1211 02:01:29.024571 6208 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1211 02:01:29.024582 6208 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1211 02:01:29.024593 6208 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1211 02:01:29.024604 6208 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1211 02:01:29.024620 6208 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1211 02:01:29.024705 6208 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ead58694813806d307bbbc6a7b3cf41730d1c059b231a3002684bf09c10c7f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f6pgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:43Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:43 crc kubenswrapper[4824]: I1211 02:01:43.233206 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-br4j9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce125194-9c9d-4ea4-8a72-2c30d5722295\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9798f904052e08196a0db1745060dab11e2710308bca01ff160475910f2665df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fc5s8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-br4j9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:43Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:43 crc kubenswrapper[4824]: I1211 02:01:43.246595 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8v4jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99a6bfee-afca-4726-a964-b95a948945a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b23b9d8f18db3eeadb60f83b36b05b1778cd0fa31669dad2680bb335ae004acf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4hk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3acc1940a0534ba0e6096175e41834fdd0fbcbb4c3fb6e9a38e76aedae55d2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4hk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-8v4jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:43Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:43 crc kubenswrapper[4824]: I1211 02:01:43.259377 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-grnpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09a52948-0404-45ff-85b1-51479268bf71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmmzm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmmzm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-grnpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:43Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:43 crc kubenswrapper[4824]: I1211 02:01:43.270723 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:43 crc kubenswrapper[4824]: I1211 02:01:43.270765 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:43 crc kubenswrapper[4824]: I1211 02:01:43.270777 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:43 crc kubenswrapper[4824]: I1211 02:01:43.270796 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:43 crc kubenswrapper[4824]: I1211 02:01:43.270809 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:43Z","lastTransitionTime":"2025-12-11T02:01:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:43 crc kubenswrapper[4824]: I1211 02:01:43.275971 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:43Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:43 crc kubenswrapper[4824]: I1211 02:01:43.289236 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kh8g9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"50c69d41-a65b-463e-a829-ea349bd2fe46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dca6be41bd7a3ce12ecd4567c868f696b65e8eb1bad436feda6e4f9bfb8e525\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69af1f33fd0bafc929b94e702b4f46c6c7d7a362d5552b121729c477892d91d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69af1f33fd0bafc929b94e702b4f46c6c7d7a362d5552b121729c477892d91d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd9675d0afded587bc660a3862fd0c5dd6885961afacda957aa9ee8501d68a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd9675d0afded587bc660a3862fd0c5dd6885961afacda957aa9ee8501d68a69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aac94149458bd6321be90b26b48ee26428d273e7f61407d95f9c9d6b3aeae4b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aac94149458bd6321be90b26b48ee26428d273e7f61407d95f9c9d6b3aeae4b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbee11bd7cb65e031415990869b52d2b8f81bf8b56e6cc935d38e09695db7b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbee11bd7cb65e031415990869b52d2b8f81bf8b56e6cc935d38e09695db7b48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86f693ade72bd593bc16021fe4b7564f1914ec2f1cda7039474ac2fb13de10f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86f693ade72bd593bc16021fe4b7564f1914ec2f1cda7039474ac2fb13de10f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7d7674252f67bde0ded2ec8ad1d925a92a26d3e3f741dee0061b735cb0b7a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7d7674252f67bde0ded2ec8ad1d925a92a26d3e3f741dee0061b735cb0b7a02\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kh8g9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:43Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:43 crc kubenswrapper[4824]: I1211 02:01:43.305466 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6bkc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8022f696-d0e7-437b-848a-3cd25bd1f364\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3e28f6778ec0dba6ea3e6b69e5a2c1c73c0d5db1abeead241fb58efdbfaefe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hh5kn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6bkc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:43Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:43 crc kubenswrapper[4824]: I1211 02:01:43.321089 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d14986666febd651039afb52ddf4db8969aca55e88eda9bbb20a67e35ba722ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:43Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:43 crc kubenswrapper[4824]: I1211 02:01:43.336311 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:43Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:43 crc kubenswrapper[4824]: I1211 02:01:43.348853 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44b156e4-64a4-4d45-aa5e-9b10a862faed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7368d92144b188e95a7796fabf497b0f8d9c0b240293bdc7b4aec1ff75497bc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w62dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcf289ac27dbda4f24aabbde3fb5119329e7b4fc7d6e018168da9a69661a4c57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w62dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gx6xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:43Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:43 crc kubenswrapper[4824]: I1211 02:01:43.373888 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:43 crc kubenswrapper[4824]: I1211 02:01:43.373928 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:43 crc kubenswrapper[4824]: I1211 02:01:43.373940 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:43 crc kubenswrapper[4824]: I1211 02:01:43.373957 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:43 crc kubenswrapper[4824]: I1211 02:01:43.373969 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:43Z","lastTransitionTime":"2025-12-11T02:01:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:43 crc kubenswrapper[4824]: I1211 02:01:43.389025 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" Dec 11 02:01:43 crc kubenswrapper[4824]: I1211 02:01:43.476625 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:43 crc kubenswrapper[4824]: I1211 02:01:43.476688 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:43 crc kubenswrapper[4824]: I1211 02:01:43.476706 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:43 crc kubenswrapper[4824]: I1211 02:01:43.476732 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:43 crc kubenswrapper[4824]: I1211 02:01:43.476755 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:43Z","lastTransitionTime":"2025-12-11T02:01:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:43 crc kubenswrapper[4824]: I1211 02:01:43.579389 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:43 crc kubenswrapper[4824]: I1211 02:01:43.579438 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:43 crc kubenswrapper[4824]: I1211 02:01:43.579447 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:43 crc kubenswrapper[4824]: I1211 02:01:43.579460 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:43 crc kubenswrapper[4824]: I1211 02:01:43.579468 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:43Z","lastTransitionTime":"2025-12-11T02:01:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:43 crc kubenswrapper[4824]: I1211 02:01:43.632432 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 02:01:43 crc kubenswrapper[4824]: I1211 02:01:43.632448 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 02:01:43 crc kubenswrapper[4824]: I1211 02:01:43.632459 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 02:01:43 crc kubenswrapper[4824]: E1211 02:01:43.632533 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 02:01:43 crc kubenswrapper[4824]: I1211 02:01:43.632622 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-grnpt" Dec 11 02:01:43 crc kubenswrapper[4824]: E1211 02:01:43.632799 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 02:01:43 crc kubenswrapper[4824]: E1211 02:01:43.633005 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-grnpt" podUID="09a52948-0404-45ff-85b1-51479268bf71" Dec 11 02:01:43 crc kubenswrapper[4824]: E1211 02:01:43.633079 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 02:01:43 crc kubenswrapper[4824]: I1211 02:01:43.682087 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:43 crc kubenswrapper[4824]: I1211 02:01:43.682178 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:43 crc kubenswrapper[4824]: I1211 02:01:43.682202 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:43 crc kubenswrapper[4824]: I1211 02:01:43.682239 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:43 crc kubenswrapper[4824]: I1211 02:01:43.682263 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:43Z","lastTransitionTime":"2025-12-11T02:01:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:43 crc kubenswrapper[4824]: I1211 02:01:43.785313 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:43 crc kubenswrapper[4824]: I1211 02:01:43.785388 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:43 crc kubenswrapper[4824]: I1211 02:01:43.785412 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:43 crc kubenswrapper[4824]: I1211 02:01:43.785441 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:43 crc kubenswrapper[4824]: I1211 02:01:43.785464 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:43Z","lastTransitionTime":"2025-12-11T02:01:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:43 crc kubenswrapper[4824]: I1211 02:01:43.889167 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:43 crc kubenswrapper[4824]: I1211 02:01:43.889228 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:43 crc kubenswrapper[4824]: I1211 02:01:43.889244 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:43 crc kubenswrapper[4824]: I1211 02:01:43.889284 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:43 crc kubenswrapper[4824]: I1211 02:01:43.889307 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:43Z","lastTransitionTime":"2025-12-11T02:01:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:43 crc kubenswrapper[4824]: I1211 02:01:43.991976 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:43 crc kubenswrapper[4824]: I1211 02:01:43.992052 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:43 crc kubenswrapper[4824]: I1211 02:01:43.992077 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:43 crc kubenswrapper[4824]: I1211 02:01:43.992145 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:43 crc kubenswrapper[4824]: I1211 02:01:43.992165 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:43Z","lastTransitionTime":"2025-12-11T02:01:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:44 crc kubenswrapper[4824]: I1211 02:01:44.080062 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f6pgc_5b193bd2-b536-4056-92f1-94c9836ab2eb/ovnkube-controller/2.log" Dec 11 02:01:44 crc kubenswrapper[4824]: I1211 02:01:44.081177 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f6pgc_5b193bd2-b536-4056-92f1-94c9836ab2eb/ovnkube-controller/1.log" Dec 11 02:01:44 crc kubenswrapper[4824]: I1211 02:01:44.085931 4824 generic.go:334] "Generic (PLEG): container finished" podID="5b193bd2-b536-4056-92f1-94c9836ab2eb" containerID="09bb8b4c85e1fd5fa8d62db23ebee77a25a0faa274da3bb22ab679987597f26b" exitCode=1 Dec 11 02:01:44 crc kubenswrapper[4824]: I1211 02:01:44.086009 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" event={"ID":"5b193bd2-b536-4056-92f1-94c9836ab2eb","Type":"ContainerDied","Data":"09bb8b4c85e1fd5fa8d62db23ebee77a25a0faa274da3bb22ab679987597f26b"} Dec 11 02:01:44 crc kubenswrapper[4824]: I1211 02:01:44.086152 4824 scope.go:117] "RemoveContainer" containerID="c803a25bdafcd4b9b8a4dfbcf8b67d6ade84b490f9d57f46873c32f9c763401d" Dec 11 02:01:44 crc kubenswrapper[4824]: I1211 02:01:44.087001 4824 scope.go:117] "RemoveContainer" containerID="09bb8b4c85e1fd5fa8d62db23ebee77a25a0faa274da3bb22ab679987597f26b" Dec 11 02:01:44 crc kubenswrapper[4824]: E1211 02:01:44.087308 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-f6pgc_openshift-ovn-kubernetes(5b193bd2-b536-4056-92f1-94c9836ab2eb)\"" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" podUID="5b193bd2-b536-4056-92f1-94c9836ab2eb" Dec 11 02:01:44 crc kubenswrapper[4824]: I1211 02:01:44.094578 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:44 crc kubenswrapper[4824]: I1211 02:01:44.094656 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:44 crc kubenswrapper[4824]: I1211 02:01:44.094683 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:44 crc kubenswrapper[4824]: I1211 02:01:44.094866 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:44 crc kubenswrapper[4824]: I1211 02:01:44.094892 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:44Z","lastTransitionTime":"2025-12-11T02:01:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:44 crc kubenswrapper[4824]: I1211 02:01:44.108812 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"253d4291-2b2f-487c-8862-3d305f0bdc55\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44b29bb7a6eda7cb890d36f86375b5f3f197ae5cf390f16e71139b0904873e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://922af7f67b99a0ce08776f3f75c0c377e6ea69af1720b6178a6d1e2f9b4b8ba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://136c68020325db8e6930d0f40b0d0d7e013c14c4dd5e7d1449d01d6d458104d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c89afaf56b4a9597f6543ec2e3b176c7a6dab7dc341941fc3bc8ec85cc4edaf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:00:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:44Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:44 crc kubenswrapper[4824]: I1211 02:01:44.131936 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:44Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:44 crc kubenswrapper[4824]: I1211 02:01:44.153962 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"beab386b-6c65-4bc3-9ef0-dc3f68b4403f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c26f06d2ed5e6e98b61a51f073130675f2c49acda106b8aedcbb2a6d4f7580c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20af9b10a67c8b21a7abcd23e42bc412798d159c8422e7c8ff58f5b997ca19a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9bbfc6edfe050b3fa9dbeee5f98d416bcb2a1533a38608a881b75396fd0b059\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7c766f8de73ca762ffde0c18b2d1c46bb152b2b37535a611c5a0e454a21b7d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d45819f0611c91b20c83eeee1092031a3441fb69f3015241632ad419880227d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 02:01:11.070250 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 02:01:11.071673 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2752882565/tls.crt::/tmp/serving-cert-2752882565/tls.key\\\\\\\"\\\\nI1211 02:01:17.042610 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 02:01:17.048046 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 02:01:17.048080 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 02:01:17.048172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 02:01:17.048188 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 02:01:17.058248 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 02:01:17.058294 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 02:01:17.058303 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 02:01:17.058315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 02:01:17.058322 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 02:01:17.058328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 02:01:17.058335 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 02:01:17.058808 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 02:01:17.060225 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14a940a72b2404a6cbd52a69e150b03266bcd7f5c1d3f4d07b51c0c69e51786\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:00:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:44Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:44 crc kubenswrapper[4824]: I1211 02:01:44.179391 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ded146c045085be2b58821749052413a36850ede00ae1024fe2dae9fa5ac84b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:44Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:44 crc kubenswrapper[4824]: I1211 02:01:44.198749 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:44 crc kubenswrapper[4824]: I1211 02:01:44.198804 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:44 crc kubenswrapper[4824]: I1211 02:01:44.198821 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:44 crc kubenswrapper[4824]: I1211 02:01:44.198845 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:44 crc kubenswrapper[4824]: I1211 02:01:44.198863 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:44Z","lastTransitionTime":"2025-12-11T02:01:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:44 crc kubenswrapper[4824]: I1211 02:01:44.200163 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76b10bbd1e891b8c682f4b2e17d023fd787a8ed44ed94fbf7d97314d641ccf36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f32838f33832b9b67e7dd9c53340680e17721cf1c7bdf0170230c5f8feafe64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:44Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:44 crc kubenswrapper[4824]: I1211 02:01:44.214970 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kptws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68568dba-83a5-46c1-8d01-337359679d91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f965a6d6db619995132f61b5d51b66f0b2b3fdbef3fbec63e231b9a62ccd0a9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd7nf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kptws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:44Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:44 crc kubenswrapper[4824]: I1211 02:01:44.232754 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8v4jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99a6bfee-afca-4726-a964-b95a948945a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b23b9d8f18db3eeadb60f83b36b05b1778cd0fa31669dad2680bb335ae004acf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4hk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3acc1940a0534ba0e6096175e41834fdd0fbcbb4c3fb6e9a38e76aedae55d2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4hk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-8v4jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:44Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:44 crc kubenswrapper[4824]: I1211 02:01:44.248047 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-grnpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09a52948-0404-45ff-85b1-51479268bf71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmmzm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmmzm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-grnpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:44Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:44 crc kubenswrapper[4824]: I1211 02:01:44.262511 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:44Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:44 crc kubenswrapper[4824]: I1211 02:01:44.280531 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kh8g9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"50c69d41-a65b-463e-a829-ea349bd2fe46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dca6be41bd7a3ce12ecd4567c868f696b65e8eb1bad436feda6e4f9bfb8e525\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69af1f33fd0bafc929b94e702b4f46c6c7d7a362d5552b121729c477892d91d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69af1f33fd0bafc929b94e702b4f46c6c7d7a362d5552b121729c477892d91d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd9675d0afded587bc660a3862fd0c5dd6885961afacda957aa9ee8501d68a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd9675d0afded587bc660a3862fd0c5dd6885961afacda957aa9ee8501d68a69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aac94149458bd6321be90b26b48ee26428d273e7f61407d95f9c9d6b3aeae4b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aac94149458bd6321be90b26b48ee26428d273e7f61407d95f9c9d6b3aeae4b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbee11bd7cb65e031415990869b52d2b8f81bf8b56e6cc935d38e09695db7b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbee11bd7cb65e031415990869b52d2b8f81bf8b56e6cc935d38e09695db7b48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86f693ade72bd593bc16021fe4b7564f1914ec2f1cda7039474ac2fb13de10f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86f693ade72bd593bc16021fe4b7564f1914ec2f1cda7039474ac2fb13de10f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7d7674252f67bde0ded2ec8ad1d925a92a26d3e3f741dee0061b735cb0b7a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7d7674252f67bde0ded2ec8ad1d925a92a26d3e3f741dee0061b735cb0b7a02\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kh8g9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:44Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:44 crc kubenswrapper[4824]: I1211 02:01:44.302494 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:44 crc kubenswrapper[4824]: I1211 02:01:44.302464 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6bkc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8022f696-d0e7-437b-848a-3cd25bd1f364\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3e28f6778ec0dba6ea3e6b69e5a2c1c73c0d5db1abeead241fb58efdbfaefe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hh5kn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6bkc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:44Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:44 crc kubenswrapper[4824]: I1211 02:01:44.302550 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:44 crc kubenswrapper[4824]: I1211 02:01:44.302569 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:44 crc kubenswrapper[4824]: I1211 02:01:44.302594 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:44 crc kubenswrapper[4824]: I1211 02:01:44.302612 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:44Z","lastTransitionTime":"2025-12-11T02:01:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:44 crc kubenswrapper[4824]: I1211 02:01:44.332607 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b193bd2-b536-4056-92f1-94c9836ab2eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51f3c41d86ddc44f7eca118029876ad353317f753d0ad173ed4d30c2a15f10e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c345850891744a0240781277e9a89a59f6899c4c944995f62e44650f807440a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed2ae40b1fe28111efc76d9d72345edecb5f3ea34522b238f3730812b29f2bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce7aa078800e6f299cfe03770fa7e3b64fa3742fe0c0f383c87d84cf48569ddf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6aa9fc1f1bff192c69bd59d32a05d4a72d09641d9ce4b265bc4517147a7e15dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b4a3bc50eb80777188206adab774e8acad8b2e4952a8415f91b1c29ba9a0ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09bb8b4c85e1fd5fa8d62db23ebee77a25a0faa274da3bb22ab679987597f26b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c803a25bdafcd4b9b8a4dfbcf8b67d6ade84b490f9d57f46873c32f9c763401d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T02:01:29Z\\\",\\\"message\\\":\\\"ode_controller event handlers\\\\nI1211 02:01:29.024367 6208 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1211 02:01:29.024404 6208 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1211 02:01:29.024418 6208 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1211 02:01:29.024439 6208 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1211 02:01:29.024462 6208 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1211 02:01:29.024470 6208 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1211 02:01:29.024492 6208 factory.go:656] Stopping watch factory\\\\nI1211 02:01:29.024509 6208 ovnkube.go:599] Stopped ovnkube\\\\nI1211 02:01:29.024540 6208 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1211 02:01:29.024559 6208 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1211 02:01:29.024571 6208 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1211 02:01:29.024582 6208 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1211 02:01:29.024593 6208 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1211 02:01:29.024604 6208 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1211 02:01:29.024620 6208 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1211 02:01:29.024705 6208 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09bb8b4c85e1fd5fa8d62db23ebee77a25a0faa274da3bb22ab679987597f26b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T02:01:43Z\\\",\\\"message\\\":\\\"ng]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-webhook\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.254\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1211 02:01:43.627342 6417 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-target-xd92c in node crc\\\\nI1211 02:01:43.627315 6417 services_controller.go:443] Built service openshift-machine-config-operator/machine-config-daemon LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.43\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:9001, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}, services.lbConfig{vips:[]string{\\\\\\\"10.217.4.43\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8798, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLo\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ead58694813806d307bbbc6a7b3cf41730d1c059b231a3002684bf09c10c7f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f6pgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:44Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:44 crc kubenswrapper[4824]: I1211 02:01:44.349925 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-br4j9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce125194-9c9d-4ea4-8a72-2c30d5722295\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9798f904052e08196a0db1745060dab11e2710308bca01ff160475910f2665df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fc5s8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-br4j9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:44Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:44 crc kubenswrapper[4824]: I1211 02:01:44.369243 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d14986666febd651039afb52ddf4db8969aca55e88eda9bbb20a67e35ba722ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:44Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:44 crc kubenswrapper[4824]: I1211 02:01:44.389918 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:44Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:44 crc kubenswrapper[4824]: I1211 02:01:44.405765 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:44 crc kubenswrapper[4824]: I1211 02:01:44.405806 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:44 crc kubenswrapper[4824]: I1211 02:01:44.405822 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:44 crc kubenswrapper[4824]: I1211 02:01:44.405846 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:44 crc kubenswrapper[4824]: I1211 02:01:44.405863 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:44Z","lastTransitionTime":"2025-12-11T02:01:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:44 crc kubenswrapper[4824]: I1211 02:01:44.409022 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44b156e4-64a4-4d45-aa5e-9b10a862faed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7368d92144b188e95a7796fabf497b0f8d9c0b240293bdc7b4aec1ff75497bc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w62dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcf289ac27dbda4f24aabbde3fb5119329e7b4fc7d6e018168da9a69661a4c57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w62dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gx6xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:44Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:44 crc kubenswrapper[4824]: I1211 02:01:44.508680 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:44 crc kubenswrapper[4824]: I1211 02:01:44.508740 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:44 crc kubenswrapper[4824]: I1211 02:01:44.508763 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:44 crc kubenswrapper[4824]: I1211 02:01:44.508793 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:44 crc kubenswrapper[4824]: I1211 02:01:44.508815 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:44Z","lastTransitionTime":"2025-12-11T02:01:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:44 crc kubenswrapper[4824]: I1211 02:01:44.611679 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:44 crc kubenswrapper[4824]: I1211 02:01:44.611723 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:44 crc kubenswrapper[4824]: I1211 02:01:44.611741 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:44 crc kubenswrapper[4824]: I1211 02:01:44.611763 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:44 crc kubenswrapper[4824]: I1211 02:01:44.611780 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:44Z","lastTransitionTime":"2025-12-11T02:01:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:44 crc kubenswrapper[4824]: I1211 02:01:44.715294 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:44 crc kubenswrapper[4824]: I1211 02:01:44.715356 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:44 crc kubenswrapper[4824]: I1211 02:01:44.715374 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:44 crc kubenswrapper[4824]: I1211 02:01:44.715397 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:44 crc kubenswrapper[4824]: I1211 02:01:44.715420 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:44Z","lastTransitionTime":"2025-12-11T02:01:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:44 crc kubenswrapper[4824]: I1211 02:01:44.818361 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:44 crc kubenswrapper[4824]: I1211 02:01:44.818419 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:44 crc kubenswrapper[4824]: I1211 02:01:44.818441 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:44 crc kubenswrapper[4824]: I1211 02:01:44.818470 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:44 crc kubenswrapper[4824]: I1211 02:01:44.818492 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:44Z","lastTransitionTime":"2025-12-11T02:01:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:44 crc kubenswrapper[4824]: I1211 02:01:44.921881 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:44 crc kubenswrapper[4824]: I1211 02:01:44.921942 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:44 crc kubenswrapper[4824]: I1211 02:01:44.921964 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:44 crc kubenswrapper[4824]: I1211 02:01:44.921991 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:44 crc kubenswrapper[4824]: I1211 02:01:44.922012 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:44Z","lastTransitionTime":"2025-12-11T02:01:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:45 crc kubenswrapper[4824]: I1211 02:01:45.025381 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:45 crc kubenswrapper[4824]: I1211 02:01:45.025526 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:45 crc kubenswrapper[4824]: I1211 02:01:45.025548 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:45 crc kubenswrapper[4824]: I1211 02:01:45.025579 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:45 crc kubenswrapper[4824]: I1211 02:01:45.025608 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:45Z","lastTransitionTime":"2025-12-11T02:01:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:45 crc kubenswrapper[4824]: I1211 02:01:45.092181 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f6pgc_5b193bd2-b536-4056-92f1-94c9836ab2eb/ovnkube-controller/2.log" Dec 11 02:01:45 crc kubenswrapper[4824]: I1211 02:01:45.097974 4824 scope.go:117] "RemoveContainer" containerID="09bb8b4c85e1fd5fa8d62db23ebee77a25a0faa274da3bb22ab679987597f26b" Dec 11 02:01:45 crc kubenswrapper[4824]: E1211 02:01:45.098271 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-f6pgc_openshift-ovn-kubernetes(5b193bd2-b536-4056-92f1-94c9836ab2eb)\"" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" podUID="5b193bd2-b536-4056-92f1-94c9836ab2eb" Dec 11 02:01:45 crc kubenswrapper[4824]: I1211 02:01:45.114420 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kptws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68568dba-83a5-46c1-8d01-337359679d91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f965a6d6db619995132f61b5d51b66f0b2b3fdbef3fbec63e231b9a62ccd0a9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd7nf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kptws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:45Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:45 crc kubenswrapper[4824]: I1211 02:01:45.128847 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:45 crc kubenswrapper[4824]: I1211 02:01:45.128957 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:45 crc kubenswrapper[4824]: I1211 02:01:45.128978 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:45 crc kubenswrapper[4824]: I1211 02:01:45.129007 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:45 crc kubenswrapper[4824]: I1211 02:01:45.129028 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:45Z","lastTransitionTime":"2025-12-11T02:01:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:45 crc kubenswrapper[4824]: I1211 02:01:45.136785 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"beab386b-6c65-4bc3-9ef0-dc3f68b4403f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c26f06d2ed5e6e98b61a51f073130675f2c49acda106b8aedcbb2a6d4f7580c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20af9b10a67c8b21a7abcd23e42bc412798d159c8422e7c8ff58f5b997ca19a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9bbfc6edfe050b3fa9dbeee5f98d416bcb2a1533a38608a881b75396fd0b059\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7c766f8de73ca762ffde0c18b2d1c46bb152b2b37535a611c5a0e454a21b7d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d45819f0611c91b20c83eeee1092031a3441fb69f3015241632ad419880227d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 02:01:11.070250 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 02:01:11.071673 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2752882565/tls.crt::/tmp/serving-cert-2752882565/tls.key\\\\\\\"\\\\nI1211 02:01:17.042610 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 02:01:17.048046 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 02:01:17.048080 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 02:01:17.048172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 02:01:17.048188 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 02:01:17.058248 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 02:01:17.058294 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 02:01:17.058303 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 02:01:17.058315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 02:01:17.058322 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 02:01:17.058328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 02:01:17.058335 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 02:01:17.058808 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 02:01:17.060225 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14a940a72b2404a6cbd52a69e150b03266bcd7f5c1d3f4d07b51c0c69e51786\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:00:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:45Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:45 crc kubenswrapper[4824]: I1211 02:01:45.158786 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ded146c045085be2b58821749052413a36850ede00ae1024fe2dae9fa5ac84b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:45Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:45 crc kubenswrapper[4824]: I1211 02:01:45.178856 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76b10bbd1e891b8c682f4b2e17d023fd787a8ed44ed94fbf7d97314d641ccf36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f32838f33832b9b67e7dd9c53340680e17721cf1c7bdf0170230c5f8feafe64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:45Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:45 crc kubenswrapper[4824]: I1211 02:01:45.195221 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-br4j9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce125194-9c9d-4ea4-8a72-2c30d5722295\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9798f904052e08196a0db1745060dab11e2710308bca01ff160475910f2665df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fc5s8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-br4j9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:45Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:45 crc kubenswrapper[4824]: I1211 02:01:45.213322 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8v4jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99a6bfee-afca-4726-a964-b95a948945a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b23b9d8f18db3eeadb60f83b36b05b1778cd0fa31669dad2680bb335ae004acf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4hk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3acc1940a0534ba0e6096175e41834fdd0fbcbb4c3fb6e9a38e76aedae55d2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4hk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-8v4jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:45Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:45 crc kubenswrapper[4824]: I1211 02:01:45.230771 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-grnpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09a52948-0404-45ff-85b1-51479268bf71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmmzm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmmzm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-grnpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:45Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:45 crc kubenswrapper[4824]: I1211 02:01:45.231953 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:45 crc kubenswrapper[4824]: I1211 02:01:45.232010 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:45 crc kubenswrapper[4824]: I1211 02:01:45.232028 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:45 crc kubenswrapper[4824]: I1211 02:01:45.232058 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:45 crc kubenswrapper[4824]: I1211 02:01:45.232080 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:45Z","lastTransitionTime":"2025-12-11T02:01:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:45 crc kubenswrapper[4824]: I1211 02:01:45.249654 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:45Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:45 crc kubenswrapper[4824]: I1211 02:01:45.273306 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kh8g9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"50c69d41-a65b-463e-a829-ea349bd2fe46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dca6be41bd7a3ce12ecd4567c868f696b65e8eb1bad436feda6e4f9bfb8e525\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69af1f33fd0bafc929b94e702b4f46c6c7d7a362d5552b121729c477892d91d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69af1f33fd0bafc929b94e702b4f46c6c7d7a362d5552b121729c477892d91d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd9675d0afded587bc660a3862fd0c5dd6885961afacda957aa9ee8501d68a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd9675d0afded587bc660a3862fd0c5dd6885961afacda957aa9ee8501d68a69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aac94149458bd6321be90b26b48ee26428d273e7f61407d95f9c9d6b3aeae4b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aac94149458bd6321be90b26b48ee26428d273e7f61407d95f9c9d6b3aeae4b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbee11bd7cb65e031415990869b52d2b8f81bf8b56e6cc935d38e09695db7b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbee11bd7cb65e031415990869b52d2b8f81bf8b56e6cc935d38e09695db7b48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86f693ade72bd593bc16021fe4b7564f1914ec2f1cda7039474ac2fb13de10f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86f693ade72bd593bc16021fe4b7564f1914ec2f1cda7039474ac2fb13de10f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7d7674252f67bde0ded2ec8ad1d925a92a26d3e3f741dee0061b735cb0b7a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7d7674252f67bde0ded2ec8ad1d925a92a26d3e3f741dee0061b735cb0b7a02\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kh8g9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:45Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:45 crc kubenswrapper[4824]: I1211 02:01:45.293590 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6bkc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8022f696-d0e7-437b-848a-3cd25bd1f364\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3e28f6778ec0dba6ea3e6b69e5a2c1c73c0d5db1abeead241fb58efdbfaefe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hh5kn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6bkc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:45Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:45 crc kubenswrapper[4824]: I1211 02:01:45.324420 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b193bd2-b536-4056-92f1-94c9836ab2eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51f3c41d86ddc44f7eca118029876ad353317f753d0ad173ed4d30c2a15f10e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c345850891744a0240781277e9a89a59f6899c4c944995f62e44650f807440a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed2ae40b1fe28111efc76d9d72345edecb5f3ea34522b238f3730812b29f2bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce7aa078800e6f299cfe03770fa7e3b64fa3742fe0c0f383c87d84cf48569ddf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6aa9fc1f1bff192c69bd59d32a05d4a72d09641d9ce4b265bc4517147a7e15dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b4a3bc50eb80777188206adab774e8acad8b2e4952a8415f91b1c29ba9a0ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09bb8b4c85e1fd5fa8d62db23ebee77a25a0faa274da3bb22ab679987597f26b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09bb8b4c85e1fd5fa8d62db23ebee77a25a0faa274da3bb22ab679987597f26b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T02:01:43Z\\\",\\\"message\\\":\\\"ng]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-webhook\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.254\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1211 02:01:43.627342 6417 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-target-xd92c in node crc\\\\nI1211 02:01:43.627315 6417 services_controller.go:443] Built service openshift-machine-config-operator/machine-config-daemon LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.43\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:9001, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}, services.lbConfig{vips:[]string{\\\\\\\"10.217.4.43\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8798, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLo\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:42Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-f6pgc_openshift-ovn-kubernetes(5b193bd2-b536-4056-92f1-94c9836ab2eb)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ead58694813806d307bbbc6a7b3cf41730d1c059b231a3002684bf09c10c7f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f6pgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:45Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:45 crc kubenswrapper[4824]: I1211 02:01:45.335691 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:45 crc kubenswrapper[4824]: I1211 02:01:45.335835 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:45 crc kubenswrapper[4824]: I1211 02:01:45.335863 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:45 crc kubenswrapper[4824]: I1211 02:01:45.335893 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:45 crc kubenswrapper[4824]: I1211 02:01:45.335916 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:45Z","lastTransitionTime":"2025-12-11T02:01:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:45 crc kubenswrapper[4824]: I1211 02:01:45.348569 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d14986666febd651039afb52ddf4db8969aca55e88eda9bbb20a67e35ba722ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:45Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:45 crc kubenswrapper[4824]: I1211 02:01:45.369446 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:45Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:45 crc kubenswrapper[4824]: I1211 02:01:45.386892 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44b156e4-64a4-4d45-aa5e-9b10a862faed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7368d92144b188e95a7796fabf497b0f8d9c0b240293bdc7b4aec1ff75497bc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w62dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcf289ac27dbda4f24aabbde3fb5119329e7b4fc7d6e018168da9a69661a4c57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w62dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gx6xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:45Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:45 crc kubenswrapper[4824]: I1211 02:01:45.405938 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"253d4291-2b2f-487c-8862-3d305f0bdc55\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44b29bb7a6eda7cb890d36f86375b5f3f197ae5cf390f16e71139b0904873e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://922af7f67b99a0ce08776f3f75c0c377e6ea69af1720b6178a6d1e2f9b4b8ba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://136c68020325db8e6930d0f40b0d0d7e013c14c4dd5e7d1449d01d6d458104d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c89afaf56b4a9597f6543ec2e3b176c7a6dab7dc341941fc3bc8ec85cc4edaf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:00:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:45Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:45 crc kubenswrapper[4824]: I1211 02:01:45.426346 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:45Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:45 crc kubenswrapper[4824]: I1211 02:01:45.438911 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:45 crc kubenswrapper[4824]: I1211 02:01:45.438974 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:45 crc kubenswrapper[4824]: I1211 02:01:45.438997 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:45 crc kubenswrapper[4824]: I1211 02:01:45.439027 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:45 crc kubenswrapper[4824]: I1211 02:01:45.439049 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:45Z","lastTransitionTime":"2025-12-11T02:01:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:45 crc kubenswrapper[4824]: I1211 02:01:45.541582 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:45 crc kubenswrapper[4824]: I1211 02:01:45.541663 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:45 crc kubenswrapper[4824]: I1211 02:01:45.541686 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:45 crc kubenswrapper[4824]: I1211 02:01:45.541716 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:45 crc kubenswrapper[4824]: I1211 02:01:45.541740 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:45Z","lastTransitionTime":"2025-12-11T02:01:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:45 crc kubenswrapper[4824]: I1211 02:01:45.631721 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-grnpt" Dec 11 02:01:45 crc kubenswrapper[4824]: I1211 02:01:45.631754 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 02:01:45 crc kubenswrapper[4824]: I1211 02:01:45.631735 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 02:01:45 crc kubenswrapper[4824]: E1211 02:01:45.631852 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-grnpt" podUID="09a52948-0404-45ff-85b1-51479268bf71" Dec 11 02:01:45 crc kubenswrapper[4824]: E1211 02:01:45.631935 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 02:01:45 crc kubenswrapper[4824]: I1211 02:01:45.631997 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 02:01:45 crc kubenswrapper[4824]: E1211 02:01:45.632184 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 02:01:45 crc kubenswrapper[4824]: E1211 02:01:45.632259 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 02:01:45 crc kubenswrapper[4824]: I1211 02:01:45.644791 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:45 crc kubenswrapper[4824]: I1211 02:01:45.644842 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:45 crc kubenswrapper[4824]: I1211 02:01:45.644859 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:45 crc kubenswrapper[4824]: I1211 02:01:45.644885 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:45 crc kubenswrapper[4824]: I1211 02:01:45.644903 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:45Z","lastTransitionTime":"2025-12-11T02:01:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:45 crc kubenswrapper[4824]: I1211 02:01:45.748836 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:45 crc kubenswrapper[4824]: I1211 02:01:45.748914 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:45 crc kubenswrapper[4824]: I1211 02:01:45.748936 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:45 crc kubenswrapper[4824]: I1211 02:01:45.748967 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:45 crc kubenswrapper[4824]: I1211 02:01:45.748992 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:45Z","lastTransitionTime":"2025-12-11T02:01:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:45 crc kubenswrapper[4824]: I1211 02:01:45.852080 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:45 crc kubenswrapper[4824]: I1211 02:01:45.852165 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:45 crc kubenswrapper[4824]: I1211 02:01:45.852182 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:45 crc kubenswrapper[4824]: I1211 02:01:45.852205 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:45 crc kubenswrapper[4824]: I1211 02:01:45.852224 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:45Z","lastTransitionTime":"2025-12-11T02:01:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:45 crc kubenswrapper[4824]: I1211 02:01:45.954606 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:45 crc kubenswrapper[4824]: I1211 02:01:45.954683 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:45 crc kubenswrapper[4824]: I1211 02:01:45.954701 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:45 crc kubenswrapper[4824]: I1211 02:01:45.954730 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:45 crc kubenswrapper[4824]: I1211 02:01:45.954750 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:45Z","lastTransitionTime":"2025-12-11T02:01:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:46 crc kubenswrapper[4824]: I1211 02:01:46.056891 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:46 crc kubenswrapper[4824]: I1211 02:01:46.056969 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:46 crc kubenswrapper[4824]: I1211 02:01:46.056981 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:46 crc kubenswrapper[4824]: I1211 02:01:46.056998 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:46 crc kubenswrapper[4824]: I1211 02:01:46.057009 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:46Z","lastTransitionTime":"2025-12-11T02:01:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:46 crc kubenswrapper[4824]: I1211 02:01:46.160241 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:46 crc kubenswrapper[4824]: I1211 02:01:46.160316 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:46 crc kubenswrapper[4824]: I1211 02:01:46.160339 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:46 crc kubenswrapper[4824]: I1211 02:01:46.160368 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:46 crc kubenswrapper[4824]: I1211 02:01:46.160390 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:46Z","lastTransitionTime":"2025-12-11T02:01:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:46 crc kubenswrapper[4824]: I1211 02:01:46.263279 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:46 crc kubenswrapper[4824]: I1211 02:01:46.263372 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:46 crc kubenswrapper[4824]: I1211 02:01:46.263390 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:46 crc kubenswrapper[4824]: I1211 02:01:46.263417 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:46 crc kubenswrapper[4824]: I1211 02:01:46.263433 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:46Z","lastTransitionTime":"2025-12-11T02:01:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:46 crc kubenswrapper[4824]: I1211 02:01:46.366681 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:46 crc kubenswrapper[4824]: I1211 02:01:46.366749 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:46 crc kubenswrapper[4824]: I1211 02:01:46.366765 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:46 crc kubenswrapper[4824]: I1211 02:01:46.366790 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:46 crc kubenswrapper[4824]: I1211 02:01:46.366813 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:46Z","lastTransitionTime":"2025-12-11T02:01:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:46 crc kubenswrapper[4824]: I1211 02:01:46.469984 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:46 crc kubenswrapper[4824]: I1211 02:01:46.470057 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:46 crc kubenswrapper[4824]: I1211 02:01:46.470074 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:46 crc kubenswrapper[4824]: I1211 02:01:46.470097 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:46 crc kubenswrapper[4824]: I1211 02:01:46.470146 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:46Z","lastTransitionTime":"2025-12-11T02:01:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:46 crc kubenswrapper[4824]: I1211 02:01:46.572419 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:46 crc kubenswrapper[4824]: I1211 02:01:46.572481 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:46 crc kubenswrapper[4824]: I1211 02:01:46.572496 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:46 crc kubenswrapper[4824]: I1211 02:01:46.572515 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:46 crc kubenswrapper[4824]: I1211 02:01:46.572528 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:46Z","lastTransitionTime":"2025-12-11T02:01:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:46 crc kubenswrapper[4824]: I1211 02:01:46.676168 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:46 crc kubenswrapper[4824]: I1211 02:01:46.676254 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:46 crc kubenswrapper[4824]: I1211 02:01:46.676272 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:46 crc kubenswrapper[4824]: I1211 02:01:46.676306 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:46 crc kubenswrapper[4824]: I1211 02:01:46.676328 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:46Z","lastTransitionTime":"2025-12-11T02:01:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:46 crc kubenswrapper[4824]: I1211 02:01:46.779053 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:46 crc kubenswrapper[4824]: I1211 02:01:46.779252 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:46 crc kubenswrapper[4824]: I1211 02:01:46.779281 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:46 crc kubenswrapper[4824]: I1211 02:01:46.779315 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:46 crc kubenswrapper[4824]: I1211 02:01:46.779343 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:46Z","lastTransitionTime":"2025-12-11T02:01:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:46 crc kubenswrapper[4824]: I1211 02:01:46.882510 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:46 crc kubenswrapper[4824]: I1211 02:01:46.882647 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:46 crc kubenswrapper[4824]: I1211 02:01:46.882669 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:46 crc kubenswrapper[4824]: I1211 02:01:46.882692 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:46 crc kubenswrapper[4824]: I1211 02:01:46.882713 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:46Z","lastTransitionTime":"2025-12-11T02:01:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:46 crc kubenswrapper[4824]: I1211 02:01:46.985327 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:46 crc kubenswrapper[4824]: I1211 02:01:46.985401 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:46 crc kubenswrapper[4824]: I1211 02:01:46.985419 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:46 crc kubenswrapper[4824]: I1211 02:01:46.985444 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:46 crc kubenswrapper[4824]: I1211 02:01:46.985464 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:46Z","lastTransitionTime":"2025-12-11T02:01:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:47 crc kubenswrapper[4824]: I1211 02:01:47.087967 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:47 crc kubenswrapper[4824]: I1211 02:01:47.088036 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:47 crc kubenswrapper[4824]: I1211 02:01:47.088054 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:47 crc kubenswrapper[4824]: I1211 02:01:47.088082 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:47 crc kubenswrapper[4824]: I1211 02:01:47.088099 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:47Z","lastTransitionTime":"2025-12-11T02:01:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:47 crc kubenswrapper[4824]: I1211 02:01:47.190904 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:47 crc kubenswrapper[4824]: I1211 02:01:47.190976 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:47 crc kubenswrapper[4824]: I1211 02:01:47.191001 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:47 crc kubenswrapper[4824]: I1211 02:01:47.191030 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:47 crc kubenswrapper[4824]: I1211 02:01:47.191052 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:47Z","lastTransitionTime":"2025-12-11T02:01:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:47 crc kubenswrapper[4824]: I1211 02:01:47.294076 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:47 crc kubenswrapper[4824]: I1211 02:01:47.294211 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:47 crc kubenswrapper[4824]: I1211 02:01:47.294228 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:47 crc kubenswrapper[4824]: I1211 02:01:47.294253 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:47 crc kubenswrapper[4824]: I1211 02:01:47.294272 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:47Z","lastTransitionTime":"2025-12-11T02:01:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:47 crc kubenswrapper[4824]: I1211 02:01:47.397621 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:47 crc kubenswrapper[4824]: I1211 02:01:47.397710 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:47 crc kubenswrapper[4824]: I1211 02:01:47.397730 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:47 crc kubenswrapper[4824]: I1211 02:01:47.397774 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:47 crc kubenswrapper[4824]: I1211 02:01:47.397799 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:47Z","lastTransitionTime":"2025-12-11T02:01:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:47 crc kubenswrapper[4824]: I1211 02:01:47.501039 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:47 crc kubenswrapper[4824]: I1211 02:01:47.501144 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:47 crc kubenswrapper[4824]: I1211 02:01:47.501164 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:47 crc kubenswrapper[4824]: I1211 02:01:47.501191 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:47 crc kubenswrapper[4824]: I1211 02:01:47.501215 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:47Z","lastTransitionTime":"2025-12-11T02:01:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:47 crc kubenswrapper[4824]: I1211 02:01:47.588455 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/09a52948-0404-45ff-85b1-51479268bf71-metrics-certs\") pod \"network-metrics-daemon-grnpt\" (UID: \"09a52948-0404-45ff-85b1-51479268bf71\") " pod="openshift-multus/network-metrics-daemon-grnpt" Dec 11 02:01:47 crc kubenswrapper[4824]: E1211 02:01:47.588738 4824 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 11 02:01:47 crc kubenswrapper[4824]: E1211 02:01:47.588855 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/09a52948-0404-45ff-85b1-51479268bf71-metrics-certs podName:09a52948-0404-45ff-85b1-51479268bf71 nodeName:}" failed. No retries permitted until 2025-12-11 02:02:03.588826836 +0000 UTC m=+65.277864245 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/09a52948-0404-45ff-85b1-51479268bf71-metrics-certs") pod "network-metrics-daemon-grnpt" (UID: "09a52948-0404-45ff-85b1-51479268bf71") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 11 02:01:47 crc kubenswrapper[4824]: I1211 02:01:47.604394 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:47 crc kubenswrapper[4824]: I1211 02:01:47.604451 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:47 crc kubenswrapper[4824]: I1211 02:01:47.604475 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:47 crc kubenswrapper[4824]: I1211 02:01:47.604531 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:47 crc kubenswrapper[4824]: I1211 02:01:47.604552 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:47Z","lastTransitionTime":"2025-12-11T02:01:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:47 crc kubenswrapper[4824]: I1211 02:01:47.631887 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 02:01:47 crc kubenswrapper[4824]: I1211 02:01:47.631935 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 02:01:47 crc kubenswrapper[4824]: I1211 02:01:47.632016 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-grnpt" Dec 11 02:01:47 crc kubenswrapper[4824]: E1211 02:01:47.632019 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 02:01:47 crc kubenswrapper[4824]: E1211 02:01:47.632150 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-grnpt" podUID="09a52948-0404-45ff-85b1-51479268bf71" Dec 11 02:01:47 crc kubenswrapper[4824]: E1211 02:01:47.632226 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 02:01:47 crc kubenswrapper[4824]: I1211 02:01:47.632310 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 02:01:47 crc kubenswrapper[4824]: E1211 02:01:47.632370 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 02:01:47 crc kubenswrapper[4824]: I1211 02:01:47.707643 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:47 crc kubenswrapper[4824]: I1211 02:01:47.707680 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:47 crc kubenswrapper[4824]: I1211 02:01:47.707690 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:47 crc kubenswrapper[4824]: I1211 02:01:47.707705 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:47 crc kubenswrapper[4824]: I1211 02:01:47.707716 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:47Z","lastTransitionTime":"2025-12-11T02:01:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:47 crc kubenswrapper[4824]: I1211 02:01:47.810572 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:47 crc kubenswrapper[4824]: I1211 02:01:47.810612 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:47 crc kubenswrapper[4824]: I1211 02:01:47.810623 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:47 crc kubenswrapper[4824]: I1211 02:01:47.810640 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:47 crc kubenswrapper[4824]: I1211 02:01:47.810651 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:47Z","lastTransitionTime":"2025-12-11T02:01:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:47 crc kubenswrapper[4824]: I1211 02:01:47.913828 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:47 crc kubenswrapper[4824]: I1211 02:01:47.913894 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:47 crc kubenswrapper[4824]: I1211 02:01:47.913921 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:47 crc kubenswrapper[4824]: I1211 02:01:47.913952 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:47 crc kubenswrapper[4824]: I1211 02:01:47.913974 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:47Z","lastTransitionTime":"2025-12-11T02:01:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:48 crc kubenswrapper[4824]: I1211 02:01:48.016463 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:48 crc kubenswrapper[4824]: I1211 02:01:48.016517 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:48 crc kubenswrapper[4824]: I1211 02:01:48.016534 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:48 crc kubenswrapper[4824]: I1211 02:01:48.016557 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:48 crc kubenswrapper[4824]: I1211 02:01:48.016575 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:48Z","lastTransitionTime":"2025-12-11T02:01:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:48 crc kubenswrapper[4824]: I1211 02:01:48.118958 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:48 crc kubenswrapper[4824]: I1211 02:01:48.118998 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:48 crc kubenswrapper[4824]: I1211 02:01:48.119008 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:48 crc kubenswrapper[4824]: I1211 02:01:48.119024 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:48 crc kubenswrapper[4824]: I1211 02:01:48.119036 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:48Z","lastTransitionTime":"2025-12-11T02:01:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:48 crc kubenswrapper[4824]: I1211 02:01:48.221645 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:48 crc kubenswrapper[4824]: I1211 02:01:48.221686 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:48 crc kubenswrapper[4824]: I1211 02:01:48.221698 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:48 crc kubenswrapper[4824]: I1211 02:01:48.221715 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:48 crc kubenswrapper[4824]: I1211 02:01:48.221726 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:48Z","lastTransitionTime":"2025-12-11T02:01:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:48 crc kubenswrapper[4824]: I1211 02:01:48.324915 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:48 crc kubenswrapper[4824]: I1211 02:01:48.324979 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:48 crc kubenswrapper[4824]: I1211 02:01:48.325001 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:48 crc kubenswrapper[4824]: I1211 02:01:48.325029 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:48 crc kubenswrapper[4824]: I1211 02:01:48.325050 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:48Z","lastTransitionTime":"2025-12-11T02:01:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:48 crc kubenswrapper[4824]: I1211 02:01:48.427986 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:48 crc kubenswrapper[4824]: I1211 02:01:48.428039 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:48 crc kubenswrapper[4824]: I1211 02:01:48.428057 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:48 crc kubenswrapper[4824]: I1211 02:01:48.428080 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:48 crc kubenswrapper[4824]: I1211 02:01:48.428097 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:48Z","lastTransitionTime":"2025-12-11T02:01:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:48 crc kubenswrapper[4824]: I1211 02:01:48.531711 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:48 crc kubenswrapper[4824]: I1211 02:01:48.531761 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:48 crc kubenswrapper[4824]: I1211 02:01:48.531778 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:48 crc kubenswrapper[4824]: I1211 02:01:48.531799 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:48 crc kubenswrapper[4824]: I1211 02:01:48.531816 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:48Z","lastTransitionTime":"2025-12-11T02:01:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:48 crc kubenswrapper[4824]: I1211 02:01:48.636634 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:48 crc kubenswrapper[4824]: I1211 02:01:48.636691 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:48 crc kubenswrapper[4824]: I1211 02:01:48.636707 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:48 crc kubenswrapper[4824]: I1211 02:01:48.636737 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:48 crc kubenswrapper[4824]: I1211 02:01:48.636758 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:48Z","lastTransitionTime":"2025-12-11T02:01:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:48 crc kubenswrapper[4824]: I1211 02:01:48.652793 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:48Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:48 crc kubenswrapper[4824]: I1211 02:01:48.673016 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kh8g9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"50c69d41-a65b-463e-a829-ea349bd2fe46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dca6be41bd7a3ce12ecd4567c868f696b65e8eb1bad436feda6e4f9bfb8e525\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69af1f33fd0bafc929b94e702b4f46c6c7d7a362d5552b121729c477892d91d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69af1f33fd0bafc929b94e702b4f46c6c7d7a362d5552b121729c477892d91d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd9675d0afded587bc660a3862fd0c5dd6885961afacda957aa9ee8501d68a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd9675d0afded587bc660a3862fd0c5dd6885961afacda957aa9ee8501d68a69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aac94149458bd6321be90b26b48ee26428d273e7f61407d95f9c9d6b3aeae4b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aac94149458bd6321be90b26b48ee26428d273e7f61407d95f9c9d6b3aeae4b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbee11bd7cb65e031415990869b52d2b8f81bf8b56e6cc935d38e09695db7b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbee11bd7cb65e031415990869b52d2b8f81bf8b56e6cc935d38e09695db7b48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86f693ade72bd593bc16021fe4b7564f1914ec2f1cda7039474ac2fb13de10f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86f693ade72bd593bc16021fe4b7564f1914ec2f1cda7039474ac2fb13de10f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7d7674252f67bde0ded2ec8ad1d925a92a26d3e3f741dee0061b735cb0b7a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7d7674252f67bde0ded2ec8ad1d925a92a26d3e3f741dee0061b735cb0b7a02\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kh8g9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:48Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:48 crc kubenswrapper[4824]: I1211 02:01:48.685815 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6bkc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8022f696-d0e7-437b-848a-3cd25bd1f364\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3e28f6778ec0dba6ea3e6b69e5a2c1c73c0d5db1abeead241fb58efdbfaefe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hh5kn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6bkc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:48Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:48 crc kubenswrapper[4824]: I1211 02:01:48.715766 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b193bd2-b536-4056-92f1-94c9836ab2eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51f3c41d86ddc44f7eca118029876ad353317f753d0ad173ed4d30c2a15f10e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c345850891744a0240781277e9a89a59f6899c4c944995f62e44650f807440a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed2ae40b1fe28111efc76d9d72345edecb5f3ea34522b238f3730812b29f2bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce7aa078800e6f299cfe03770fa7e3b64fa3742fe0c0f383c87d84cf48569ddf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6aa9fc1f1bff192c69bd59d32a05d4a72d09641d9ce4b265bc4517147a7e15dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b4a3bc50eb80777188206adab774e8acad8b2e4952a8415f91b1c29ba9a0ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09bb8b4c85e1fd5fa8d62db23ebee77a25a0faa274da3bb22ab679987597f26b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09bb8b4c85e1fd5fa8d62db23ebee77a25a0faa274da3bb22ab679987597f26b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T02:01:43Z\\\",\\\"message\\\":\\\"ng]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-webhook\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.254\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1211 02:01:43.627342 6417 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-target-xd92c in node crc\\\\nI1211 02:01:43.627315 6417 services_controller.go:443] Built service openshift-machine-config-operator/machine-config-daemon LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.43\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:9001, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}, services.lbConfig{vips:[]string{\\\\\\\"10.217.4.43\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8798, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLo\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:42Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-f6pgc_openshift-ovn-kubernetes(5b193bd2-b536-4056-92f1-94c9836ab2eb)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ead58694813806d307bbbc6a7b3cf41730d1c059b231a3002684bf09c10c7f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f6pgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:48Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:48 crc kubenswrapper[4824]: I1211 02:01:48.731182 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-br4j9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce125194-9c9d-4ea4-8a72-2c30d5722295\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9798f904052e08196a0db1745060dab11e2710308bca01ff160475910f2665df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fc5s8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-br4j9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:48Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:48 crc kubenswrapper[4824]: I1211 02:01:48.739416 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:48 crc kubenswrapper[4824]: I1211 02:01:48.739460 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:48 crc kubenswrapper[4824]: I1211 02:01:48.739492 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:48 crc kubenswrapper[4824]: I1211 02:01:48.739517 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:48 crc kubenswrapper[4824]: I1211 02:01:48.739534 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:48Z","lastTransitionTime":"2025-12-11T02:01:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:48 crc kubenswrapper[4824]: I1211 02:01:48.756234 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8v4jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99a6bfee-afca-4726-a964-b95a948945a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b23b9d8f18db3eeadb60f83b36b05b1778cd0fa31669dad2680bb335ae004acf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4hk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3acc1940a0534ba0e6096175e41834fdd0fbcbb4c3fb6e9a38e76aedae55d2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4hk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-8v4jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:48Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:48 crc kubenswrapper[4824]: I1211 02:01:48.773271 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-grnpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09a52948-0404-45ff-85b1-51479268bf71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmmzm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmmzm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-grnpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:48Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:48 crc kubenswrapper[4824]: I1211 02:01:48.800356 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d14986666febd651039afb52ddf4db8969aca55e88eda9bbb20a67e35ba722ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:48Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:48 crc kubenswrapper[4824]: I1211 02:01:48.818758 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:48Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:48 crc kubenswrapper[4824]: I1211 02:01:48.835816 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44b156e4-64a4-4d45-aa5e-9b10a862faed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7368d92144b188e95a7796fabf497b0f8d9c0b240293bdc7b4aec1ff75497bc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w62dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcf289ac27dbda4f24aabbde3fb5119329e7b4fc7d6e018168da9a69661a4c57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w62dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gx6xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:48Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:48 crc kubenswrapper[4824]: I1211 02:01:48.842341 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:48 crc kubenswrapper[4824]: I1211 02:01:48.842403 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:48 crc kubenswrapper[4824]: I1211 02:01:48.842423 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:48 crc kubenswrapper[4824]: I1211 02:01:48.842448 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:48 crc kubenswrapper[4824]: I1211 02:01:48.842467 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:48Z","lastTransitionTime":"2025-12-11T02:01:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:48 crc kubenswrapper[4824]: I1211 02:01:48.853348 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"253d4291-2b2f-487c-8862-3d305f0bdc55\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44b29bb7a6eda7cb890d36f86375b5f3f197ae5cf390f16e71139b0904873e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://922af7f67b99a0ce08776f3f75c0c377e6ea69af1720b6178a6d1e2f9b4b8ba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://136c68020325db8e6930d0f40b0d0d7e013c14c4dd5e7d1449d01d6d458104d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c89afaf56b4a9597f6543ec2e3b176c7a6dab7dc341941fc3bc8ec85cc4edaf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:00:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:48Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:48 crc kubenswrapper[4824]: I1211 02:01:48.881882 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:48Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:48 crc kubenswrapper[4824]: I1211 02:01:48.913266 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"beab386b-6c65-4bc3-9ef0-dc3f68b4403f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c26f06d2ed5e6e98b61a51f073130675f2c49acda106b8aedcbb2a6d4f7580c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20af9b10a67c8b21a7abcd23e42bc412798d159c8422e7c8ff58f5b997ca19a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9bbfc6edfe050b3fa9dbeee5f98d416bcb2a1533a38608a881b75396fd0b059\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7c766f8de73ca762ffde0c18b2d1c46bb152b2b37535a611c5a0e454a21b7d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d45819f0611c91b20c83eeee1092031a3441fb69f3015241632ad419880227d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 02:01:11.070250 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 02:01:11.071673 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2752882565/tls.crt::/tmp/serving-cert-2752882565/tls.key\\\\\\\"\\\\nI1211 02:01:17.042610 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 02:01:17.048046 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 02:01:17.048080 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 02:01:17.048172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 02:01:17.048188 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 02:01:17.058248 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 02:01:17.058294 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 02:01:17.058303 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 02:01:17.058315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 02:01:17.058322 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 02:01:17.058328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 02:01:17.058335 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 02:01:17.058808 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 02:01:17.060225 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14a940a72b2404a6cbd52a69e150b03266bcd7f5c1d3f4d07b51c0c69e51786\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:00:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:48Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:48 crc kubenswrapper[4824]: I1211 02:01:48.939181 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ded146c045085be2b58821749052413a36850ede00ae1024fe2dae9fa5ac84b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:48Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:48 crc kubenswrapper[4824]: I1211 02:01:48.944142 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:48 crc kubenswrapper[4824]: I1211 02:01:48.944176 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:48 crc kubenswrapper[4824]: I1211 02:01:48.944184 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:48 crc kubenswrapper[4824]: I1211 02:01:48.944198 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:48 crc kubenswrapper[4824]: I1211 02:01:48.944208 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:48Z","lastTransitionTime":"2025-12-11T02:01:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:48 crc kubenswrapper[4824]: I1211 02:01:48.952943 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76b10bbd1e891b8c682f4b2e17d023fd787a8ed44ed94fbf7d97314d641ccf36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f32838f33832b9b67e7dd9c53340680e17721cf1c7bdf0170230c5f8feafe64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:48Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:48 crc kubenswrapper[4824]: I1211 02:01:48.962437 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kptws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68568dba-83a5-46c1-8d01-337359679d91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f965a6d6db619995132f61b5d51b66f0b2b3fdbef3fbec63e231b9a62ccd0a9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd7nf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kptws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:48Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:49 crc kubenswrapper[4824]: I1211 02:01:49.046788 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:49 crc kubenswrapper[4824]: I1211 02:01:49.046845 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:49 crc kubenswrapper[4824]: I1211 02:01:49.046862 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:49 crc kubenswrapper[4824]: I1211 02:01:49.046885 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:49 crc kubenswrapper[4824]: I1211 02:01:49.046902 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:49Z","lastTransitionTime":"2025-12-11T02:01:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:49 crc kubenswrapper[4824]: I1211 02:01:49.149409 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:49 crc kubenswrapper[4824]: I1211 02:01:49.149473 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:49 crc kubenswrapper[4824]: I1211 02:01:49.149489 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:49 crc kubenswrapper[4824]: I1211 02:01:49.149514 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:49 crc kubenswrapper[4824]: I1211 02:01:49.149534 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:49Z","lastTransitionTime":"2025-12-11T02:01:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:49 crc kubenswrapper[4824]: I1211 02:01:49.252669 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:49 crc kubenswrapper[4824]: I1211 02:01:49.252722 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:49 crc kubenswrapper[4824]: I1211 02:01:49.252740 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:49 crc kubenswrapper[4824]: I1211 02:01:49.252764 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:49 crc kubenswrapper[4824]: I1211 02:01:49.252781 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:49Z","lastTransitionTime":"2025-12-11T02:01:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:49 crc kubenswrapper[4824]: I1211 02:01:49.355421 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:49 crc kubenswrapper[4824]: I1211 02:01:49.355474 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:49 crc kubenswrapper[4824]: I1211 02:01:49.355491 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:49 crc kubenswrapper[4824]: I1211 02:01:49.355516 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:49 crc kubenswrapper[4824]: I1211 02:01:49.355534 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:49Z","lastTransitionTime":"2025-12-11T02:01:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:49 crc kubenswrapper[4824]: I1211 02:01:49.458663 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:49 crc kubenswrapper[4824]: I1211 02:01:49.458720 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:49 crc kubenswrapper[4824]: I1211 02:01:49.458737 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:49 crc kubenswrapper[4824]: I1211 02:01:49.458764 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:49 crc kubenswrapper[4824]: I1211 02:01:49.458784 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:49Z","lastTransitionTime":"2025-12-11T02:01:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:49 crc kubenswrapper[4824]: I1211 02:01:49.509221 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 02:01:49 crc kubenswrapper[4824]: E1211 02:01:49.509386 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 02:02:21.509358035 +0000 UTC m=+83.198395444 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:01:49 crc kubenswrapper[4824]: I1211 02:01:49.561826 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:49 crc kubenswrapper[4824]: I1211 02:01:49.561866 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:49 crc kubenswrapper[4824]: I1211 02:01:49.561882 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:49 crc kubenswrapper[4824]: I1211 02:01:49.561904 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:49 crc kubenswrapper[4824]: I1211 02:01:49.561921 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:49Z","lastTransitionTime":"2025-12-11T02:01:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:49 crc kubenswrapper[4824]: I1211 02:01:49.610491 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 02:01:49 crc kubenswrapper[4824]: I1211 02:01:49.610568 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 02:01:49 crc kubenswrapper[4824]: I1211 02:01:49.610608 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 02:01:49 crc kubenswrapper[4824]: I1211 02:01:49.610644 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 02:01:49 crc kubenswrapper[4824]: E1211 02:01:49.610683 4824 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 11 02:01:49 crc kubenswrapper[4824]: E1211 02:01:49.610769 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-11 02:02:21.610743512 +0000 UTC m=+83.299780931 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 11 02:01:49 crc kubenswrapper[4824]: E1211 02:01:49.610769 4824 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 11 02:01:49 crc kubenswrapper[4824]: E1211 02:01:49.610790 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 11 02:01:49 crc kubenswrapper[4824]: E1211 02:01:49.610804 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 11 02:01:49 crc kubenswrapper[4824]: E1211 02:01:49.610825 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 11 02:01:49 crc kubenswrapper[4824]: E1211 02:01:49.610850 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 11 02:01:49 crc kubenswrapper[4824]: E1211 02:01:49.610867 4824 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 02:01:49 crc kubenswrapper[4824]: E1211 02:01:49.610877 4824 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 02:01:49 crc kubenswrapper[4824]: E1211 02:01:49.610934 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-11 02:02:21.610904417 +0000 UTC m=+83.299941826 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 11 02:01:49 crc kubenswrapper[4824]: E1211 02:01:49.610963 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-11 02:02:21.610950368 +0000 UTC m=+83.299987777 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 02:01:49 crc kubenswrapper[4824]: E1211 02:01:49.610985 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-11 02:02:21.610974589 +0000 UTC m=+83.300012008 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 02:01:49 crc kubenswrapper[4824]: I1211 02:01:49.631583 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 02:01:49 crc kubenswrapper[4824]: I1211 02:01:49.631638 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 02:01:49 crc kubenswrapper[4824]: I1211 02:01:49.631645 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-grnpt" Dec 11 02:01:49 crc kubenswrapper[4824]: I1211 02:01:49.631611 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 02:01:49 crc kubenswrapper[4824]: E1211 02:01:49.631812 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 02:01:49 crc kubenswrapper[4824]: E1211 02:01:49.631940 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 02:01:49 crc kubenswrapper[4824]: E1211 02:01:49.632047 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-grnpt" podUID="09a52948-0404-45ff-85b1-51479268bf71" Dec 11 02:01:49 crc kubenswrapper[4824]: E1211 02:01:49.632179 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 02:01:49 crc kubenswrapper[4824]: I1211 02:01:49.664913 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:49 crc kubenswrapper[4824]: I1211 02:01:49.664954 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:49 crc kubenswrapper[4824]: I1211 02:01:49.664970 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:49 crc kubenswrapper[4824]: I1211 02:01:49.664991 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:49 crc kubenswrapper[4824]: I1211 02:01:49.665008 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:49Z","lastTransitionTime":"2025-12-11T02:01:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:49 crc kubenswrapper[4824]: I1211 02:01:49.767723 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:49 crc kubenswrapper[4824]: I1211 02:01:49.767775 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:49 crc kubenswrapper[4824]: I1211 02:01:49.767791 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:49 crc kubenswrapper[4824]: I1211 02:01:49.767852 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:49 crc kubenswrapper[4824]: I1211 02:01:49.767871 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:49Z","lastTransitionTime":"2025-12-11T02:01:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:49 crc kubenswrapper[4824]: I1211 02:01:49.870939 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:49 crc kubenswrapper[4824]: I1211 02:01:49.871005 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:49 crc kubenswrapper[4824]: I1211 02:01:49.871028 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:49 crc kubenswrapper[4824]: I1211 02:01:49.871055 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:49 crc kubenswrapper[4824]: I1211 02:01:49.871077 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:49Z","lastTransitionTime":"2025-12-11T02:01:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:49 crc kubenswrapper[4824]: I1211 02:01:49.974275 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:49 crc kubenswrapper[4824]: I1211 02:01:49.974327 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:49 crc kubenswrapper[4824]: I1211 02:01:49.974342 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:49 crc kubenswrapper[4824]: I1211 02:01:49.974366 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:49 crc kubenswrapper[4824]: I1211 02:01:49.974383 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:49Z","lastTransitionTime":"2025-12-11T02:01:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:50 crc kubenswrapper[4824]: I1211 02:01:50.077064 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:50 crc kubenswrapper[4824]: I1211 02:01:50.077168 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:50 crc kubenswrapper[4824]: I1211 02:01:50.077187 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:50 crc kubenswrapper[4824]: I1211 02:01:50.077212 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:50 crc kubenswrapper[4824]: I1211 02:01:50.077229 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:50Z","lastTransitionTime":"2025-12-11T02:01:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:50 crc kubenswrapper[4824]: I1211 02:01:50.180356 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:50 crc kubenswrapper[4824]: I1211 02:01:50.180406 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:50 crc kubenswrapper[4824]: I1211 02:01:50.180423 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:50 crc kubenswrapper[4824]: I1211 02:01:50.180447 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:50 crc kubenswrapper[4824]: I1211 02:01:50.180465 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:50Z","lastTransitionTime":"2025-12-11T02:01:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:50 crc kubenswrapper[4824]: I1211 02:01:50.283104 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:50 crc kubenswrapper[4824]: I1211 02:01:50.283205 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:50 crc kubenswrapper[4824]: I1211 02:01:50.283228 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:50 crc kubenswrapper[4824]: I1211 02:01:50.283251 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:50 crc kubenswrapper[4824]: I1211 02:01:50.283267 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:50Z","lastTransitionTime":"2025-12-11T02:01:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:50 crc kubenswrapper[4824]: I1211 02:01:50.386144 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:50 crc kubenswrapper[4824]: I1211 02:01:50.386186 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:50 crc kubenswrapper[4824]: I1211 02:01:50.386197 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:50 crc kubenswrapper[4824]: I1211 02:01:50.386213 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:50 crc kubenswrapper[4824]: I1211 02:01:50.386225 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:50Z","lastTransitionTime":"2025-12-11T02:01:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:50 crc kubenswrapper[4824]: I1211 02:01:50.489106 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:50 crc kubenswrapper[4824]: I1211 02:01:50.489199 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:50 crc kubenswrapper[4824]: I1211 02:01:50.489215 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:50 crc kubenswrapper[4824]: I1211 02:01:50.489239 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:50 crc kubenswrapper[4824]: I1211 02:01:50.489257 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:50Z","lastTransitionTime":"2025-12-11T02:01:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:50 crc kubenswrapper[4824]: I1211 02:01:50.591951 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:50 crc kubenswrapper[4824]: I1211 02:01:50.592001 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:50 crc kubenswrapper[4824]: I1211 02:01:50.592019 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:50 crc kubenswrapper[4824]: I1211 02:01:50.592038 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:50 crc kubenswrapper[4824]: I1211 02:01:50.592054 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:50Z","lastTransitionTime":"2025-12-11T02:01:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:50 crc kubenswrapper[4824]: I1211 02:01:50.695229 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:50 crc kubenswrapper[4824]: I1211 02:01:50.695790 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:50 crc kubenswrapper[4824]: I1211 02:01:50.696014 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:50 crc kubenswrapper[4824]: I1211 02:01:50.696291 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:50 crc kubenswrapper[4824]: I1211 02:01:50.696496 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:50Z","lastTransitionTime":"2025-12-11T02:01:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:50 crc kubenswrapper[4824]: I1211 02:01:50.799540 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:50 crc kubenswrapper[4824]: I1211 02:01:50.799610 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:50 crc kubenswrapper[4824]: I1211 02:01:50.799628 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:50 crc kubenswrapper[4824]: I1211 02:01:50.799651 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:50 crc kubenswrapper[4824]: I1211 02:01:50.799672 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:50Z","lastTransitionTime":"2025-12-11T02:01:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:50 crc kubenswrapper[4824]: I1211 02:01:50.903217 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:50 crc kubenswrapper[4824]: I1211 02:01:50.903327 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:50 crc kubenswrapper[4824]: I1211 02:01:50.903346 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:50 crc kubenswrapper[4824]: I1211 02:01:50.903370 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:50 crc kubenswrapper[4824]: I1211 02:01:50.903387 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:50Z","lastTransitionTime":"2025-12-11T02:01:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.006710 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.006760 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.006777 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.006801 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.006823 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:51Z","lastTransitionTime":"2025-12-11T02:01:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.109705 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.109756 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.109773 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.109798 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.109814 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:51Z","lastTransitionTime":"2025-12-11T02:01:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.212896 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.213329 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.213388 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.213417 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.213435 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:51Z","lastTransitionTime":"2025-12-11T02:01:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.247789 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.247833 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.247852 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.247876 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.247893 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:51Z","lastTransitionTime":"2025-12-11T02:01:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:51 crc kubenswrapper[4824]: E1211 02:01:51.268329 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:01:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:01:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:01:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:01:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d57bece6-6fc6-4a84-b8a8-e1678f9e50db\\\",\\\"systemUUID\\\":\\\"f76b32e4-2764-4219-bbf4-f7e56d76df3b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:51Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.274259 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.274301 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.274316 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.274338 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.274354 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:51Z","lastTransitionTime":"2025-12-11T02:01:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.277293 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.291968 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 11 02:01:51 crc kubenswrapper[4824]: E1211 02:01:51.300200 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:01:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:01:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:01:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:01:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d57bece6-6fc6-4a84-b8a8-e1678f9e50db\\\",\\\"systemUUID\\\":\\\"f76b32e4-2764-4219-bbf4-f7e56d76df3b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:51Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.302578 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:51Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.306201 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.306247 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.306264 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.306294 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.306313 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:51Z","lastTransitionTime":"2025-12-11T02:01:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.325461 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"253d4291-2b2f-487c-8862-3d305f0bdc55\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44b29bb7a6eda7cb890d36f86375b5f3f197ae5cf390f16e71139b0904873e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://922af7f67b99a0ce08776f3f75c0c377e6ea69af1720b6178a6d1e2f9b4b8ba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://136c68020325db8e6930d0f40b0d0d7e013c14c4dd5e7d1449d01d6d458104d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c89afaf56b4a9597f6543ec2e3b176c7a6dab7dc341941fc3bc8ec85cc4edaf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:00:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:51Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:51 crc kubenswrapper[4824]: E1211 02:01:51.330063 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:01:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:01:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:01:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:01:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d57bece6-6fc6-4a84-b8a8-e1678f9e50db\\\",\\\"systemUUID\\\":\\\"f76b32e4-2764-4219-bbf4-f7e56d76df3b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:51Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.335432 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.335488 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.335508 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.335529 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.335545 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:51Z","lastTransitionTime":"2025-12-11T02:01:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.347039 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ded146c045085be2b58821749052413a36850ede00ae1024fe2dae9fa5ac84b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:51Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:51 crc kubenswrapper[4824]: E1211 02:01:51.357034 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:01:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:01:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:01:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:01:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d57bece6-6fc6-4a84-b8a8-e1678f9e50db\\\",\\\"systemUUID\\\":\\\"f76b32e4-2764-4219-bbf4-f7e56d76df3b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:51Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.362584 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.362648 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.362672 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.362701 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.362723 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:51Z","lastTransitionTime":"2025-12-11T02:01:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.367398 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76b10bbd1e891b8c682f4b2e17d023fd787a8ed44ed94fbf7d97314d641ccf36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f32838f33832b9b67e7dd9c53340680e17721cf1c7bdf0170230c5f8feafe64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:51Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:51 crc kubenswrapper[4824]: E1211 02:01:51.383233 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:01:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:01:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:01:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:01:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d57bece6-6fc6-4a84-b8a8-e1678f9e50db\\\",\\\"systemUUID\\\":\\\"f76b32e4-2764-4219-bbf4-f7e56d76df3b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:51Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:51 crc kubenswrapper[4824]: E1211 02:01:51.383656 4824 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.385690 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.385731 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.385747 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.385768 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.385784 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:51Z","lastTransitionTime":"2025-12-11T02:01:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.387708 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kptws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68568dba-83a5-46c1-8d01-337359679d91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f965a6d6db619995132f61b5d51b66f0b2b3fdbef3fbec63e231b9a62ccd0a9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd7nf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kptws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:51Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.409082 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"beab386b-6c65-4bc3-9ef0-dc3f68b4403f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c26f06d2ed5e6e98b61a51f073130675f2c49acda106b8aedcbb2a6d4f7580c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20af9b10a67c8b21a7abcd23e42bc412798d159c8422e7c8ff58f5b997ca19a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9bbfc6edfe050b3fa9dbeee5f98d416bcb2a1533a38608a881b75396fd0b059\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7c766f8de73ca762ffde0c18b2d1c46bb152b2b37535a611c5a0e454a21b7d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d45819f0611c91b20c83eeee1092031a3441fb69f3015241632ad419880227d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 02:01:11.070250 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 02:01:11.071673 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2752882565/tls.crt::/tmp/serving-cert-2752882565/tls.key\\\\\\\"\\\\nI1211 02:01:17.042610 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 02:01:17.048046 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 02:01:17.048080 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 02:01:17.048172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 02:01:17.048188 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 02:01:17.058248 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 02:01:17.058294 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 02:01:17.058303 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 02:01:17.058315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 02:01:17.058322 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 02:01:17.058328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 02:01:17.058335 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 02:01:17.058808 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 02:01:17.060225 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14a940a72b2404a6cbd52a69e150b03266bcd7f5c1d3f4d07b51c0c69e51786\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:00:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:51Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.430030 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6bkc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8022f696-d0e7-437b-848a-3cd25bd1f364\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3e28f6778ec0dba6ea3e6b69e5a2c1c73c0d5db1abeead241fb58efdbfaefe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hh5kn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6bkc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:51Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.460904 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b193bd2-b536-4056-92f1-94c9836ab2eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51f3c41d86ddc44f7eca118029876ad353317f753d0ad173ed4d30c2a15f10e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c345850891744a0240781277e9a89a59f6899c4c944995f62e44650f807440a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed2ae40b1fe28111efc76d9d72345edecb5f3ea34522b238f3730812b29f2bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce7aa078800e6f299cfe03770fa7e3b64fa3742fe0c0f383c87d84cf48569ddf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6aa9fc1f1bff192c69bd59d32a05d4a72d09641d9ce4b265bc4517147a7e15dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b4a3bc50eb80777188206adab774e8acad8b2e4952a8415f91b1c29ba9a0ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09bb8b4c85e1fd5fa8d62db23ebee77a25a0faa274da3bb22ab679987597f26b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09bb8b4c85e1fd5fa8d62db23ebee77a25a0faa274da3bb22ab679987597f26b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T02:01:43Z\\\",\\\"message\\\":\\\"ng]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-webhook\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.254\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1211 02:01:43.627342 6417 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-target-xd92c in node crc\\\\nI1211 02:01:43.627315 6417 services_controller.go:443] Built service openshift-machine-config-operator/machine-config-daemon LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.43\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:9001, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}, services.lbConfig{vips:[]string{\\\\\\\"10.217.4.43\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8798, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLo\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:42Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-f6pgc_openshift-ovn-kubernetes(5b193bd2-b536-4056-92f1-94c9836ab2eb)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ead58694813806d307bbbc6a7b3cf41730d1c059b231a3002684bf09c10c7f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f6pgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:51Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.478766 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-br4j9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce125194-9c9d-4ea4-8a72-2c30d5722295\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9798f904052e08196a0db1745060dab11e2710308bca01ff160475910f2665df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fc5s8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-br4j9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:51Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.488905 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.488968 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.488990 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.489019 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.489040 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:51Z","lastTransitionTime":"2025-12-11T02:01:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.496437 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8v4jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99a6bfee-afca-4726-a964-b95a948945a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b23b9d8f18db3eeadb60f83b36b05b1778cd0fa31669dad2680bb335ae004acf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4hk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3acc1940a0534ba0e6096175e41834fdd0fbcbb4c3fb6e9a38e76aedae55d2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4hk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-8v4jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:51Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.511388 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-grnpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09a52948-0404-45ff-85b1-51479268bf71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmmzm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmmzm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-grnpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:51Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.530008 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:51Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.555319 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kh8g9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"50c69d41-a65b-463e-a829-ea349bd2fe46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dca6be41bd7a3ce12ecd4567c868f696b65e8eb1bad436feda6e4f9bfb8e525\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69af1f33fd0bafc929b94e702b4f46c6c7d7a362d5552b121729c477892d91d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69af1f33fd0bafc929b94e702b4f46c6c7d7a362d5552b121729c477892d91d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd9675d0afded587bc660a3862fd0c5dd6885961afacda957aa9ee8501d68a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd9675d0afded587bc660a3862fd0c5dd6885961afacda957aa9ee8501d68a69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aac94149458bd6321be90b26b48ee26428d273e7f61407d95f9c9d6b3aeae4b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aac94149458bd6321be90b26b48ee26428d273e7f61407d95f9c9d6b3aeae4b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbee11bd7cb65e031415990869b52d2b8f81bf8b56e6cc935d38e09695db7b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbee11bd7cb65e031415990869b52d2b8f81bf8b56e6cc935d38e09695db7b48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86f693ade72bd593bc16021fe4b7564f1914ec2f1cda7039474ac2fb13de10f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86f693ade72bd593bc16021fe4b7564f1914ec2f1cda7039474ac2fb13de10f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7d7674252f67bde0ded2ec8ad1d925a92a26d3e3f741dee0061b735cb0b7a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7d7674252f67bde0ded2ec8ad1d925a92a26d3e3f741dee0061b735cb0b7a02\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kh8g9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:51Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.573779 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44b156e4-64a4-4d45-aa5e-9b10a862faed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7368d92144b188e95a7796fabf497b0f8d9c0b240293bdc7b4aec1ff75497bc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w62dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcf289ac27dbda4f24aabbde3fb5119329e7b4fc7d6e018168da9a69661a4c57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w62dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gx6xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:51Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.592351 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d14986666febd651039afb52ddf4db8969aca55e88eda9bbb20a67e35ba722ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:51Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.592912 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.592984 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.593008 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.593038 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.593063 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:51Z","lastTransitionTime":"2025-12-11T02:01:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.612847 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:51Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.632254 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.632316 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 02:01:51 crc kubenswrapper[4824]: E1211 02:01:51.632479 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.632252 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 02:01:51 crc kubenswrapper[4824]: E1211 02:01:51.632696 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.632814 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-grnpt" Dec 11 02:01:51 crc kubenswrapper[4824]: E1211 02:01:51.632959 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 02:01:51 crc kubenswrapper[4824]: E1211 02:01:51.633061 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-grnpt" podUID="09a52948-0404-45ff-85b1-51479268bf71" Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.695976 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.696048 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.696071 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.696100 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.696158 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:51Z","lastTransitionTime":"2025-12-11T02:01:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.798993 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.799057 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.799075 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.799100 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.799147 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:51Z","lastTransitionTime":"2025-12-11T02:01:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.901860 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.902311 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.902508 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.902698 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:51 crc kubenswrapper[4824]: I1211 02:01:51.902873 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:51Z","lastTransitionTime":"2025-12-11T02:01:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:52 crc kubenswrapper[4824]: I1211 02:01:52.005388 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:52 crc kubenswrapper[4824]: I1211 02:01:52.005453 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:52 crc kubenswrapper[4824]: I1211 02:01:52.005477 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:52 crc kubenswrapper[4824]: I1211 02:01:52.005507 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:52 crc kubenswrapper[4824]: I1211 02:01:52.005530 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:52Z","lastTransitionTime":"2025-12-11T02:01:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:52 crc kubenswrapper[4824]: I1211 02:01:52.109155 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:52 crc kubenswrapper[4824]: I1211 02:01:52.109224 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:52 crc kubenswrapper[4824]: I1211 02:01:52.109241 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:52 crc kubenswrapper[4824]: I1211 02:01:52.109263 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:52 crc kubenswrapper[4824]: I1211 02:01:52.109282 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:52Z","lastTransitionTime":"2025-12-11T02:01:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:52 crc kubenswrapper[4824]: I1211 02:01:52.211910 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:52 crc kubenswrapper[4824]: I1211 02:01:52.212392 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:52 crc kubenswrapper[4824]: I1211 02:01:52.212587 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:52 crc kubenswrapper[4824]: I1211 02:01:52.212770 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:52 crc kubenswrapper[4824]: I1211 02:01:52.212951 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:52Z","lastTransitionTime":"2025-12-11T02:01:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:52 crc kubenswrapper[4824]: I1211 02:01:52.316264 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:52 crc kubenswrapper[4824]: I1211 02:01:52.316631 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:52 crc kubenswrapper[4824]: I1211 02:01:52.316808 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:52 crc kubenswrapper[4824]: I1211 02:01:52.316939 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:52 crc kubenswrapper[4824]: I1211 02:01:52.317092 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:52Z","lastTransitionTime":"2025-12-11T02:01:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:52 crc kubenswrapper[4824]: I1211 02:01:52.420432 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:52 crc kubenswrapper[4824]: I1211 02:01:52.420489 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:52 crc kubenswrapper[4824]: I1211 02:01:52.420507 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:52 crc kubenswrapper[4824]: I1211 02:01:52.420532 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:52 crc kubenswrapper[4824]: I1211 02:01:52.420551 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:52Z","lastTransitionTime":"2025-12-11T02:01:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:52 crc kubenswrapper[4824]: I1211 02:01:52.523373 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:52 crc kubenswrapper[4824]: I1211 02:01:52.523433 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:52 crc kubenswrapper[4824]: I1211 02:01:52.523452 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:52 crc kubenswrapper[4824]: I1211 02:01:52.523476 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:52 crc kubenswrapper[4824]: I1211 02:01:52.523497 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:52Z","lastTransitionTime":"2025-12-11T02:01:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:52 crc kubenswrapper[4824]: I1211 02:01:52.626239 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:52 crc kubenswrapper[4824]: I1211 02:01:52.626577 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:52 crc kubenswrapper[4824]: I1211 02:01:52.626724 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:52 crc kubenswrapper[4824]: I1211 02:01:52.626852 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:52 crc kubenswrapper[4824]: I1211 02:01:52.626986 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:52Z","lastTransitionTime":"2025-12-11T02:01:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:52 crc kubenswrapper[4824]: I1211 02:01:52.730358 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:52 crc kubenswrapper[4824]: I1211 02:01:52.730425 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:52 crc kubenswrapper[4824]: I1211 02:01:52.730446 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:52 crc kubenswrapper[4824]: I1211 02:01:52.730473 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:52 crc kubenswrapper[4824]: I1211 02:01:52.730493 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:52Z","lastTransitionTime":"2025-12-11T02:01:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:52 crc kubenswrapper[4824]: I1211 02:01:52.833599 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:52 crc kubenswrapper[4824]: I1211 02:01:52.833648 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:52 crc kubenswrapper[4824]: I1211 02:01:52.833662 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:52 crc kubenswrapper[4824]: I1211 02:01:52.833680 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:52 crc kubenswrapper[4824]: I1211 02:01:52.833692 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:52Z","lastTransitionTime":"2025-12-11T02:01:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:52 crc kubenswrapper[4824]: I1211 02:01:52.936480 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:52 crc kubenswrapper[4824]: I1211 02:01:52.936831 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:52 crc kubenswrapper[4824]: I1211 02:01:52.937024 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:52 crc kubenswrapper[4824]: I1211 02:01:52.937276 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:52 crc kubenswrapper[4824]: I1211 02:01:52.937459 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:52Z","lastTransitionTime":"2025-12-11T02:01:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:53 crc kubenswrapper[4824]: I1211 02:01:53.040015 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:53 crc kubenswrapper[4824]: I1211 02:01:53.040739 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:53 crc kubenswrapper[4824]: I1211 02:01:53.040917 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:53 crc kubenswrapper[4824]: I1211 02:01:53.041065 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:53 crc kubenswrapper[4824]: I1211 02:01:53.041249 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:53Z","lastTransitionTime":"2025-12-11T02:01:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:53 crc kubenswrapper[4824]: I1211 02:01:53.143983 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:53 crc kubenswrapper[4824]: I1211 02:01:53.144030 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:53 crc kubenswrapper[4824]: I1211 02:01:53.144046 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:53 crc kubenswrapper[4824]: I1211 02:01:53.144068 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:53 crc kubenswrapper[4824]: I1211 02:01:53.144084 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:53Z","lastTransitionTime":"2025-12-11T02:01:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:53 crc kubenswrapper[4824]: I1211 02:01:53.247289 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:53 crc kubenswrapper[4824]: I1211 02:01:53.247614 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:53 crc kubenswrapper[4824]: I1211 02:01:53.247831 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:53 crc kubenswrapper[4824]: I1211 02:01:53.248046 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:53 crc kubenswrapper[4824]: I1211 02:01:53.248241 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:53Z","lastTransitionTime":"2025-12-11T02:01:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:53 crc kubenswrapper[4824]: I1211 02:01:53.351600 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:53 crc kubenswrapper[4824]: I1211 02:01:53.351671 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:53 crc kubenswrapper[4824]: I1211 02:01:53.351695 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:53 crc kubenswrapper[4824]: I1211 02:01:53.351721 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:53 crc kubenswrapper[4824]: I1211 02:01:53.351738 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:53Z","lastTransitionTime":"2025-12-11T02:01:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:53 crc kubenswrapper[4824]: I1211 02:01:53.454947 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:53 crc kubenswrapper[4824]: I1211 02:01:53.455385 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:53 crc kubenswrapper[4824]: I1211 02:01:53.455582 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:53 crc kubenswrapper[4824]: I1211 02:01:53.455737 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:53 crc kubenswrapper[4824]: I1211 02:01:53.455877 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:53Z","lastTransitionTime":"2025-12-11T02:01:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:53 crc kubenswrapper[4824]: I1211 02:01:53.559582 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:53 crc kubenswrapper[4824]: I1211 02:01:53.559842 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:53 crc kubenswrapper[4824]: I1211 02:01:53.559941 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:53 crc kubenswrapper[4824]: I1211 02:01:53.560036 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:53 crc kubenswrapper[4824]: I1211 02:01:53.560157 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:53Z","lastTransitionTime":"2025-12-11T02:01:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:53 crc kubenswrapper[4824]: I1211 02:01:53.632563 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 02:01:53 crc kubenswrapper[4824]: I1211 02:01:53.632619 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 02:01:53 crc kubenswrapper[4824]: I1211 02:01:53.632686 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-grnpt" Dec 11 02:01:53 crc kubenswrapper[4824]: E1211 02:01:53.633220 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 02:01:53 crc kubenswrapper[4824]: I1211 02:01:53.632729 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 02:01:53 crc kubenswrapper[4824]: E1211 02:01:53.633035 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 02:01:53 crc kubenswrapper[4824]: E1211 02:01:53.633433 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-grnpt" podUID="09a52948-0404-45ff-85b1-51479268bf71" Dec 11 02:01:53 crc kubenswrapper[4824]: E1211 02:01:53.633525 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 02:01:53 crc kubenswrapper[4824]: I1211 02:01:53.662099 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:53 crc kubenswrapper[4824]: I1211 02:01:53.662268 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:53 crc kubenswrapper[4824]: I1211 02:01:53.662293 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:53 crc kubenswrapper[4824]: I1211 02:01:53.662309 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:53 crc kubenswrapper[4824]: I1211 02:01:53.662320 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:53Z","lastTransitionTime":"2025-12-11T02:01:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:53 crc kubenswrapper[4824]: I1211 02:01:53.764897 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:53 crc kubenswrapper[4824]: I1211 02:01:53.764961 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:53 crc kubenswrapper[4824]: I1211 02:01:53.764980 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:53 crc kubenswrapper[4824]: I1211 02:01:53.765007 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:53 crc kubenswrapper[4824]: I1211 02:01:53.765027 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:53Z","lastTransitionTime":"2025-12-11T02:01:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:53 crc kubenswrapper[4824]: I1211 02:01:53.867775 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:53 crc kubenswrapper[4824]: I1211 02:01:53.867821 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:53 crc kubenswrapper[4824]: I1211 02:01:53.867838 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:53 crc kubenswrapper[4824]: I1211 02:01:53.867861 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:53 crc kubenswrapper[4824]: I1211 02:01:53.867879 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:53Z","lastTransitionTime":"2025-12-11T02:01:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:53 crc kubenswrapper[4824]: I1211 02:01:53.971505 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:53 crc kubenswrapper[4824]: I1211 02:01:53.971640 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:53 crc kubenswrapper[4824]: I1211 02:01:53.971666 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:53 crc kubenswrapper[4824]: I1211 02:01:53.971697 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:53 crc kubenswrapper[4824]: I1211 02:01:53.971718 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:53Z","lastTransitionTime":"2025-12-11T02:01:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:54 crc kubenswrapper[4824]: I1211 02:01:54.075839 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:54 crc kubenswrapper[4824]: I1211 02:01:54.075910 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:54 crc kubenswrapper[4824]: I1211 02:01:54.075927 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:54 crc kubenswrapper[4824]: I1211 02:01:54.075951 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:54 crc kubenswrapper[4824]: I1211 02:01:54.075969 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:54Z","lastTransitionTime":"2025-12-11T02:01:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:54 crc kubenswrapper[4824]: I1211 02:01:54.178779 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:54 crc kubenswrapper[4824]: I1211 02:01:54.179173 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:54 crc kubenswrapper[4824]: I1211 02:01:54.179308 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:54 crc kubenswrapper[4824]: I1211 02:01:54.179448 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:54 crc kubenswrapper[4824]: I1211 02:01:54.179589 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:54Z","lastTransitionTime":"2025-12-11T02:01:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:54 crc kubenswrapper[4824]: I1211 02:01:54.282167 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:54 crc kubenswrapper[4824]: I1211 02:01:54.282541 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:54 crc kubenswrapper[4824]: I1211 02:01:54.282767 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:54 crc kubenswrapper[4824]: I1211 02:01:54.282988 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:54 crc kubenswrapper[4824]: I1211 02:01:54.283239 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:54Z","lastTransitionTime":"2025-12-11T02:01:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:54 crc kubenswrapper[4824]: I1211 02:01:54.386296 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:54 crc kubenswrapper[4824]: I1211 02:01:54.386345 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:54 crc kubenswrapper[4824]: I1211 02:01:54.386361 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:54 crc kubenswrapper[4824]: I1211 02:01:54.386384 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:54 crc kubenswrapper[4824]: I1211 02:01:54.386400 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:54Z","lastTransitionTime":"2025-12-11T02:01:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:54 crc kubenswrapper[4824]: I1211 02:01:54.489284 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:54 crc kubenswrapper[4824]: I1211 02:01:54.489555 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:54 crc kubenswrapper[4824]: I1211 02:01:54.489720 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:54 crc kubenswrapper[4824]: I1211 02:01:54.489919 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:54 crc kubenswrapper[4824]: I1211 02:01:54.490101 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:54Z","lastTransitionTime":"2025-12-11T02:01:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:54 crc kubenswrapper[4824]: I1211 02:01:54.594346 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:54 crc kubenswrapper[4824]: I1211 02:01:54.594406 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:54 crc kubenswrapper[4824]: I1211 02:01:54.594423 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:54 crc kubenswrapper[4824]: I1211 02:01:54.594448 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:54 crc kubenswrapper[4824]: I1211 02:01:54.594467 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:54Z","lastTransitionTime":"2025-12-11T02:01:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:54 crc kubenswrapper[4824]: I1211 02:01:54.697041 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:54 crc kubenswrapper[4824]: I1211 02:01:54.697081 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:54 crc kubenswrapper[4824]: I1211 02:01:54.697108 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:54 crc kubenswrapper[4824]: I1211 02:01:54.697157 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:54 crc kubenswrapper[4824]: I1211 02:01:54.697173 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:54Z","lastTransitionTime":"2025-12-11T02:01:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:54 crc kubenswrapper[4824]: I1211 02:01:54.801036 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:54 crc kubenswrapper[4824]: I1211 02:01:54.801465 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:54 crc kubenswrapper[4824]: I1211 02:01:54.801612 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:54 crc kubenswrapper[4824]: I1211 02:01:54.801752 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:54 crc kubenswrapper[4824]: I1211 02:01:54.801905 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:54Z","lastTransitionTime":"2025-12-11T02:01:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:54 crc kubenswrapper[4824]: I1211 02:01:54.905920 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:54 crc kubenswrapper[4824]: I1211 02:01:54.905985 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:54 crc kubenswrapper[4824]: I1211 02:01:54.906005 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:54 crc kubenswrapper[4824]: I1211 02:01:54.906032 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:54 crc kubenswrapper[4824]: I1211 02:01:54.906050 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:54Z","lastTransitionTime":"2025-12-11T02:01:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:55 crc kubenswrapper[4824]: I1211 02:01:55.009500 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:55 crc kubenswrapper[4824]: I1211 02:01:55.009564 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:55 crc kubenswrapper[4824]: I1211 02:01:55.009589 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:55 crc kubenswrapper[4824]: I1211 02:01:55.009617 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:55 crc kubenswrapper[4824]: I1211 02:01:55.009637 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:55Z","lastTransitionTime":"2025-12-11T02:01:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:55 crc kubenswrapper[4824]: I1211 02:01:55.112175 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:55 crc kubenswrapper[4824]: I1211 02:01:55.112237 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:55 crc kubenswrapper[4824]: I1211 02:01:55.112254 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:55 crc kubenswrapper[4824]: I1211 02:01:55.112278 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:55 crc kubenswrapper[4824]: I1211 02:01:55.112294 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:55Z","lastTransitionTime":"2025-12-11T02:01:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:55 crc kubenswrapper[4824]: I1211 02:01:55.216232 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:55 crc kubenswrapper[4824]: I1211 02:01:55.216298 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:55 crc kubenswrapper[4824]: I1211 02:01:55.216315 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:55 crc kubenswrapper[4824]: I1211 02:01:55.216339 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:55 crc kubenswrapper[4824]: I1211 02:01:55.216359 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:55Z","lastTransitionTime":"2025-12-11T02:01:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:55 crc kubenswrapper[4824]: I1211 02:01:55.320558 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:55 crc kubenswrapper[4824]: I1211 02:01:55.320671 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:55 crc kubenswrapper[4824]: I1211 02:01:55.320699 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:55 crc kubenswrapper[4824]: I1211 02:01:55.320727 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:55 crc kubenswrapper[4824]: I1211 02:01:55.320752 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:55Z","lastTransitionTime":"2025-12-11T02:01:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:55 crc kubenswrapper[4824]: I1211 02:01:55.424546 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:55 crc kubenswrapper[4824]: I1211 02:01:55.424880 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:55 crc kubenswrapper[4824]: I1211 02:01:55.425031 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:55 crc kubenswrapper[4824]: I1211 02:01:55.425211 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:55 crc kubenswrapper[4824]: I1211 02:01:55.425373 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:55Z","lastTransitionTime":"2025-12-11T02:01:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:55 crc kubenswrapper[4824]: I1211 02:01:55.527921 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:55 crc kubenswrapper[4824]: I1211 02:01:55.527979 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:55 crc kubenswrapper[4824]: I1211 02:01:55.528001 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:55 crc kubenswrapper[4824]: I1211 02:01:55.528030 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:55 crc kubenswrapper[4824]: I1211 02:01:55.528051 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:55Z","lastTransitionTime":"2025-12-11T02:01:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:55 crc kubenswrapper[4824]: I1211 02:01:55.636953 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 02:01:55 crc kubenswrapper[4824]: E1211 02:01:55.637107 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 02:01:55 crc kubenswrapper[4824]: I1211 02:01:55.637296 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 02:01:55 crc kubenswrapper[4824]: E1211 02:01:55.637371 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 02:01:55 crc kubenswrapper[4824]: I1211 02:01:55.637509 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-grnpt" Dec 11 02:01:55 crc kubenswrapper[4824]: E1211 02:01:55.637655 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-grnpt" podUID="09a52948-0404-45ff-85b1-51479268bf71" Dec 11 02:01:55 crc kubenswrapper[4824]: I1211 02:01:55.638050 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 02:01:55 crc kubenswrapper[4824]: E1211 02:01:55.638433 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 02:01:55 crc kubenswrapper[4824]: I1211 02:01:55.638959 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:55 crc kubenswrapper[4824]: I1211 02:01:55.639249 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:55 crc kubenswrapper[4824]: I1211 02:01:55.639447 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:55 crc kubenswrapper[4824]: I1211 02:01:55.639673 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:55 crc kubenswrapper[4824]: I1211 02:01:55.639863 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:55Z","lastTransitionTime":"2025-12-11T02:01:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:55 crc kubenswrapper[4824]: I1211 02:01:55.742324 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:55 crc kubenswrapper[4824]: I1211 02:01:55.742481 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:55 crc kubenswrapper[4824]: I1211 02:01:55.742571 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:55 crc kubenswrapper[4824]: I1211 02:01:55.742649 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:55 crc kubenswrapper[4824]: I1211 02:01:55.742725 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:55Z","lastTransitionTime":"2025-12-11T02:01:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:55 crc kubenswrapper[4824]: I1211 02:01:55.845789 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:55 crc kubenswrapper[4824]: I1211 02:01:55.846021 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:55 crc kubenswrapper[4824]: I1211 02:01:55.846283 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:55 crc kubenswrapper[4824]: I1211 02:01:55.846480 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:55 crc kubenswrapper[4824]: I1211 02:01:55.846623 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:55Z","lastTransitionTime":"2025-12-11T02:01:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:55 crc kubenswrapper[4824]: I1211 02:01:55.949640 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:55 crc kubenswrapper[4824]: I1211 02:01:55.950033 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:55 crc kubenswrapper[4824]: I1211 02:01:55.950220 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:55 crc kubenswrapper[4824]: I1211 02:01:55.950372 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:55 crc kubenswrapper[4824]: I1211 02:01:55.950501 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:55Z","lastTransitionTime":"2025-12-11T02:01:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:56 crc kubenswrapper[4824]: I1211 02:01:56.053298 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:56 crc kubenswrapper[4824]: I1211 02:01:56.053345 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:56 crc kubenswrapper[4824]: I1211 02:01:56.053363 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:56 crc kubenswrapper[4824]: I1211 02:01:56.053385 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:56 crc kubenswrapper[4824]: I1211 02:01:56.053402 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:56Z","lastTransitionTime":"2025-12-11T02:01:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:56 crc kubenswrapper[4824]: I1211 02:01:56.156266 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:56 crc kubenswrapper[4824]: I1211 02:01:56.156925 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:56 crc kubenswrapper[4824]: I1211 02:01:56.157035 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:56 crc kubenswrapper[4824]: I1211 02:01:56.157134 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:56 crc kubenswrapper[4824]: I1211 02:01:56.157229 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:56Z","lastTransitionTime":"2025-12-11T02:01:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:56 crc kubenswrapper[4824]: I1211 02:01:56.394241 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:56 crc kubenswrapper[4824]: I1211 02:01:56.394301 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:56 crc kubenswrapper[4824]: I1211 02:01:56.394322 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:56 crc kubenswrapper[4824]: I1211 02:01:56.394349 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:56 crc kubenswrapper[4824]: I1211 02:01:56.394366 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:56Z","lastTransitionTime":"2025-12-11T02:01:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:56 crc kubenswrapper[4824]: I1211 02:01:56.497891 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:56 crc kubenswrapper[4824]: I1211 02:01:56.497950 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:56 crc kubenswrapper[4824]: I1211 02:01:56.497971 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:56 crc kubenswrapper[4824]: I1211 02:01:56.497996 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:56 crc kubenswrapper[4824]: I1211 02:01:56.498013 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:56Z","lastTransitionTime":"2025-12-11T02:01:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:56 crc kubenswrapper[4824]: I1211 02:01:56.600505 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:56 crc kubenswrapper[4824]: I1211 02:01:56.600593 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:56 crc kubenswrapper[4824]: I1211 02:01:56.600620 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:56 crc kubenswrapper[4824]: I1211 02:01:56.600655 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:56 crc kubenswrapper[4824]: I1211 02:01:56.600679 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:56Z","lastTransitionTime":"2025-12-11T02:01:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:56 crc kubenswrapper[4824]: I1211 02:01:56.703643 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:56 crc kubenswrapper[4824]: I1211 02:01:56.703694 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:56 crc kubenswrapper[4824]: I1211 02:01:56.703711 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:56 crc kubenswrapper[4824]: I1211 02:01:56.703735 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:56 crc kubenswrapper[4824]: I1211 02:01:56.703753 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:56Z","lastTransitionTime":"2025-12-11T02:01:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:56 crc kubenswrapper[4824]: I1211 02:01:56.807340 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:56 crc kubenswrapper[4824]: I1211 02:01:56.807412 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:56 crc kubenswrapper[4824]: I1211 02:01:56.807436 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:56 crc kubenswrapper[4824]: I1211 02:01:56.807468 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:56 crc kubenswrapper[4824]: I1211 02:01:56.807490 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:56Z","lastTransitionTime":"2025-12-11T02:01:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:56 crc kubenswrapper[4824]: I1211 02:01:56.910471 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:56 crc kubenswrapper[4824]: I1211 02:01:56.910535 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:56 crc kubenswrapper[4824]: I1211 02:01:56.910552 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:56 crc kubenswrapper[4824]: I1211 02:01:56.910577 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:56 crc kubenswrapper[4824]: I1211 02:01:56.910595 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:56Z","lastTransitionTime":"2025-12-11T02:01:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:57 crc kubenswrapper[4824]: I1211 02:01:57.013459 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:57 crc kubenswrapper[4824]: I1211 02:01:57.013534 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:57 crc kubenswrapper[4824]: I1211 02:01:57.013552 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:57 crc kubenswrapper[4824]: I1211 02:01:57.013580 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:57 crc kubenswrapper[4824]: I1211 02:01:57.013601 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:57Z","lastTransitionTime":"2025-12-11T02:01:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:57 crc kubenswrapper[4824]: I1211 02:01:57.116938 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:57 crc kubenswrapper[4824]: I1211 02:01:57.117026 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:57 crc kubenswrapper[4824]: I1211 02:01:57.117049 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:57 crc kubenswrapper[4824]: I1211 02:01:57.117078 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:57 crc kubenswrapper[4824]: I1211 02:01:57.117096 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:57Z","lastTransitionTime":"2025-12-11T02:01:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:57 crc kubenswrapper[4824]: I1211 02:01:57.220023 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:57 crc kubenswrapper[4824]: I1211 02:01:57.220078 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:57 crc kubenswrapper[4824]: I1211 02:01:57.220095 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:57 crc kubenswrapper[4824]: I1211 02:01:57.220155 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:57 crc kubenswrapper[4824]: I1211 02:01:57.220174 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:57Z","lastTransitionTime":"2025-12-11T02:01:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:57 crc kubenswrapper[4824]: I1211 02:01:57.322603 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:57 crc kubenswrapper[4824]: I1211 02:01:57.322659 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:57 crc kubenswrapper[4824]: I1211 02:01:57.322675 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:57 crc kubenswrapper[4824]: I1211 02:01:57.322697 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:57 crc kubenswrapper[4824]: I1211 02:01:57.322715 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:57Z","lastTransitionTime":"2025-12-11T02:01:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:57 crc kubenswrapper[4824]: I1211 02:01:57.425824 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:57 crc kubenswrapper[4824]: I1211 02:01:57.425880 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:57 crc kubenswrapper[4824]: I1211 02:01:57.425898 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:57 crc kubenswrapper[4824]: I1211 02:01:57.425922 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:57 crc kubenswrapper[4824]: I1211 02:01:57.425939 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:57Z","lastTransitionTime":"2025-12-11T02:01:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:57 crc kubenswrapper[4824]: I1211 02:01:57.528898 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:57 crc kubenswrapper[4824]: I1211 02:01:57.528960 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:57 crc kubenswrapper[4824]: I1211 02:01:57.528977 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:57 crc kubenswrapper[4824]: I1211 02:01:57.529000 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:57 crc kubenswrapper[4824]: I1211 02:01:57.529019 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:57Z","lastTransitionTime":"2025-12-11T02:01:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:57 crc kubenswrapper[4824]: I1211 02:01:57.631634 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-grnpt" Dec 11 02:01:57 crc kubenswrapper[4824]: I1211 02:01:57.631669 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 02:01:57 crc kubenswrapper[4824]: I1211 02:01:57.631688 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 02:01:57 crc kubenswrapper[4824]: I1211 02:01:57.631634 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 02:01:57 crc kubenswrapper[4824]: E1211 02:01:57.631807 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-grnpt" podUID="09a52948-0404-45ff-85b1-51479268bf71" Dec 11 02:01:57 crc kubenswrapper[4824]: E1211 02:01:57.631947 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 02:01:57 crc kubenswrapper[4824]: E1211 02:01:57.632061 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 02:01:57 crc kubenswrapper[4824]: E1211 02:01:57.632228 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 02:01:57 crc kubenswrapper[4824]: I1211 02:01:57.632400 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:57 crc kubenswrapper[4824]: I1211 02:01:57.632429 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:57 crc kubenswrapper[4824]: I1211 02:01:57.632445 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:57 crc kubenswrapper[4824]: I1211 02:01:57.632465 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:57 crc kubenswrapper[4824]: I1211 02:01:57.632483 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:57Z","lastTransitionTime":"2025-12-11T02:01:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:57 crc kubenswrapper[4824]: I1211 02:01:57.735478 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:57 crc kubenswrapper[4824]: I1211 02:01:57.735557 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:57 crc kubenswrapper[4824]: I1211 02:01:57.735574 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:57 crc kubenswrapper[4824]: I1211 02:01:57.735597 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:57 crc kubenswrapper[4824]: I1211 02:01:57.735615 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:57Z","lastTransitionTime":"2025-12-11T02:01:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:57 crc kubenswrapper[4824]: I1211 02:01:57.839166 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:57 crc kubenswrapper[4824]: I1211 02:01:57.839244 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:57 crc kubenswrapper[4824]: I1211 02:01:57.839271 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:57 crc kubenswrapper[4824]: I1211 02:01:57.839299 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:57 crc kubenswrapper[4824]: I1211 02:01:57.839319 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:57Z","lastTransitionTime":"2025-12-11T02:01:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:57 crc kubenswrapper[4824]: I1211 02:01:57.942938 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:57 crc kubenswrapper[4824]: I1211 02:01:57.943399 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:57 crc kubenswrapper[4824]: I1211 02:01:57.943614 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:57 crc kubenswrapper[4824]: I1211 02:01:57.943828 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:57 crc kubenswrapper[4824]: I1211 02:01:57.944044 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:57Z","lastTransitionTime":"2025-12-11T02:01:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:58 crc kubenswrapper[4824]: I1211 02:01:58.046964 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:58 crc kubenswrapper[4824]: I1211 02:01:58.047341 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:58 crc kubenswrapper[4824]: I1211 02:01:58.047540 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:58 crc kubenswrapper[4824]: I1211 02:01:58.047731 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:58 crc kubenswrapper[4824]: I1211 02:01:58.047924 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:58Z","lastTransitionTime":"2025-12-11T02:01:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:58 crc kubenswrapper[4824]: I1211 02:01:58.150819 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:58 crc kubenswrapper[4824]: I1211 02:01:58.150869 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:58 crc kubenswrapper[4824]: I1211 02:01:58.150883 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:58 crc kubenswrapper[4824]: I1211 02:01:58.150902 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:58 crc kubenswrapper[4824]: I1211 02:01:58.150917 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:58Z","lastTransitionTime":"2025-12-11T02:01:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:58 crc kubenswrapper[4824]: I1211 02:01:58.253868 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:58 crc kubenswrapper[4824]: I1211 02:01:58.253941 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:58 crc kubenswrapper[4824]: I1211 02:01:58.253966 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:58 crc kubenswrapper[4824]: I1211 02:01:58.253996 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:58 crc kubenswrapper[4824]: I1211 02:01:58.254023 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:58Z","lastTransitionTime":"2025-12-11T02:01:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:58 crc kubenswrapper[4824]: I1211 02:01:58.361614 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:58 crc kubenswrapper[4824]: I1211 02:01:58.361660 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:58 crc kubenswrapper[4824]: I1211 02:01:58.361673 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:58 crc kubenswrapper[4824]: I1211 02:01:58.361689 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:58 crc kubenswrapper[4824]: I1211 02:01:58.361703 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:58Z","lastTransitionTime":"2025-12-11T02:01:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:58 crc kubenswrapper[4824]: I1211 02:01:58.464578 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:58 crc kubenswrapper[4824]: I1211 02:01:58.464630 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:58 crc kubenswrapper[4824]: I1211 02:01:58.464647 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:58 crc kubenswrapper[4824]: I1211 02:01:58.464671 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:58 crc kubenswrapper[4824]: I1211 02:01:58.464691 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:58Z","lastTransitionTime":"2025-12-11T02:01:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:58 crc kubenswrapper[4824]: I1211 02:01:58.567638 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:58 crc kubenswrapper[4824]: I1211 02:01:58.567684 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:58 crc kubenswrapper[4824]: I1211 02:01:58.567701 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:58 crc kubenswrapper[4824]: I1211 02:01:58.567724 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:58 crc kubenswrapper[4824]: I1211 02:01:58.567741 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:58Z","lastTransitionTime":"2025-12-11T02:01:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:58 crc kubenswrapper[4824]: I1211 02:01:58.633477 4824 scope.go:117] "RemoveContainer" containerID="09bb8b4c85e1fd5fa8d62db23ebee77a25a0faa274da3bb22ab679987597f26b" Dec 11 02:01:58 crc kubenswrapper[4824]: E1211 02:01:58.633834 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-f6pgc_openshift-ovn-kubernetes(5b193bd2-b536-4056-92f1-94c9836ab2eb)\"" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" podUID="5b193bd2-b536-4056-92f1-94c9836ab2eb" Dec 11 02:01:58 crc kubenswrapper[4824]: I1211 02:01:58.657296 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:58Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:58 crc kubenswrapper[4824]: I1211 02:01:58.670815 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:58 crc kubenswrapper[4824]: I1211 02:01:58.671236 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:58 crc kubenswrapper[4824]: I1211 02:01:58.671495 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:58 crc kubenswrapper[4824]: I1211 02:01:58.671760 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:58 crc kubenswrapper[4824]: I1211 02:01:58.671966 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:58Z","lastTransitionTime":"2025-12-11T02:01:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:58 crc kubenswrapper[4824]: I1211 02:01:58.675330 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44b156e4-64a4-4d45-aa5e-9b10a862faed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7368d92144b188e95a7796fabf497b0f8d9c0b240293bdc7b4aec1ff75497bc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w62dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcf289ac27dbda4f24aabbde3fb5119329e7b4fc7d6e018168da9a69661a4c57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w62dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gx6xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:58Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:58 crc kubenswrapper[4824]: I1211 02:01:58.694288 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d14986666febd651039afb52ddf4db8969aca55e88eda9bbb20a67e35ba722ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:58Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:58 crc kubenswrapper[4824]: I1211 02:01:58.714646 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"253d4291-2b2f-487c-8862-3d305f0bdc55\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44b29bb7a6eda7cb890d36f86375b5f3f197ae5cf390f16e71139b0904873e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://922af7f67b99a0ce08776f3f75c0c377e6ea69af1720b6178a6d1e2f9b4b8ba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://136c68020325db8e6930d0f40b0d0d7e013c14c4dd5e7d1449d01d6d458104d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c89afaf56b4a9597f6543ec2e3b176c7a6dab7dc341941fc3bc8ec85cc4edaf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:00:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:58Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:58 crc kubenswrapper[4824]: I1211 02:01:58.729963 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98027cfa-5dde-4297-9f74-c4d662c7f82b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f410073059c5898660c50f775a270bebe51836f73e2a2f02db38a2aabf15931d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://335a774b8a3aff18695de4aba0b522d346aecc8ed8d96a80150301f5debce32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://339d619250c1c0a12f988a9b0ebc03874db3ac41a9d3b3f8573795467578cd0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e27af6c438009a44288f8bc10f8ebf2b61d9309f87e8d955e8d255a7190174f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e27af6c438009a44288f8bc10f8ebf2b61d9309f87e8d955e8d255a7190174f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:00:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:58Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:58 crc kubenswrapper[4824]: I1211 02:01:58.742578 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:58Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:58 crc kubenswrapper[4824]: I1211 02:01:58.757980 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"beab386b-6c65-4bc3-9ef0-dc3f68b4403f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c26f06d2ed5e6e98b61a51f073130675f2c49acda106b8aedcbb2a6d4f7580c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20af9b10a67c8b21a7abcd23e42bc412798d159c8422e7c8ff58f5b997ca19a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9bbfc6edfe050b3fa9dbeee5f98d416bcb2a1533a38608a881b75396fd0b059\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7c766f8de73ca762ffde0c18b2d1c46bb152b2b37535a611c5a0e454a21b7d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d45819f0611c91b20c83eeee1092031a3441fb69f3015241632ad419880227d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 02:01:11.070250 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 02:01:11.071673 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2752882565/tls.crt::/tmp/serving-cert-2752882565/tls.key\\\\\\\"\\\\nI1211 02:01:17.042610 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 02:01:17.048046 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 02:01:17.048080 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 02:01:17.048172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 02:01:17.048188 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 02:01:17.058248 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 02:01:17.058294 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 02:01:17.058303 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 02:01:17.058315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 02:01:17.058322 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 02:01:17.058328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 02:01:17.058335 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 02:01:17.058808 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 02:01:17.060225 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14a940a72b2404a6cbd52a69e150b03266bcd7f5c1d3f4d07b51c0c69e51786\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:00:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:58Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:58 crc kubenswrapper[4824]: I1211 02:01:58.773376 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ded146c045085be2b58821749052413a36850ede00ae1024fe2dae9fa5ac84b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:58Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:58 crc kubenswrapper[4824]: I1211 02:01:58.775499 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:58 crc kubenswrapper[4824]: I1211 02:01:58.775567 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:58 crc kubenswrapper[4824]: I1211 02:01:58.775584 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:58 crc kubenswrapper[4824]: I1211 02:01:58.775610 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:58 crc kubenswrapper[4824]: I1211 02:01:58.775628 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:58Z","lastTransitionTime":"2025-12-11T02:01:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:58 crc kubenswrapper[4824]: I1211 02:01:58.792882 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76b10bbd1e891b8c682f4b2e17d023fd787a8ed44ed94fbf7d97314d641ccf36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f32838f33832b9b67e7dd9c53340680e17721cf1c7bdf0170230c5f8feafe64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:58Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:58 crc kubenswrapper[4824]: I1211 02:01:58.808913 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kptws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68568dba-83a5-46c1-8d01-337359679d91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f965a6d6db619995132f61b5d51b66f0b2b3fdbef3fbec63e231b9a62ccd0a9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd7nf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kptws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:58Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:58 crc kubenswrapper[4824]: I1211 02:01:58.830430 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kh8g9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"50c69d41-a65b-463e-a829-ea349bd2fe46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dca6be41bd7a3ce12ecd4567c868f696b65e8eb1bad436feda6e4f9bfb8e525\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69af1f33fd0bafc929b94e702b4f46c6c7d7a362d5552b121729c477892d91d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69af1f33fd0bafc929b94e702b4f46c6c7d7a362d5552b121729c477892d91d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd9675d0afded587bc660a3862fd0c5dd6885961afacda957aa9ee8501d68a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd9675d0afded587bc660a3862fd0c5dd6885961afacda957aa9ee8501d68a69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aac94149458bd6321be90b26b48ee26428d273e7f61407d95f9c9d6b3aeae4b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aac94149458bd6321be90b26b48ee26428d273e7f61407d95f9c9d6b3aeae4b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbee11bd7cb65e031415990869b52d2b8f81bf8b56e6cc935d38e09695db7b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbee11bd7cb65e031415990869b52d2b8f81bf8b56e6cc935d38e09695db7b48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86f693ade72bd593bc16021fe4b7564f1914ec2f1cda7039474ac2fb13de10f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86f693ade72bd593bc16021fe4b7564f1914ec2f1cda7039474ac2fb13de10f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7d7674252f67bde0ded2ec8ad1d925a92a26d3e3f741dee0061b735cb0b7a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7d7674252f67bde0ded2ec8ad1d925a92a26d3e3f741dee0061b735cb0b7a02\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kh8g9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:58Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:58 crc kubenswrapper[4824]: I1211 02:01:58.849884 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6bkc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8022f696-d0e7-437b-848a-3cd25bd1f364\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3e28f6778ec0dba6ea3e6b69e5a2c1c73c0d5db1abeead241fb58efdbfaefe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hh5kn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6bkc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:58Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:58 crc kubenswrapper[4824]: I1211 02:01:58.870937 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b193bd2-b536-4056-92f1-94c9836ab2eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51f3c41d86ddc44f7eca118029876ad353317f753d0ad173ed4d30c2a15f10e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c345850891744a0240781277e9a89a59f6899c4c944995f62e44650f807440a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed2ae40b1fe28111efc76d9d72345edecb5f3ea34522b238f3730812b29f2bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce7aa078800e6f299cfe03770fa7e3b64fa3742fe0c0f383c87d84cf48569ddf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6aa9fc1f1bff192c69bd59d32a05d4a72d09641d9ce4b265bc4517147a7e15dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b4a3bc50eb80777188206adab774e8acad8b2e4952a8415f91b1c29ba9a0ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09bb8b4c85e1fd5fa8d62db23ebee77a25a0faa274da3bb22ab679987597f26b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09bb8b4c85e1fd5fa8d62db23ebee77a25a0faa274da3bb22ab679987597f26b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T02:01:43Z\\\",\\\"message\\\":\\\"ng]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-webhook\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.254\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1211 02:01:43.627342 6417 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-target-xd92c in node crc\\\\nI1211 02:01:43.627315 6417 services_controller.go:443] Built service openshift-machine-config-operator/machine-config-daemon LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.43\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:9001, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}, services.lbConfig{vips:[]string{\\\\\\\"10.217.4.43\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8798, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLo\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:42Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-f6pgc_openshift-ovn-kubernetes(5b193bd2-b536-4056-92f1-94c9836ab2eb)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ead58694813806d307bbbc6a7b3cf41730d1c059b231a3002684bf09c10c7f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f6pgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:58Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:58 crc kubenswrapper[4824]: I1211 02:01:58.877349 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:58 crc kubenswrapper[4824]: I1211 02:01:58.877373 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:58 crc kubenswrapper[4824]: I1211 02:01:58.877380 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:58 crc kubenswrapper[4824]: I1211 02:01:58.877393 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:58 crc kubenswrapper[4824]: I1211 02:01:58.877402 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:58Z","lastTransitionTime":"2025-12-11T02:01:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:58 crc kubenswrapper[4824]: I1211 02:01:58.884717 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-br4j9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce125194-9c9d-4ea4-8a72-2c30d5722295\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9798f904052e08196a0db1745060dab11e2710308bca01ff160475910f2665df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fc5s8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-br4j9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:58Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:58 crc kubenswrapper[4824]: I1211 02:01:58.899572 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8v4jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99a6bfee-afca-4726-a964-b95a948945a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b23b9d8f18db3eeadb60f83b36b05b1778cd0fa31669dad2680bb335ae004acf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4hk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3acc1940a0534ba0e6096175e41834fdd0fbcbb4c3fb6e9a38e76aedae55d2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4hk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-8v4jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:58Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:58 crc kubenswrapper[4824]: I1211 02:01:58.912857 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-grnpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09a52948-0404-45ff-85b1-51479268bf71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmmzm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmmzm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-grnpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:58Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:58 crc kubenswrapper[4824]: I1211 02:01:58.926805 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:01:58Z is after 2025-08-24T17:21:41Z" Dec 11 02:01:58 crc kubenswrapper[4824]: I1211 02:01:58.979912 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:58 crc kubenswrapper[4824]: I1211 02:01:58.979977 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:58 crc kubenswrapper[4824]: I1211 02:01:58.979994 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:58 crc kubenswrapper[4824]: I1211 02:01:58.980017 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:58 crc kubenswrapper[4824]: I1211 02:01:58.980035 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:58Z","lastTransitionTime":"2025-12-11T02:01:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:59 crc kubenswrapper[4824]: I1211 02:01:59.083561 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:59 crc kubenswrapper[4824]: I1211 02:01:59.083616 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:59 crc kubenswrapper[4824]: I1211 02:01:59.083634 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:59 crc kubenswrapper[4824]: I1211 02:01:59.083657 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:59 crc kubenswrapper[4824]: I1211 02:01:59.083675 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:59Z","lastTransitionTime":"2025-12-11T02:01:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:59 crc kubenswrapper[4824]: I1211 02:01:59.186536 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:59 crc kubenswrapper[4824]: I1211 02:01:59.186594 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:59 crc kubenswrapper[4824]: I1211 02:01:59.186611 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:59 crc kubenswrapper[4824]: I1211 02:01:59.186635 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:59 crc kubenswrapper[4824]: I1211 02:01:59.186653 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:59Z","lastTransitionTime":"2025-12-11T02:01:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:59 crc kubenswrapper[4824]: I1211 02:01:59.290077 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:59 crc kubenswrapper[4824]: I1211 02:01:59.290202 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:59 crc kubenswrapper[4824]: I1211 02:01:59.290228 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:59 crc kubenswrapper[4824]: I1211 02:01:59.290259 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:59 crc kubenswrapper[4824]: I1211 02:01:59.290282 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:59Z","lastTransitionTime":"2025-12-11T02:01:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:59 crc kubenswrapper[4824]: I1211 02:01:59.393504 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:59 crc kubenswrapper[4824]: I1211 02:01:59.393598 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:59 crc kubenswrapper[4824]: I1211 02:01:59.393618 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:59 crc kubenswrapper[4824]: I1211 02:01:59.393641 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:59 crc kubenswrapper[4824]: I1211 02:01:59.393659 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:59Z","lastTransitionTime":"2025-12-11T02:01:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:59 crc kubenswrapper[4824]: I1211 02:01:59.497001 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:59 crc kubenswrapper[4824]: I1211 02:01:59.497067 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:59 crc kubenswrapper[4824]: I1211 02:01:59.497085 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:59 crc kubenswrapper[4824]: I1211 02:01:59.497137 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:59 crc kubenswrapper[4824]: I1211 02:01:59.497155 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:59Z","lastTransitionTime":"2025-12-11T02:01:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:59 crc kubenswrapper[4824]: I1211 02:01:59.600475 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:59 crc kubenswrapper[4824]: I1211 02:01:59.600550 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:59 crc kubenswrapper[4824]: I1211 02:01:59.600568 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:59 crc kubenswrapper[4824]: I1211 02:01:59.600593 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:59 crc kubenswrapper[4824]: I1211 02:01:59.600615 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:59Z","lastTransitionTime":"2025-12-11T02:01:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:59 crc kubenswrapper[4824]: I1211 02:01:59.632179 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 02:01:59 crc kubenswrapper[4824]: I1211 02:01:59.632254 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 02:01:59 crc kubenswrapper[4824]: I1211 02:01:59.632265 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-grnpt" Dec 11 02:01:59 crc kubenswrapper[4824]: I1211 02:01:59.632383 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 02:01:59 crc kubenswrapper[4824]: E1211 02:01:59.632376 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 02:01:59 crc kubenswrapper[4824]: E1211 02:01:59.632505 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-grnpt" podUID="09a52948-0404-45ff-85b1-51479268bf71" Dec 11 02:01:59 crc kubenswrapper[4824]: E1211 02:01:59.632684 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 02:01:59 crc kubenswrapper[4824]: E1211 02:01:59.632856 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 02:01:59 crc kubenswrapper[4824]: I1211 02:01:59.703312 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:59 crc kubenswrapper[4824]: I1211 02:01:59.703372 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:59 crc kubenswrapper[4824]: I1211 02:01:59.703390 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:59 crc kubenswrapper[4824]: I1211 02:01:59.703413 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:59 crc kubenswrapper[4824]: I1211 02:01:59.703431 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:59Z","lastTransitionTime":"2025-12-11T02:01:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:59 crc kubenswrapper[4824]: I1211 02:01:59.807299 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:59 crc kubenswrapper[4824]: I1211 02:01:59.807369 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:59 crc kubenswrapper[4824]: I1211 02:01:59.807381 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:59 crc kubenswrapper[4824]: I1211 02:01:59.807415 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:59 crc kubenswrapper[4824]: I1211 02:01:59.807435 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:59Z","lastTransitionTime":"2025-12-11T02:01:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:01:59 crc kubenswrapper[4824]: I1211 02:01:59.910580 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:01:59 crc kubenswrapper[4824]: I1211 02:01:59.910652 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:01:59 crc kubenswrapper[4824]: I1211 02:01:59.910666 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:01:59 crc kubenswrapper[4824]: I1211 02:01:59.910688 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:01:59 crc kubenswrapper[4824]: I1211 02:01:59.910701 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:01:59Z","lastTransitionTime":"2025-12-11T02:01:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:00 crc kubenswrapper[4824]: I1211 02:02:00.019787 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:00 crc kubenswrapper[4824]: I1211 02:02:00.019876 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:00 crc kubenswrapper[4824]: I1211 02:02:00.019889 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:00 crc kubenswrapper[4824]: I1211 02:02:00.019911 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:00 crc kubenswrapper[4824]: I1211 02:02:00.019925 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:00Z","lastTransitionTime":"2025-12-11T02:02:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:00 crc kubenswrapper[4824]: I1211 02:02:00.122862 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:00 crc kubenswrapper[4824]: I1211 02:02:00.122919 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:00 crc kubenswrapper[4824]: I1211 02:02:00.122936 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:00 crc kubenswrapper[4824]: I1211 02:02:00.122960 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:00 crc kubenswrapper[4824]: I1211 02:02:00.122980 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:00Z","lastTransitionTime":"2025-12-11T02:02:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:00 crc kubenswrapper[4824]: I1211 02:02:00.225900 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:00 crc kubenswrapper[4824]: I1211 02:02:00.225994 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:00 crc kubenswrapper[4824]: I1211 02:02:00.226011 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:00 crc kubenswrapper[4824]: I1211 02:02:00.226035 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:00 crc kubenswrapper[4824]: I1211 02:02:00.226053 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:00Z","lastTransitionTime":"2025-12-11T02:02:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:00 crc kubenswrapper[4824]: I1211 02:02:00.329043 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:00 crc kubenswrapper[4824]: I1211 02:02:00.329143 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:00 crc kubenswrapper[4824]: I1211 02:02:00.329167 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:00 crc kubenswrapper[4824]: I1211 02:02:00.329196 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:00 crc kubenswrapper[4824]: I1211 02:02:00.329216 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:00Z","lastTransitionTime":"2025-12-11T02:02:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:00 crc kubenswrapper[4824]: I1211 02:02:00.432691 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:00 crc kubenswrapper[4824]: I1211 02:02:00.432748 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:00 crc kubenswrapper[4824]: I1211 02:02:00.432757 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:00 crc kubenswrapper[4824]: I1211 02:02:00.432779 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:00 crc kubenswrapper[4824]: I1211 02:02:00.432791 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:00Z","lastTransitionTime":"2025-12-11T02:02:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:00 crc kubenswrapper[4824]: I1211 02:02:00.536562 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:00 crc kubenswrapper[4824]: I1211 02:02:00.536707 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:00 crc kubenswrapper[4824]: I1211 02:02:00.536790 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:00 crc kubenswrapper[4824]: I1211 02:02:00.536885 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:00 crc kubenswrapper[4824]: I1211 02:02:00.536916 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:00Z","lastTransitionTime":"2025-12-11T02:02:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:00 crc kubenswrapper[4824]: I1211 02:02:00.641786 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:00 crc kubenswrapper[4824]: I1211 02:02:00.641896 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:00 crc kubenswrapper[4824]: I1211 02:02:00.641927 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:00 crc kubenswrapper[4824]: I1211 02:02:00.641970 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:00 crc kubenswrapper[4824]: I1211 02:02:00.641997 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:00Z","lastTransitionTime":"2025-12-11T02:02:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:00 crc kubenswrapper[4824]: I1211 02:02:00.744907 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:00 crc kubenswrapper[4824]: I1211 02:02:00.744977 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:00 crc kubenswrapper[4824]: I1211 02:02:00.744994 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:00 crc kubenswrapper[4824]: I1211 02:02:00.745018 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:00 crc kubenswrapper[4824]: I1211 02:02:00.745037 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:00Z","lastTransitionTime":"2025-12-11T02:02:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:00 crc kubenswrapper[4824]: I1211 02:02:00.848582 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:00 crc kubenswrapper[4824]: I1211 02:02:00.848637 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:00 crc kubenswrapper[4824]: I1211 02:02:00.848649 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:00 crc kubenswrapper[4824]: I1211 02:02:00.848668 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:00 crc kubenswrapper[4824]: I1211 02:02:00.848680 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:00Z","lastTransitionTime":"2025-12-11T02:02:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:00 crc kubenswrapper[4824]: I1211 02:02:00.951497 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:00 crc kubenswrapper[4824]: I1211 02:02:00.951563 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:00 crc kubenswrapper[4824]: I1211 02:02:00.951584 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:00 crc kubenswrapper[4824]: I1211 02:02:00.951609 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:00 crc kubenswrapper[4824]: I1211 02:02:00.951626 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:00Z","lastTransitionTime":"2025-12-11T02:02:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:01 crc kubenswrapper[4824]: I1211 02:02:01.054944 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:01 crc kubenswrapper[4824]: I1211 02:02:01.055019 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:01 crc kubenswrapper[4824]: I1211 02:02:01.055036 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:01 crc kubenswrapper[4824]: I1211 02:02:01.055060 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:01 crc kubenswrapper[4824]: I1211 02:02:01.055078 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:01Z","lastTransitionTime":"2025-12-11T02:02:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:01 crc kubenswrapper[4824]: I1211 02:02:01.158175 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:01 crc kubenswrapper[4824]: I1211 02:02:01.158237 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:01 crc kubenswrapper[4824]: I1211 02:02:01.158255 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:01 crc kubenswrapper[4824]: I1211 02:02:01.158278 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:01 crc kubenswrapper[4824]: I1211 02:02:01.158297 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:01Z","lastTransitionTime":"2025-12-11T02:02:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:01 crc kubenswrapper[4824]: I1211 02:02:01.261270 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:01 crc kubenswrapper[4824]: I1211 02:02:01.261355 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:01 crc kubenswrapper[4824]: I1211 02:02:01.261377 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:01 crc kubenswrapper[4824]: I1211 02:02:01.261400 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:01 crc kubenswrapper[4824]: I1211 02:02:01.261418 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:01Z","lastTransitionTime":"2025-12-11T02:02:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:01 crc kubenswrapper[4824]: I1211 02:02:01.364634 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:01 crc kubenswrapper[4824]: I1211 02:02:01.364764 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:01 crc kubenswrapper[4824]: I1211 02:02:01.364794 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:01 crc kubenswrapper[4824]: I1211 02:02:01.364826 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:01 crc kubenswrapper[4824]: I1211 02:02:01.364851 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:01Z","lastTransitionTime":"2025-12-11T02:02:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:01 crc kubenswrapper[4824]: I1211 02:02:01.467458 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:01 crc kubenswrapper[4824]: I1211 02:02:01.467555 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:01 crc kubenswrapper[4824]: I1211 02:02:01.467572 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:01 crc kubenswrapper[4824]: I1211 02:02:01.467595 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:01 crc kubenswrapper[4824]: I1211 02:02:01.467613 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:01Z","lastTransitionTime":"2025-12-11T02:02:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:01 crc kubenswrapper[4824]: I1211 02:02:01.569868 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:01 crc kubenswrapper[4824]: I1211 02:02:01.569939 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:01 crc kubenswrapper[4824]: I1211 02:02:01.569958 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:01 crc kubenswrapper[4824]: I1211 02:02:01.569981 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:01 crc kubenswrapper[4824]: I1211 02:02:01.569999 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:01Z","lastTransitionTime":"2025-12-11T02:02:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:01 crc kubenswrapper[4824]: I1211 02:02:01.614882 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:01 crc kubenswrapper[4824]: I1211 02:02:01.614948 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:01 crc kubenswrapper[4824]: I1211 02:02:01.614970 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:01 crc kubenswrapper[4824]: I1211 02:02:01.614998 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:01 crc kubenswrapper[4824]: I1211 02:02:01.615020 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:01Z","lastTransitionTime":"2025-12-11T02:02:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:01 crc kubenswrapper[4824]: I1211 02:02:01.631718 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 02:02:01 crc kubenswrapper[4824]: I1211 02:02:01.631778 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-grnpt" Dec 11 02:02:01 crc kubenswrapper[4824]: I1211 02:02:01.631815 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 02:02:01 crc kubenswrapper[4824]: I1211 02:02:01.631776 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 02:02:01 crc kubenswrapper[4824]: E1211 02:02:01.631892 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 02:02:01 crc kubenswrapper[4824]: E1211 02:02:01.632000 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-grnpt" podUID="09a52948-0404-45ff-85b1-51479268bf71" Dec 11 02:02:01 crc kubenswrapper[4824]: E1211 02:02:01.632207 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 02:02:01 crc kubenswrapper[4824]: E1211 02:02:01.632360 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 02:02:01 crc kubenswrapper[4824]: E1211 02:02:01.637808 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:02:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:02:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:02:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:02:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d57bece6-6fc6-4a84-b8a8-e1678f9e50db\\\",\\\"systemUUID\\\":\\\"f76b32e4-2764-4219-bbf4-f7e56d76df3b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:01Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:01 crc kubenswrapper[4824]: I1211 02:02:01.642824 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:01 crc kubenswrapper[4824]: I1211 02:02:01.642887 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:01 crc kubenswrapper[4824]: I1211 02:02:01.642906 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:01 crc kubenswrapper[4824]: I1211 02:02:01.642932 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:01 crc kubenswrapper[4824]: I1211 02:02:01.642952 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:01Z","lastTransitionTime":"2025-12-11T02:02:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:01 crc kubenswrapper[4824]: E1211 02:02:01.663389 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:02:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:02:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:02:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:02:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d57bece6-6fc6-4a84-b8a8-e1678f9e50db\\\",\\\"systemUUID\\\":\\\"f76b32e4-2764-4219-bbf4-f7e56d76df3b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:01Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:01 crc kubenswrapper[4824]: I1211 02:02:01.667556 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:01 crc kubenswrapper[4824]: I1211 02:02:01.667606 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:01 crc kubenswrapper[4824]: I1211 02:02:01.667622 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:01 crc kubenswrapper[4824]: I1211 02:02:01.667646 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:01 crc kubenswrapper[4824]: I1211 02:02:01.667667 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:01Z","lastTransitionTime":"2025-12-11T02:02:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:01 crc kubenswrapper[4824]: E1211 02:02:01.685546 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:02:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:02:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:02:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:02:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d57bece6-6fc6-4a84-b8a8-e1678f9e50db\\\",\\\"systemUUID\\\":\\\"f76b32e4-2764-4219-bbf4-f7e56d76df3b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:01Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:01 crc kubenswrapper[4824]: I1211 02:02:01.689746 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:01 crc kubenswrapper[4824]: I1211 02:02:01.689817 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:01 crc kubenswrapper[4824]: I1211 02:02:01.689840 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:01 crc kubenswrapper[4824]: I1211 02:02:01.689878 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:01 crc kubenswrapper[4824]: I1211 02:02:01.689901 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:01Z","lastTransitionTime":"2025-12-11T02:02:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:01 crc kubenswrapper[4824]: E1211 02:02:01.710520 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:02:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:02:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:02:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:02:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d57bece6-6fc6-4a84-b8a8-e1678f9e50db\\\",\\\"systemUUID\\\":\\\"f76b32e4-2764-4219-bbf4-f7e56d76df3b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:01Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:01 crc kubenswrapper[4824]: I1211 02:02:01.715999 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:01 crc kubenswrapper[4824]: I1211 02:02:01.716066 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:01 crc kubenswrapper[4824]: I1211 02:02:01.716089 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:01 crc kubenswrapper[4824]: I1211 02:02:01.716183 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:01 crc kubenswrapper[4824]: I1211 02:02:01.716212 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:01Z","lastTransitionTime":"2025-12-11T02:02:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:01 crc kubenswrapper[4824]: E1211 02:02:01.736148 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:02:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:02:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:02:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:02:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d57bece6-6fc6-4a84-b8a8-e1678f9e50db\\\",\\\"systemUUID\\\":\\\"f76b32e4-2764-4219-bbf4-f7e56d76df3b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:01Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:01 crc kubenswrapper[4824]: E1211 02:02:01.736417 4824 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 11 02:02:01 crc kubenswrapper[4824]: I1211 02:02:01.739869 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:01 crc kubenswrapper[4824]: I1211 02:02:01.739935 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:01 crc kubenswrapper[4824]: I1211 02:02:01.739953 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:01 crc kubenswrapper[4824]: I1211 02:02:01.739978 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:01 crc kubenswrapper[4824]: I1211 02:02:01.739995 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:01Z","lastTransitionTime":"2025-12-11T02:02:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:01 crc kubenswrapper[4824]: I1211 02:02:01.843454 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:01 crc kubenswrapper[4824]: I1211 02:02:01.843563 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:01 crc kubenswrapper[4824]: I1211 02:02:01.843576 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:01 crc kubenswrapper[4824]: I1211 02:02:01.843594 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:01 crc kubenswrapper[4824]: I1211 02:02:01.843612 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:01Z","lastTransitionTime":"2025-12-11T02:02:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:01 crc kubenswrapper[4824]: I1211 02:02:01.946633 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:01 crc kubenswrapper[4824]: I1211 02:02:01.946691 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:01 crc kubenswrapper[4824]: I1211 02:02:01.946710 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:01 crc kubenswrapper[4824]: I1211 02:02:01.946738 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:01 crc kubenswrapper[4824]: I1211 02:02:01.946762 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:01Z","lastTransitionTime":"2025-12-11T02:02:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:02 crc kubenswrapper[4824]: I1211 02:02:02.049699 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:02 crc kubenswrapper[4824]: I1211 02:02:02.049769 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:02 crc kubenswrapper[4824]: I1211 02:02:02.049786 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:02 crc kubenswrapper[4824]: I1211 02:02:02.049808 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:02 crc kubenswrapper[4824]: I1211 02:02:02.049825 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:02Z","lastTransitionTime":"2025-12-11T02:02:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:02 crc kubenswrapper[4824]: I1211 02:02:02.152051 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:02 crc kubenswrapper[4824]: I1211 02:02:02.152136 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:02 crc kubenswrapper[4824]: I1211 02:02:02.152156 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:02 crc kubenswrapper[4824]: I1211 02:02:02.152180 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:02 crc kubenswrapper[4824]: I1211 02:02:02.152197 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:02Z","lastTransitionTime":"2025-12-11T02:02:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:02 crc kubenswrapper[4824]: I1211 02:02:02.254271 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:02 crc kubenswrapper[4824]: I1211 02:02:02.254333 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:02 crc kubenswrapper[4824]: I1211 02:02:02.254343 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:02 crc kubenswrapper[4824]: I1211 02:02:02.254359 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:02 crc kubenswrapper[4824]: I1211 02:02:02.254371 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:02Z","lastTransitionTime":"2025-12-11T02:02:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:02 crc kubenswrapper[4824]: I1211 02:02:02.357280 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:02 crc kubenswrapper[4824]: I1211 02:02:02.357344 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:02 crc kubenswrapper[4824]: I1211 02:02:02.357360 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:02 crc kubenswrapper[4824]: I1211 02:02:02.357384 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:02 crc kubenswrapper[4824]: I1211 02:02:02.357401 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:02Z","lastTransitionTime":"2025-12-11T02:02:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:02 crc kubenswrapper[4824]: I1211 02:02:02.460652 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:02 crc kubenswrapper[4824]: I1211 02:02:02.460701 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:02 crc kubenswrapper[4824]: I1211 02:02:02.460714 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:02 crc kubenswrapper[4824]: I1211 02:02:02.460731 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:02 crc kubenswrapper[4824]: I1211 02:02:02.460742 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:02Z","lastTransitionTime":"2025-12-11T02:02:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:02 crc kubenswrapper[4824]: I1211 02:02:02.563665 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:02 crc kubenswrapper[4824]: I1211 02:02:02.563704 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:02 crc kubenswrapper[4824]: I1211 02:02:02.563713 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:02 crc kubenswrapper[4824]: I1211 02:02:02.563726 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:02 crc kubenswrapper[4824]: I1211 02:02:02.563736 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:02Z","lastTransitionTime":"2025-12-11T02:02:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:02 crc kubenswrapper[4824]: I1211 02:02:02.666434 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:02 crc kubenswrapper[4824]: I1211 02:02:02.666513 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:02 crc kubenswrapper[4824]: I1211 02:02:02.666536 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:02 crc kubenswrapper[4824]: I1211 02:02:02.666568 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:02 crc kubenswrapper[4824]: I1211 02:02:02.666591 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:02Z","lastTransitionTime":"2025-12-11T02:02:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:02 crc kubenswrapper[4824]: I1211 02:02:02.769414 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:02 crc kubenswrapper[4824]: I1211 02:02:02.769468 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:02 crc kubenswrapper[4824]: I1211 02:02:02.769485 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:02 crc kubenswrapper[4824]: I1211 02:02:02.769507 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:02 crc kubenswrapper[4824]: I1211 02:02:02.769523 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:02Z","lastTransitionTime":"2025-12-11T02:02:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:02 crc kubenswrapper[4824]: I1211 02:02:02.872012 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:02 crc kubenswrapper[4824]: I1211 02:02:02.872063 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:02 crc kubenswrapper[4824]: I1211 02:02:02.872078 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:02 crc kubenswrapper[4824]: I1211 02:02:02.872101 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:02 crc kubenswrapper[4824]: I1211 02:02:02.872143 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:02Z","lastTransitionTime":"2025-12-11T02:02:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:02 crc kubenswrapper[4824]: I1211 02:02:02.974905 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:02 crc kubenswrapper[4824]: I1211 02:02:02.975031 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:02 crc kubenswrapper[4824]: I1211 02:02:02.975053 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:02 crc kubenswrapper[4824]: I1211 02:02:02.975156 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:02 crc kubenswrapper[4824]: I1211 02:02:02.975196 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:02Z","lastTransitionTime":"2025-12-11T02:02:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:03 crc kubenswrapper[4824]: I1211 02:02:03.078833 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:03 crc kubenswrapper[4824]: I1211 02:02:03.078878 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:03 crc kubenswrapper[4824]: I1211 02:02:03.078890 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:03 crc kubenswrapper[4824]: I1211 02:02:03.078906 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:03 crc kubenswrapper[4824]: I1211 02:02:03.078917 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:03Z","lastTransitionTime":"2025-12-11T02:02:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:03 crc kubenswrapper[4824]: I1211 02:02:03.180809 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:03 crc kubenswrapper[4824]: I1211 02:02:03.180853 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:03 crc kubenswrapper[4824]: I1211 02:02:03.180866 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:03 crc kubenswrapper[4824]: I1211 02:02:03.180883 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:03 crc kubenswrapper[4824]: I1211 02:02:03.180895 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:03Z","lastTransitionTime":"2025-12-11T02:02:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:03 crc kubenswrapper[4824]: I1211 02:02:03.284148 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:03 crc kubenswrapper[4824]: I1211 02:02:03.284207 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:03 crc kubenswrapper[4824]: I1211 02:02:03.284224 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:03 crc kubenswrapper[4824]: I1211 02:02:03.284247 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:03 crc kubenswrapper[4824]: I1211 02:02:03.284265 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:03Z","lastTransitionTime":"2025-12-11T02:02:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:03 crc kubenswrapper[4824]: I1211 02:02:03.388046 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:03 crc kubenswrapper[4824]: I1211 02:02:03.388147 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:03 crc kubenswrapper[4824]: I1211 02:02:03.388166 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:03 crc kubenswrapper[4824]: I1211 02:02:03.388199 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:03 crc kubenswrapper[4824]: I1211 02:02:03.388218 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:03Z","lastTransitionTime":"2025-12-11T02:02:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:03 crc kubenswrapper[4824]: I1211 02:02:03.490851 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:03 crc kubenswrapper[4824]: I1211 02:02:03.490972 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:03 crc kubenswrapper[4824]: I1211 02:02:03.490985 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:03 crc kubenswrapper[4824]: I1211 02:02:03.491002 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:03 crc kubenswrapper[4824]: I1211 02:02:03.491011 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:03Z","lastTransitionTime":"2025-12-11T02:02:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:03 crc kubenswrapper[4824]: I1211 02:02:03.594057 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:03 crc kubenswrapper[4824]: I1211 02:02:03.594094 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:03 crc kubenswrapper[4824]: I1211 02:02:03.594104 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:03 crc kubenswrapper[4824]: I1211 02:02:03.594137 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:03 crc kubenswrapper[4824]: I1211 02:02:03.594146 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:03Z","lastTransitionTime":"2025-12-11T02:02:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:03 crc kubenswrapper[4824]: I1211 02:02:03.632242 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-grnpt" Dec 11 02:02:03 crc kubenswrapper[4824]: I1211 02:02:03.632279 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 02:02:03 crc kubenswrapper[4824]: I1211 02:02:03.632289 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 02:02:03 crc kubenswrapper[4824]: E1211 02:02:03.632426 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 02:02:03 crc kubenswrapper[4824]: I1211 02:02:03.632447 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 02:02:03 crc kubenswrapper[4824]: E1211 02:02:03.632523 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 02:02:03 crc kubenswrapper[4824]: E1211 02:02:03.632679 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 02:02:03 crc kubenswrapper[4824]: E1211 02:02:03.632863 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-grnpt" podUID="09a52948-0404-45ff-85b1-51479268bf71" Dec 11 02:02:03 crc kubenswrapper[4824]: I1211 02:02:03.664054 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/09a52948-0404-45ff-85b1-51479268bf71-metrics-certs\") pod \"network-metrics-daemon-grnpt\" (UID: \"09a52948-0404-45ff-85b1-51479268bf71\") " pod="openshift-multus/network-metrics-daemon-grnpt" Dec 11 02:02:03 crc kubenswrapper[4824]: E1211 02:02:03.664237 4824 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 11 02:02:03 crc kubenswrapper[4824]: E1211 02:02:03.664293 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/09a52948-0404-45ff-85b1-51479268bf71-metrics-certs podName:09a52948-0404-45ff-85b1-51479268bf71 nodeName:}" failed. No retries permitted until 2025-12-11 02:02:35.664275524 +0000 UTC m=+97.353313003 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/09a52948-0404-45ff-85b1-51479268bf71-metrics-certs") pod "network-metrics-daemon-grnpt" (UID: "09a52948-0404-45ff-85b1-51479268bf71") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 11 02:02:03 crc kubenswrapper[4824]: I1211 02:02:03.696206 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:03 crc kubenswrapper[4824]: I1211 02:02:03.696244 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:03 crc kubenswrapper[4824]: I1211 02:02:03.696253 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:03 crc kubenswrapper[4824]: I1211 02:02:03.696267 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:03 crc kubenswrapper[4824]: I1211 02:02:03.696276 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:03Z","lastTransitionTime":"2025-12-11T02:02:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:03 crc kubenswrapper[4824]: I1211 02:02:03.798621 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:03 crc kubenswrapper[4824]: I1211 02:02:03.798678 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:03 crc kubenswrapper[4824]: I1211 02:02:03.798687 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:03 crc kubenswrapper[4824]: I1211 02:02:03.798701 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:03 crc kubenswrapper[4824]: I1211 02:02:03.798710 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:03Z","lastTransitionTime":"2025-12-11T02:02:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:03 crc kubenswrapper[4824]: I1211 02:02:03.901212 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:03 crc kubenswrapper[4824]: I1211 02:02:03.901251 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:03 crc kubenswrapper[4824]: I1211 02:02:03.901264 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:03 crc kubenswrapper[4824]: I1211 02:02:03.901280 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:03 crc kubenswrapper[4824]: I1211 02:02:03.901293 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:03Z","lastTransitionTime":"2025-12-11T02:02:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:04 crc kubenswrapper[4824]: I1211 02:02:04.004181 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:04 crc kubenswrapper[4824]: I1211 02:02:04.004222 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:04 crc kubenswrapper[4824]: I1211 02:02:04.004231 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:04 crc kubenswrapper[4824]: I1211 02:02:04.004247 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:04 crc kubenswrapper[4824]: I1211 02:02:04.004257 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:04Z","lastTransitionTime":"2025-12-11T02:02:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:04 crc kubenswrapper[4824]: I1211 02:02:04.107476 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:04 crc kubenswrapper[4824]: I1211 02:02:04.107511 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:04 crc kubenswrapper[4824]: I1211 02:02:04.107525 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:04 crc kubenswrapper[4824]: I1211 02:02:04.107545 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:04 crc kubenswrapper[4824]: I1211 02:02:04.107556 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:04Z","lastTransitionTime":"2025-12-11T02:02:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:04 crc kubenswrapper[4824]: I1211 02:02:04.165895 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6bkc2_8022f696-d0e7-437b-848a-3cd25bd1f364/kube-multus/0.log" Dec 11 02:02:04 crc kubenswrapper[4824]: I1211 02:02:04.165973 4824 generic.go:334] "Generic (PLEG): container finished" podID="8022f696-d0e7-437b-848a-3cd25bd1f364" containerID="f3e28f6778ec0dba6ea3e6b69e5a2c1c73c0d5db1abeead241fb58efdbfaefe4" exitCode=1 Dec 11 02:02:04 crc kubenswrapper[4824]: I1211 02:02:04.166057 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-6bkc2" event={"ID":"8022f696-d0e7-437b-848a-3cd25bd1f364","Type":"ContainerDied","Data":"f3e28f6778ec0dba6ea3e6b69e5a2c1c73c0d5db1abeead241fb58efdbfaefe4"} Dec 11 02:02:04 crc kubenswrapper[4824]: I1211 02:02:04.166909 4824 scope.go:117] "RemoveContainer" containerID="f3e28f6778ec0dba6ea3e6b69e5a2c1c73c0d5db1abeead241fb58efdbfaefe4" Dec 11 02:02:04 crc kubenswrapper[4824]: I1211 02:02:04.182843 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"beab386b-6c65-4bc3-9ef0-dc3f68b4403f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c26f06d2ed5e6e98b61a51f073130675f2c49acda106b8aedcbb2a6d4f7580c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20af9b10a67c8b21a7abcd23e42bc412798d159c8422e7c8ff58f5b997ca19a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9bbfc6edfe050b3fa9dbeee5f98d416bcb2a1533a38608a881b75396fd0b059\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7c766f8de73ca762ffde0c18b2d1c46bb152b2b37535a611c5a0e454a21b7d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d45819f0611c91b20c83eeee1092031a3441fb69f3015241632ad419880227d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 02:01:11.070250 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 02:01:11.071673 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2752882565/tls.crt::/tmp/serving-cert-2752882565/tls.key\\\\\\\"\\\\nI1211 02:01:17.042610 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 02:01:17.048046 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 02:01:17.048080 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 02:01:17.048172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 02:01:17.048188 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 02:01:17.058248 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 02:01:17.058294 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 02:01:17.058303 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 02:01:17.058315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 02:01:17.058322 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 02:01:17.058328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 02:01:17.058335 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 02:01:17.058808 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 02:01:17.060225 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14a940a72b2404a6cbd52a69e150b03266bcd7f5c1d3f4d07b51c0c69e51786\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:00:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:04Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:04 crc kubenswrapper[4824]: I1211 02:02:04.195313 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ded146c045085be2b58821749052413a36850ede00ae1024fe2dae9fa5ac84b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:04Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:04 crc kubenswrapper[4824]: I1211 02:02:04.209528 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:04 crc kubenswrapper[4824]: I1211 02:02:04.209585 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:04 crc kubenswrapper[4824]: I1211 02:02:04.209607 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:04 crc kubenswrapper[4824]: I1211 02:02:04.209630 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:04 crc kubenswrapper[4824]: I1211 02:02:04.209650 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:04Z","lastTransitionTime":"2025-12-11T02:02:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:04 crc kubenswrapper[4824]: I1211 02:02:04.213057 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76b10bbd1e891b8c682f4b2e17d023fd787a8ed44ed94fbf7d97314d641ccf36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f32838f33832b9b67e7dd9c53340680e17721cf1c7bdf0170230c5f8feafe64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:04Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:04 crc kubenswrapper[4824]: I1211 02:02:04.230578 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kptws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68568dba-83a5-46c1-8d01-337359679d91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f965a6d6db619995132f61b5d51b66f0b2b3fdbef3fbec63e231b9a62ccd0a9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd7nf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kptws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:04Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:04 crc kubenswrapper[4824]: I1211 02:02:04.248257 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kh8g9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"50c69d41-a65b-463e-a829-ea349bd2fe46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dca6be41bd7a3ce12ecd4567c868f696b65e8eb1bad436feda6e4f9bfb8e525\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69af1f33fd0bafc929b94e702b4f46c6c7d7a362d5552b121729c477892d91d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69af1f33fd0bafc929b94e702b4f46c6c7d7a362d5552b121729c477892d91d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd9675d0afded587bc660a3862fd0c5dd6885961afacda957aa9ee8501d68a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd9675d0afded587bc660a3862fd0c5dd6885961afacda957aa9ee8501d68a69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aac94149458bd6321be90b26b48ee26428d273e7f61407d95f9c9d6b3aeae4b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aac94149458bd6321be90b26b48ee26428d273e7f61407d95f9c9d6b3aeae4b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbee11bd7cb65e031415990869b52d2b8f81bf8b56e6cc935d38e09695db7b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbee11bd7cb65e031415990869b52d2b8f81bf8b56e6cc935d38e09695db7b48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86f693ade72bd593bc16021fe4b7564f1914ec2f1cda7039474ac2fb13de10f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86f693ade72bd593bc16021fe4b7564f1914ec2f1cda7039474ac2fb13de10f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7d7674252f67bde0ded2ec8ad1d925a92a26d3e3f741dee0061b735cb0b7a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7d7674252f67bde0ded2ec8ad1d925a92a26d3e3f741dee0061b735cb0b7a02\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kh8g9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:04Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:04 crc kubenswrapper[4824]: I1211 02:02:04.262984 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6bkc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8022f696-d0e7-437b-848a-3cd25bd1f364\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3e28f6778ec0dba6ea3e6b69e5a2c1c73c0d5db1abeead241fb58efdbfaefe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3e28f6778ec0dba6ea3e6b69e5a2c1c73c0d5db1abeead241fb58efdbfaefe4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T02:02:03Z\\\",\\\"message\\\":\\\"2025-12-11T02:01:18+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_00fc4257-9dfb-4131-a927-0fc54bfa1304\\\\n2025-12-11T02:01:18+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_00fc4257-9dfb-4131-a927-0fc54bfa1304 to /host/opt/cni/bin/\\\\n2025-12-11T02:01:18Z [verbose] multus-daemon started\\\\n2025-12-11T02:01:18Z [verbose] Readiness Indicator file check\\\\n2025-12-11T02:02:03Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hh5kn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6bkc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:04Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:04 crc kubenswrapper[4824]: I1211 02:02:04.280532 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b193bd2-b536-4056-92f1-94c9836ab2eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51f3c41d86ddc44f7eca118029876ad353317f753d0ad173ed4d30c2a15f10e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c345850891744a0240781277e9a89a59f6899c4c944995f62e44650f807440a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed2ae40b1fe28111efc76d9d72345edecb5f3ea34522b238f3730812b29f2bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce7aa078800e6f299cfe03770fa7e3b64fa3742fe0c0f383c87d84cf48569ddf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6aa9fc1f1bff192c69bd59d32a05d4a72d09641d9ce4b265bc4517147a7e15dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b4a3bc50eb80777188206adab774e8acad8b2e4952a8415f91b1c29ba9a0ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09bb8b4c85e1fd5fa8d62db23ebee77a25a0faa274da3bb22ab679987597f26b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09bb8b4c85e1fd5fa8d62db23ebee77a25a0faa274da3bb22ab679987597f26b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T02:01:43Z\\\",\\\"message\\\":\\\"ng]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-webhook\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.254\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1211 02:01:43.627342 6417 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-target-xd92c in node crc\\\\nI1211 02:01:43.627315 6417 services_controller.go:443] Built service openshift-machine-config-operator/machine-config-daemon LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.43\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:9001, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}, services.lbConfig{vips:[]string{\\\\\\\"10.217.4.43\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8798, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLo\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:42Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-f6pgc_openshift-ovn-kubernetes(5b193bd2-b536-4056-92f1-94c9836ab2eb)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ead58694813806d307bbbc6a7b3cf41730d1c059b231a3002684bf09c10c7f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f6pgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:04Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:04 crc kubenswrapper[4824]: I1211 02:02:04.294672 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-br4j9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce125194-9c9d-4ea4-8a72-2c30d5722295\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9798f904052e08196a0db1745060dab11e2710308bca01ff160475910f2665df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fc5s8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-br4j9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:04Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:04 crc kubenswrapper[4824]: I1211 02:02:04.306981 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8v4jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99a6bfee-afca-4726-a964-b95a948945a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b23b9d8f18db3eeadb60f83b36b05b1778cd0fa31669dad2680bb335ae004acf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4hk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3acc1940a0534ba0e6096175e41834fdd0fbcbb4c3fb6e9a38e76aedae55d2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4hk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-8v4jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:04Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:04 crc kubenswrapper[4824]: I1211 02:02:04.311554 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:04 crc kubenswrapper[4824]: I1211 02:02:04.311589 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:04 crc kubenswrapper[4824]: I1211 02:02:04.311601 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:04 crc kubenswrapper[4824]: I1211 02:02:04.311619 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:04 crc kubenswrapper[4824]: I1211 02:02:04.311631 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:04Z","lastTransitionTime":"2025-12-11T02:02:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:04 crc kubenswrapper[4824]: I1211 02:02:04.323593 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-grnpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09a52948-0404-45ff-85b1-51479268bf71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmmzm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmmzm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-grnpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:04Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:04 crc kubenswrapper[4824]: I1211 02:02:04.337709 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:04Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:04 crc kubenswrapper[4824]: I1211 02:02:04.348780 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:04Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:04 crc kubenswrapper[4824]: I1211 02:02:04.361854 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44b156e4-64a4-4d45-aa5e-9b10a862faed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7368d92144b188e95a7796fabf497b0f8d9c0b240293bdc7b4aec1ff75497bc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w62dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcf289ac27dbda4f24aabbde3fb5119329e7b4fc7d6e018168da9a69661a4c57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w62dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gx6xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:04Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:04 crc kubenswrapper[4824]: I1211 02:02:04.374866 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d14986666febd651039afb52ddf4db8969aca55e88eda9bbb20a67e35ba722ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:04Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:04 crc kubenswrapper[4824]: I1211 02:02:04.388162 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"253d4291-2b2f-487c-8862-3d305f0bdc55\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44b29bb7a6eda7cb890d36f86375b5f3f197ae5cf390f16e71139b0904873e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://922af7f67b99a0ce08776f3f75c0c377e6ea69af1720b6178a6d1e2f9b4b8ba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://136c68020325db8e6930d0f40b0d0d7e013c14c4dd5e7d1449d01d6d458104d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c89afaf56b4a9597f6543ec2e3b176c7a6dab7dc341941fc3bc8ec85cc4edaf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:00:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:04Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:04 crc kubenswrapper[4824]: I1211 02:02:04.400494 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98027cfa-5dde-4297-9f74-c4d662c7f82b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f410073059c5898660c50f775a270bebe51836f73e2a2f02db38a2aabf15931d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://335a774b8a3aff18695de4aba0b522d346aecc8ed8d96a80150301f5debce32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://339d619250c1c0a12f988a9b0ebc03874db3ac41a9d3b3f8573795467578cd0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e27af6c438009a44288f8bc10f8ebf2b61d9309f87e8d955e8d255a7190174f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e27af6c438009a44288f8bc10f8ebf2b61d9309f87e8d955e8d255a7190174f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:00:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:04Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:04 crc kubenswrapper[4824]: I1211 02:02:04.412569 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:04Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:04 crc kubenswrapper[4824]: I1211 02:02:04.414139 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:04 crc kubenswrapper[4824]: I1211 02:02:04.414183 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:04 crc kubenswrapper[4824]: I1211 02:02:04.414193 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:04 crc kubenswrapper[4824]: I1211 02:02:04.414208 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:04 crc kubenswrapper[4824]: I1211 02:02:04.414217 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:04Z","lastTransitionTime":"2025-12-11T02:02:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:04 crc kubenswrapper[4824]: I1211 02:02:04.517080 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:04 crc kubenswrapper[4824]: I1211 02:02:04.517131 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:04 crc kubenswrapper[4824]: I1211 02:02:04.517140 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:04 crc kubenswrapper[4824]: I1211 02:02:04.517154 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:04 crc kubenswrapper[4824]: I1211 02:02:04.517164 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:04Z","lastTransitionTime":"2025-12-11T02:02:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:04 crc kubenswrapper[4824]: I1211 02:02:04.619222 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:04 crc kubenswrapper[4824]: I1211 02:02:04.619464 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:04 crc kubenswrapper[4824]: I1211 02:02:04.619628 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:04 crc kubenswrapper[4824]: I1211 02:02:04.619806 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:04 crc kubenswrapper[4824]: I1211 02:02:04.619958 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:04Z","lastTransitionTime":"2025-12-11T02:02:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:04 crc kubenswrapper[4824]: I1211 02:02:04.722758 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:04 crc kubenswrapper[4824]: I1211 02:02:04.722803 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:04 crc kubenswrapper[4824]: I1211 02:02:04.722814 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:04 crc kubenswrapper[4824]: I1211 02:02:04.722830 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:04 crc kubenswrapper[4824]: I1211 02:02:04.722840 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:04Z","lastTransitionTime":"2025-12-11T02:02:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:04 crc kubenswrapper[4824]: I1211 02:02:04.825789 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:04 crc kubenswrapper[4824]: I1211 02:02:04.825868 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:04 crc kubenswrapper[4824]: I1211 02:02:04.825890 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:04 crc kubenswrapper[4824]: I1211 02:02:04.825915 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:04 crc kubenswrapper[4824]: I1211 02:02:04.825936 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:04Z","lastTransitionTime":"2025-12-11T02:02:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:04 crc kubenswrapper[4824]: I1211 02:02:04.928756 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:04 crc kubenswrapper[4824]: I1211 02:02:04.928808 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:04 crc kubenswrapper[4824]: I1211 02:02:04.928820 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:04 crc kubenswrapper[4824]: I1211 02:02:04.928837 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:04 crc kubenswrapper[4824]: I1211 02:02:04.928852 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:04Z","lastTransitionTime":"2025-12-11T02:02:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:05 crc kubenswrapper[4824]: I1211 02:02:05.031556 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:05 crc kubenswrapper[4824]: I1211 02:02:05.031591 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:05 crc kubenswrapper[4824]: I1211 02:02:05.031600 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:05 crc kubenswrapper[4824]: I1211 02:02:05.031615 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:05 crc kubenswrapper[4824]: I1211 02:02:05.031625 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:05Z","lastTransitionTime":"2025-12-11T02:02:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:05 crc kubenswrapper[4824]: I1211 02:02:05.134349 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:05 crc kubenswrapper[4824]: I1211 02:02:05.134417 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:05 crc kubenswrapper[4824]: I1211 02:02:05.134435 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:05 crc kubenswrapper[4824]: I1211 02:02:05.134492 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:05 crc kubenswrapper[4824]: I1211 02:02:05.134522 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:05Z","lastTransitionTime":"2025-12-11T02:02:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:05 crc kubenswrapper[4824]: I1211 02:02:05.171327 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6bkc2_8022f696-d0e7-437b-848a-3cd25bd1f364/kube-multus/0.log" Dec 11 02:02:05 crc kubenswrapper[4824]: I1211 02:02:05.171410 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-6bkc2" event={"ID":"8022f696-d0e7-437b-848a-3cd25bd1f364","Type":"ContainerStarted","Data":"f40719f59881e077e7edbab3adb9efd101359bd461e5434c365e119e8cdbddb2"} Dec 11 02:02:05 crc kubenswrapper[4824]: I1211 02:02:05.188014 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"beab386b-6c65-4bc3-9ef0-dc3f68b4403f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c26f06d2ed5e6e98b61a51f073130675f2c49acda106b8aedcbb2a6d4f7580c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20af9b10a67c8b21a7abcd23e42bc412798d159c8422e7c8ff58f5b997ca19a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9bbfc6edfe050b3fa9dbeee5f98d416bcb2a1533a38608a881b75396fd0b059\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7c766f8de73ca762ffde0c18b2d1c46bb152b2b37535a611c5a0e454a21b7d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d45819f0611c91b20c83eeee1092031a3441fb69f3015241632ad419880227d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 02:01:11.070250 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 02:01:11.071673 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2752882565/tls.crt::/tmp/serving-cert-2752882565/tls.key\\\\\\\"\\\\nI1211 02:01:17.042610 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 02:01:17.048046 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 02:01:17.048080 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 02:01:17.048172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 02:01:17.048188 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 02:01:17.058248 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 02:01:17.058294 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 02:01:17.058303 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 02:01:17.058315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 02:01:17.058322 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 02:01:17.058328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 02:01:17.058335 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 02:01:17.058808 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 02:01:17.060225 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14a940a72b2404a6cbd52a69e150b03266bcd7f5c1d3f4d07b51c0c69e51786\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:00:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:05Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:05 crc kubenswrapper[4824]: I1211 02:02:05.200012 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ded146c045085be2b58821749052413a36850ede00ae1024fe2dae9fa5ac84b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:05Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:05 crc kubenswrapper[4824]: I1211 02:02:05.210995 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76b10bbd1e891b8c682f4b2e17d023fd787a8ed44ed94fbf7d97314d641ccf36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f32838f33832b9b67e7dd9c53340680e17721cf1c7bdf0170230c5f8feafe64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:05Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:05 crc kubenswrapper[4824]: I1211 02:02:05.219609 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kptws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68568dba-83a5-46c1-8d01-337359679d91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f965a6d6db619995132f61b5d51b66f0b2b3fdbef3fbec63e231b9a62ccd0a9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd7nf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kptws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:05Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:05 crc kubenswrapper[4824]: I1211 02:02:05.229945 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:05Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:05 crc kubenswrapper[4824]: I1211 02:02:05.237284 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:05 crc kubenswrapper[4824]: I1211 02:02:05.237383 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:05 crc kubenswrapper[4824]: I1211 02:02:05.237445 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:05 crc kubenswrapper[4824]: I1211 02:02:05.237469 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:05 crc kubenswrapper[4824]: I1211 02:02:05.237520 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:05Z","lastTransitionTime":"2025-12-11T02:02:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:05 crc kubenswrapper[4824]: I1211 02:02:05.243757 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kh8g9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"50c69d41-a65b-463e-a829-ea349bd2fe46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dca6be41bd7a3ce12ecd4567c868f696b65e8eb1bad436feda6e4f9bfb8e525\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69af1f33fd0bafc929b94e702b4f46c6c7d7a362d5552b121729c477892d91d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69af1f33fd0bafc929b94e702b4f46c6c7d7a362d5552b121729c477892d91d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd9675d0afded587bc660a3862fd0c5dd6885961afacda957aa9ee8501d68a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd9675d0afded587bc660a3862fd0c5dd6885961afacda957aa9ee8501d68a69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aac94149458bd6321be90b26b48ee26428d273e7f61407d95f9c9d6b3aeae4b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aac94149458bd6321be90b26b48ee26428d273e7f61407d95f9c9d6b3aeae4b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbee11bd7cb65e031415990869b52d2b8f81bf8b56e6cc935d38e09695db7b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbee11bd7cb65e031415990869b52d2b8f81bf8b56e6cc935d38e09695db7b48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86f693ade72bd593bc16021fe4b7564f1914ec2f1cda7039474ac2fb13de10f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86f693ade72bd593bc16021fe4b7564f1914ec2f1cda7039474ac2fb13de10f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7d7674252f67bde0ded2ec8ad1d925a92a26d3e3f741dee0061b735cb0b7a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7d7674252f67bde0ded2ec8ad1d925a92a26d3e3f741dee0061b735cb0b7a02\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kh8g9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:05Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:05 crc kubenswrapper[4824]: I1211 02:02:05.259094 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6bkc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8022f696-d0e7-437b-848a-3cd25bd1f364\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f40719f59881e077e7edbab3adb9efd101359bd461e5434c365e119e8cdbddb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3e28f6778ec0dba6ea3e6b69e5a2c1c73c0d5db1abeead241fb58efdbfaefe4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T02:02:03Z\\\",\\\"message\\\":\\\"2025-12-11T02:01:18+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_00fc4257-9dfb-4131-a927-0fc54bfa1304\\\\n2025-12-11T02:01:18+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_00fc4257-9dfb-4131-a927-0fc54bfa1304 to /host/opt/cni/bin/\\\\n2025-12-11T02:01:18Z [verbose] multus-daemon started\\\\n2025-12-11T02:01:18Z [verbose] Readiness Indicator file check\\\\n2025-12-11T02:02:03Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:02:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hh5kn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6bkc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:05Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:05 crc kubenswrapper[4824]: I1211 02:02:05.284239 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b193bd2-b536-4056-92f1-94c9836ab2eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51f3c41d86ddc44f7eca118029876ad353317f753d0ad173ed4d30c2a15f10e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c345850891744a0240781277e9a89a59f6899c4c944995f62e44650f807440a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed2ae40b1fe28111efc76d9d72345edecb5f3ea34522b238f3730812b29f2bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce7aa078800e6f299cfe03770fa7e3b64fa3742fe0c0f383c87d84cf48569ddf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6aa9fc1f1bff192c69bd59d32a05d4a72d09641d9ce4b265bc4517147a7e15dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b4a3bc50eb80777188206adab774e8acad8b2e4952a8415f91b1c29ba9a0ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09bb8b4c85e1fd5fa8d62db23ebee77a25a0faa274da3bb22ab679987597f26b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09bb8b4c85e1fd5fa8d62db23ebee77a25a0faa274da3bb22ab679987597f26b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T02:01:43Z\\\",\\\"message\\\":\\\"ng]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-webhook\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.254\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1211 02:01:43.627342 6417 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-target-xd92c in node crc\\\\nI1211 02:01:43.627315 6417 services_controller.go:443] Built service openshift-machine-config-operator/machine-config-daemon LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.43\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:9001, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}, services.lbConfig{vips:[]string{\\\\\\\"10.217.4.43\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8798, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLo\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:42Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-f6pgc_openshift-ovn-kubernetes(5b193bd2-b536-4056-92f1-94c9836ab2eb)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ead58694813806d307bbbc6a7b3cf41730d1c059b231a3002684bf09c10c7f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f6pgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:05Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:05 crc kubenswrapper[4824]: I1211 02:02:05.294823 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-br4j9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce125194-9c9d-4ea4-8a72-2c30d5722295\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9798f904052e08196a0db1745060dab11e2710308bca01ff160475910f2665df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fc5s8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-br4j9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:05Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:05 crc kubenswrapper[4824]: I1211 02:02:05.304698 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8v4jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99a6bfee-afca-4726-a964-b95a948945a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b23b9d8f18db3eeadb60f83b36b05b1778cd0fa31669dad2680bb335ae004acf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4hk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3acc1940a0534ba0e6096175e41834fdd0fbcbb4c3fb6e9a38e76aedae55d2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4hk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-8v4jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:05Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:05 crc kubenswrapper[4824]: I1211 02:02:05.313537 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-grnpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09a52948-0404-45ff-85b1-51479268bf71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmmzm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmmzm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-grnpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:05Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:05 crc kubenswrapper[4824]: I1211 02:02:05.329934 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d14986666febd651039afb52ddf4db8969aca55e88eda9bbb20a67e35ba722ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:05Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:05 crc kubenswrapper[4824]: I1211 02:02:05.340587 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:05 crc kubenswrapper[4824]: I1211 02:02:05.340621 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:05 crc kubenswrapper[4824]: I1211 02:02:05.340631 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:05 crc kubenswrapper[4824]: I1211 02:02:05.340645 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:05 crc kubenswrapper[4824]: I1211 02:02:05.340654 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:05Z","lastTransitionTime":"2025-12-11T02:02:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:05 crc kubenswrapper[4824]: I1211 02:02:05.345100 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:05Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:05 crc kubenswrapper[4824]: I1211 02:02:05.357573 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44b156e4-64a4-4d45-aa5e-9b10a862faed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7368d92144b188e95a7796fabf497b0f8d9c0b240293bdc7b4aec1ff75497bc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w62dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcf289ac27dbda4f24aabbde3fb5119329e7b4fc7d6e018168da9a69661a4c57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w62dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gx6xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:05Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:05 crc kubenswrapper[4824]: I1211 02:02:05.372883 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"253d4291-2b2f-487c-8862-3d305f0bdc55\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44b29bb7a6eda7cb890d36f86375b5f3f197ae5cf390f16e71139b0904873e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://922af7f67b99a0ce08776f3f75c0c377e6ea69af1720b6178a6d1e2f9b4b8ba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://136c68020325db8e6930d0f40b0d0d7e013c14c4dd5e7d1449d01d6d458104d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c89afaf56b4a9597f6543ec2e3b176c7a6dab7dc341941fc3bc8ec85cc4edaf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:00:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:05Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:05 crc kubenswrapper[4824]: I1211 02:02:05.384290 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98027cfa-5dde-4297-9f74-c4d662c7f82b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f410073059c5898660c50f775a270bebe51836f73e2a2f02db38a2aabf15931d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://335a774b8a3aff18695de4aba0b522d346aecc8ed8d96a80150301f5debce32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://339d619250c1c0a12f988a9b0ebc03874db3ac41a9d3b3f8573795467578cd0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e27af6c438009a44288f8bc10f8ebf2b61d9309f87e8d955e8d255a7190174f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e27af6c438009a44288f8bc10f8ebf2b61d9309f87e8d955e8d255a7190174f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:00:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:05Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:05 crc kubenswrapper[4824]: I1211 02:02:05.401051 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:05Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:05 crc kubenswrapper[4824]: I1211 02:02:05.442472 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:05 crc kubenswrapper[4824]: I1211 02:02:05.442506 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:05 crc kubenswrapper[4824]: I1211 02:02:05.442514 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:05 crc kubenswrapper[4824]: I1211 02:02:05.442527 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:05 crc kubenswrapper[4824]: I1211 02:02:05.442536 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:05Z","lastTransitionTime":"2025-12-11T02:02:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:05 crc kubenswrapper[4824]: I1211 02:02:05.544933 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:05 crc kubenswrapper[4824]: I1211 02:02:05.545102 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:05 crc kubenswrapper[4824]: I1211 02:02:05.545168 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:05 crc kubenswrapper[4824]: I1211 02:02:05.545193 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:05 crc kubenswrapper[4824]: I1211 02:02:05.545213 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:05Z","lastTransitionTime":"2025-12-11T02:02:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:05 crc kubenswrapper[4824]: I1211 02:02:05.631703 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 02:02:05 crc kubenswrapper[4824]: E1211 02:02:05.631858 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 02:02:05 crc kubenswrapper[4824]: I1211 02:02:05.632138 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 02:02:05 crc kubenswrapper[4824]: E1211 02:02:05.632262 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 02:02:05 crc kubenswrapper[4824]: I1211 02:02:05.632474 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-grnpt" Dec 11 02:02:05 crc kubenswrapper[4824]: E1211 02:02:05.632584 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-grnpt" podUID="09a52948-0404-45ff-85b1-51479268bf71" Dec 11 02:02:05 crc kubenswrapper[4824]: I1211 02:02:05.633040 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 02:02:05 crc kubenswrapper[4824]: E1211 02:02:05.633191 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 02:02:05 crc kubenswrapper[4824]: I1211 02:02:05.647597 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:05 crc kubenswrapper[4824]: I1211 02:02:05.647804 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:05 crc kubenswrapper[4824]: I1211 02:02:05.647935 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:05 crc kubenswrapper[4824]: I1211 02:02:05.648553 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:05 crc kubenswrapper[4824]: I1211 02:02:05.648986 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:05Z","lastTransitionTime":"2025-12-11T02:02:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:05 crc kubenswrapper[4824]: I1211 02:02:05.752620 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:05 crc kubenswrapper[4824]: I1211 02:02:05.752657 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:05 crc kubenswrapper[4824]: I1211 02:02:05.752665 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:05 crc kubenswrapper[4824]: I1211 02:02:05.752678 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:05 crc kubenswrapper[4824]: I1211 02:02:05.752688 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:05Z","lastTransitionTime":"2025-12-11T02:02:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:05 crc kubenswrapper[4824]: I1211 02:02:05.855127 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:05 crc kubenswrapper[4824]: I1211 02:02:05.855163 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:05 crc kubenswrapper[4824]: I1211 02:02:05.855174 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:05 crc kubenswrapper[4824]: I1211 02:02:05.855189 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:05 crc kubenswrapper[4824]: I1211 02:02:05.855198 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:05Z","lastTransitionTime":"2025-12-11T02:02:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:05 crc kubenswrapper[4824]: I1211 02:02:05.957990 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:05 crc kubenswrapper[4824]: I1211 02:02:05.958038 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:05 crc kubenswrapper[4824]: I1211 02:02:05.958050 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:05 crc kubenswrapper[4824]: I1211 02:02:05.958068 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:05 crc kubenswrapper[4824]: I1211 02:02:05.958080 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:05Z","lastTransitionTime":"2025-12-11T02:02:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:06 crc kubenswrapper[4824]: I1211 02:02:06.060465 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:06 crc kubenswrapper[4824]: I1211 02:02:06.060509 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:06 crc kubenswrapper[4824]: I1211 02:02:06.060517 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:06 crc kubenswrapper[4824]: I1211 02:02:06.060532 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:06 crc kubenswrapper[4824]: I1211 02:02:06.060546 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:06Z","lastTransitionTime":"2025-12-11T02:02:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:06 crc kubenswrapper[4824]: I1211 02:02:06.162685 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:06 crc kubenswrapper[4824]: I1211 02:02:06.162833 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:06 crc kubenswrapper[4824]: I1211 02:02:06.162926 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:06 crc kubenswrapper[4824]: I1211 02:02:06.163019 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:06 crc kubenswrapper[4824]: I1211 02:02:06.163102 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:06Z","lastTransitionTime":"2025-12-11T02:02:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:06 crc kubenswrapper[4824]: I1211 02:02:06.265920 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:06 crc kubenswrapper[4824]: I1211 02:02:06.265959 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:06 crc kubenswrapper[4824]: I1211 02:02:06.265972 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:06 crc kubenswrapper[4824]: I1211 02:02:06.265987 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:06 crc kubenswrapper[4824]: I1211 02:02:06.265997 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:06Z","lastTransitionTime":"2025-12-11T02:02:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:06 crc kubenswrapper[4824]: I1211 02:02:06.367744 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:06 crc kubenswrapper[4824]: I1211 02:02:06.367890 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:06 crc kubenswrapper[4824]: I1211 02:02:06.367952 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:06 crc kubenswrapper[4824]: I1211 02:02:06.368026 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:06 crc kubenswrapper[4824]: I1211 02:02:06.368099 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:06Z","lastTransitionTime":"2025-12-11T02:02:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:06 crc kubenswrapper[4824]: I1211 02:02:06.470105 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:06 crc kubenswrapper[4824]: I1211 02:02:06.470294 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:06 crc kubenswrapper[4824]: I1211 02:02:06.470381 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:06 crc kubenswrapper[4824]: I1211 02:02:06.470492 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:06 crc kubenswrapper[4824]: I1211 02:02:06.470577 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:06Z","lastTransitionTime":"2025-12-11T02:02:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:06 crc kubenswrapper[4824]: I1211 02:02:06.573350 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:06 crc kubenswrapper[4824]: I1211 02:02:06.573393 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:06 crc kubenswrapper[4824]: I1211 02:02:06.573405 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:06 crc kubenswrapper[4824]: I1211 02:02:06.573426 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:06 crc kubenswrapper[4824]: I1211 02:02:06.573439 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:06Z","lastTransitionTime":"2025-12-11T02:02:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:06 crc kubenswrapper[4824]: I1211 02:02:06.676293 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:06 crc kubenswrapper[4824]: I1211 02:02:06.676355 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:06 crc kubenswrapper[4824]: I1211 02:02:06.676381 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:06 crc kubenswrapper[4824]: I1211 02:02:06.676407 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:06 crc kubenswrapper[4824]: I1211 02:02:06.676430 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:06Z","lastTransitionTime":"2025-12-11T02:02:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:06 crc kubenswrapper[4824]: I1211 02:02:06.780578 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:06 crc kubenswrapper[4824]: I1211 02:02:06.780712 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:06 crc kubenswrapper[4824]: I1211 02:02:06.780795 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:06 crc kubenswrapper[4824]: I1211 02:02:06.780871 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:06 crc kubenswrapper[4824]: I1211 02:02:06.780938 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:06Z","lastTransitionTime":"2025-12-11T02:02:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:06 crc kubenswrapper[4824]: I1211 02:02:06.883614 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:06 crc kubenswrapper[4824]: I1211 02:02:06.883672 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:06 crc kubenswrapper[4824]: I1211 02:02:06.883691 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:06 crc kubenswrapper[4824]: I1211 02:02:06.883715 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:06 crc kubenswrapper[4824]: I1211 02:02:06.883732 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:06Z","lastTransitionTime":"2025-12-11T02:02:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:06 crc kubenswrapper[4824]: I1211 02:02:06.986475 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:06 crc kubenswrapper[4824]: I1211 02:02:06.986522 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:06 crc kubenswrapper[4824]: I1211 02:02:06.986533 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:06 crc kubenswrapper[4824]: I1211 02:02:06.986552 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:06 crc kubenswrapper[4824]: I1211 02:02:06.986562 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:06Z","lastTransitionTime":"2025-12-11T02:02:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:07 crc kubenswrapper[4824]: I1211 02:02:07.089381 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:07 crc kubenswrapper[4824]: I1211 02:02:07.089463 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:07 crc kubenswrapper[4824]: I1211 02:02:07.089488 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:07 crc kubenswrapper[4824]: I1211 02:02:07.089520 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:07 crc kubenswrapper[4824]: I1211 02:02:07.089543 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:07Z","lastTransitionTime":"2025-12-11T02:02:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:07 crc kubenswrapper[4824]: I1211 02:02:07.191879 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:07 crc kubenswrapper[4824]: I1211 02:02:07.191935 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:07 crc kubenswrapper[4824]: I1211 02:02:07.191954 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:07 crc kubenswrapper[4824]: I1211 02:02:07.191975 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:07 crc kubenswrapper[4824]: I1211 02:02:07.191994 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:07Z","lastTransitionTime":"2025-12-11T02:02:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:07 crc kubenswrapper[4824]: I1211 02:02:07.295242 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:07 crc kubenswrapper[4824]: I1211 02:02:07.295295 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:07 crc kubenswrapper[4824]: I1211 02:02:07.295311 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:07 crc kubenswrapper[4824]: I1211 02:02:07.295332 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:07 crc kubenswrapper[4824]: I1211 02:02:07.295348 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:07Z","lastTransitionTime":"2025-12-11T02:02:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:07 crc kubenswrapper[4824]: I1211 02:02:07.398200 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:07 crc kubenswrapper[4824]: I1211 02:02:07.398283 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:07 crc kubenswrapper[4824]: I1211 02:02:07.398293 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:07 crc kubenswrapper[4824]: I1211 02:02:07.398307 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:07 crc kubenswrapper[4824]: I1211 02:02:07.398316 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:07Z","lastTransitionTime":"2025-12-11T02:02:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:07 crc kubenswrapper[4824]: I1211 02:02:07.501544 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:07 crc kubenswrapper[4824]: I1211 02:02:07.501586 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:07 crc kubenswrapper[4824]: I1211 02:02:07.501594 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:07 crc kubenswrapper[4824]: I1211 02:02:07.501609 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:07 crc kubenswrapper[4824]: I1211 02:02:07.501625 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:07Z","lastTransitionTime":"2025-12-11T02:02:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:07 crc kubenswrapper[4824]: I1211 02:02:07.604224 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:07 crc kubenswrapper[4824]: I1211 02:02:07.604272 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:07 crc kubenswrapper[4824]: I1211 02:02:07.604287 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:07 crc kubenswrapper[4824]: I1211 02:02:07.604311 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:07 crc kubenswrapper[4824]: I1211 02:02:07.604328 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:07Z","lastTransitionTime":"2025-12-11T02:02:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:07 crc kubenswrapper[4824]: I1211 02:02:07.631574 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 02:02:07 crc kubenswrapper[4824]: I1211 02:02:07.631648 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 02:02:07 crc kubenswrapper[4824]: E1211 02:02:07.631713 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 02:02:07 crc kubenswrapper[4824]: E1211 02:02:07.631803 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 02:02:07 crc kubenswrapper[4824]: I1211 02:02:07.631845 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 02:02:07 crc kubenswrapper[4824]: E1211 02:02:07.631889 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 02:02:07 crc kubenswrapper[4824]: I1211 02:02:07.632133 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-grnpt" Dec 11 02:02:07 crc kubenswrapper[4824]: E1211 02:02:07.632355 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-grnpt" podUID="09a52948-0404-45ff-85b1-51479268bf71" Dec 11 02:02:07 crc kubenswrapper[4824]: I1211 02:02:07.708022 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:07 crc kubenswrapper[4824]: I1211 02:02:07.708396 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:07 crc kubenswrapper[4824]: I1211 02:02:07.708497 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:07 crc kubenswrapper[4824]: I1211 02:02:07.708586 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:07 crc kubenswrapper[4824]: I1211 02:02:07.708686 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:07Z","lastTransitionTime":"2025-12-11T02:02:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:07 crc kubenswrapper[4824]: I1211 02:02:07.811392 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:07 crc kubenswrapper[4824]: I1211 02:02:07.811487 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:07 crc kubenswrapper[4824]: I1211 02:02:07.811498 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:07 crc kubenswrapper[4824]: I1211 02:02:07.811512 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:07 crc kubenswrapper[4824]: I1211 02:02:07.811522 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:07Z","lastTransitionTime":"2025-12-11T02:02:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:07 crc kubenswrapper[4824]: I1211 02:02:07.913766 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:07 crc kubenswrapper[4824]: I1211 02:02:07.914093 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:07 crc kubenswrapper[4824]: I1211 02:02:07.914228 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:07 crc kubenswrapper[4824]: I1211 02:02:07.914340 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:07 crc kubenswrapper[4824]: I1211 02:02:07.914569 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:07Z","lastTransitionTime":"2025-12-11T02:02:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:08 crc kubenswrapper[4824]: I1211 02:02:08.017588 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:08 crc kubenswrapper[4824]: I1211 02:02:08.017928 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:08 crc kubenswrapper[4824]: I1211 02:02:08.018164 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:08 crc kubenswrapper[4824]: I1211 02:02:08.018309 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:08 crc kubenswrapper[4824]: I1211 02:02:08.018442 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:08Z","lastTransitionTime":"2025-12-11T02:02:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:08 crc kubenswrapper[4824]: I1211 02:02:08.121193 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:08 crc kubenswrapper[4824]: I1211 02:02:08.121479 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:08 crc kubenswrapper[4824]: I1211 02:02:08.121648 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:08 crc kubenswrapper[4824]: I1211 02:02:08.121780 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:08 crc kubenswrapper[4824]: I1211 02:02:08.121896 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:08Z","lastTransitionTime":"2025-12-11T02:02:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:08 crc kubenswrapper[4824]: I1211 02:02:08.224407 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:08 crc kubenswrapper[4824]: I1211 02:02:08.224448 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:08 crc kubenswrapper[4824]: I1211 02:02:08.224459 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:08 crc kubenswrapper[4824]: I1211 02:02:08.224477 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:08 crc kubenswrapper[4824]: I1211 02:02:08.224489 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:08Z","lastTransitionTime":"2025-12-11T02:02:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:08 crc kubenswrapper[4824]: I1211 02:02:08.327006 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:08 crc kubenswrapper[4824]: I1211 02:02:08.327038 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:08 crc kubenswrapper[4824]: I1211 02:02:08.327047 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:08 crc kubenswrapper[4824]: I1211 02:02:08.327060 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:08 crc kubenswrapper[4824]: I1211 02:02:08.327069 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:08Z","lastTransitionTime":"2025-12-11T02:02:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:08 crc kubenswrapper[4824]: I1211 02:02:08.429723 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:08 crc kubenswrapper[4824]: I1211 02:02:08.429778 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:08 crc kubenswrapper[4824]: I1211 02:02:08.429795 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:08 crc kubenswrapper[4824]: I1211 02:02:08.429822 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:08 crc kubenswrapper[4824]: I1211 02:02:08.429844 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:08Z","lastTransitionTime":"2025-12-11T02:02:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:08 crc kubenswrapper[4824]: I1211 02:02:08.531950 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:08 crc kubenswrapper[4824]: I1211 02:02:08.532002 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:08 crc kubenswrapper[4824]: I1211 02:02:08.532019 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:08 crc kubenswrapper[4824]: I1211 02:02:08.532042 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:08 crc kubenswrapper[4824]: I1211 02:02:08.532058 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:08Z","lastTransitionTime":"2025-12-11T02:02:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:08 crc kubenswrapper[4824]: I1211 02:02:08.637784 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:08 crc kubenswrapper[4824]: I1211 02:02:08.637813 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:08 crc kubenswrapper[4824]: I1211 02:02:08.637820 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:08 crc kubenswrapper[4824]: I1211 02:02:08.637836 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:08 crc kubenswrapper[4824]: I1211 02:02:08.637845 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:08Z","lastTransitionTime":"2025-12-11T02:02:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:08 crc kubenswrapper[4824]: I1211 02:02:08.648855 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76b10bbd1e891b8c682f4b2e17d023fd787a8ed44ed94fbf7d97314d641ccf36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f32838f33832b9b67e7dd9c53340680e17721cf1c7bdf0170230c5f8feafe64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:08Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:08 crc kubenswrapper[4824]: I1211 02:02:08.664043 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kptws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68568dba-83a5-46c1-8d01-337359679d91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f965a6d6db619995132f61b5d51b66f0b2b3fdbef3fbec63e231b9a62ccd0a9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd7nf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kptws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:08Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:08 crc kubenswrapper[4824]: I1211 02:02:08.685082 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"beab386b-6c65-4bc3-9ef0-dc3f68b4403f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c26f06d2ed5e6e98b61a51f073130675f2c49acda106b8aedcbb2a6d4f7580c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20af9b10a67c8b21a7abcd23e42bc412798d159c8422e7c8ff58f5b997ca19a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9bbfc6edfe050b3fa9dbeee5f98d416bcb2a1533a38608a881b75396fd0b059\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7c766f8de73ca762ffde0c18b2d1c46bb152b2b37535a611c5a0e454a21b7d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d45819f0611c91b20c83eeee1092031a3441fb69f3015241632ad419880227d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 02:01:11.070250 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 02:01:11.071673 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2752882565/tls.crt::/tmp/serving-cert-2752882565/tls.key\\\\\\\"\\\\nI1211 02:01:17.042610 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 02:01:17.048046 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 02:01:17.048080 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 02:01:17.048172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 02:01:17.048188 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 02:01:17.058248 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 02:01:17.058294 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 02:01:17.058303 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 02:01:17.058315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 02:01:17.058322 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 02:01:17.058328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 02:01:17.058335 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 02:01:17.058808 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 02:01:17.060225 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14a940a72b2404a6cbd52a69e150b03266bcd7f5c1d3f4d07b51c0c69e51786\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:00:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:08Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:08 crc kubenswrapper[4824]: I1211 02:02:08.701865 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ded146c045085be2b58821749052413a36850ede00ae1024fe2dae9fa5ac84b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:08Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:08 crc kubenswrapper[4824]: I1211 02:02:08.729019 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b193bd2-b536-4056-92f1-94c9836ab2eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51f3c41d86ddc44f7eca118029876ad353317f753d0ad173ed4d30c2a15f10e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c345850891744a0240781277e9a89a59f6899c4c944995f62e44650f807440a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed2ae40b1fe28111efc76d9d72345edecb5f3ea34522b238f3730812b29f2bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce7aa078800e6f299cfe03770fa7e3b64fa3742fe0c0f383c87d84cf48569ddf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6aa9fc1f1bff192c69bd59d32a05d4a72d09641d9ce4b265bc4517147a7e15dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b4a3bc50eb80777188206adab774e8acad8b2e4952a8415f91b1c29ba9a0ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09bb8b4c85e1fd5fa8d62db23ebee77a25a0faa274da3bb22ab679987597f26b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09bb8b4c85e1fd5fa8d62db23ebee77a25a0faa274da3bb22ab679987597f26b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T02:01:43Z\\\",\\\"message\\\":\\\"ng]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-webhook\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.254\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1211 02:01:43.627342 6417 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-target-xd92c in node crc\\\\nI1211 02:01:43.627315 6417 services_controller.go:443] Built service openshift-machine-config-operator/machine-config-daemon LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.43\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:9001, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}, services.lbConfig{vips:[]string{\\\\\\\"10.217.4.43\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8798, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLo\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:42Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-f6pgc_openshift-ovn-kubernetes(5b193bd2-b536-4056-92f1-94c9836ab2eb)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ead58694813806d307bbbc6a7b3cf41730d1c059b231a3002684bf09c10c7f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f6pgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:08Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:08 crc kubenswrapper[4824]: I1211 02:02:08.750656 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-br4j9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce125194-9c9d-4ea4-8a72-2c30d5722295\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9798f904052e08196a0db1745060dab11e2710308bca01ff160475910f2665df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fc5s8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-br4j9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:08Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:08 crc kubenswrapper[4824]: I1211 02:02:08.751836 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:08 crc kubenswrapper[4824]: I1211 02:02:08.751942 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:08 crc kubenswrapper[4824]: I1211 02:02:08.752001 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:08 crc kubenswrapper[4824]: I1211 02:02:08.752072 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:08 crc kubenswrapper[4824]: I1211 02:02:08.752150 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:08Z","lastTransitionTime":"2025-12-11T02:02:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:08 crc kubenswrapper[4824]: I1211 02:02:08.767099 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8v4jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99a6bfee-afca-4726-a964-b95a948945a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b23b9d8f18db3eeadb60f83b36b05b1778cd0fa31669dad2680bb335ae004acf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4hk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3acc1940a0534ba0e6096175e41834fdd0fbcbb4c3fb6e9a38e76aedae55d2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4hk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-8v4jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:08Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:08 crc kubenswrapper[4824]: I1211 02:02:08.780437 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-grnpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09a52948-0404-45ff-85b1-51479268bf71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmmzm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmmzm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-grnpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:08Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:08 crc kubenswrapper[4824]: I1211 02:02:08.795554 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:08Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:08 crc kubenswrapper[4824]: I1211 02:02:08.814501 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kh8g9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"50c69d41-a65b-463e-a829-ea349bd2fe46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dca6be41bd7a3ce12ecd4567c868f696b65e8eb1bad436feda6e4f9bfb8e525\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69af1f33fd0bafc929b94e702b4f46c6c7d7a362d5552b121729c477892d91d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69af1f33fd0bafc929b94e702b4f46c6c7d7a362d5552b121729c477892d91d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd9675d0afded587bc660a3862fd0c5dd6885961afacda957aa9ee8501d68a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd9675d0afded587bc660a3862fd0c5dd6885961afacda957aa9ee8501d68a69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aac94149458bd6321be90b26b48ee26428d273e7f61407d95f9c9d6b3aeae4b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aac94149458bd6321be90b26b48ee26428d273e7f61407d95f9c9d6b3aeae4b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbee11bd7cb65e031415990869b52d2b8f81bf8b56e6cc935d38e09695db7b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbee11bd7cb65e031415990869b52d2b8f81bf8b56e6cc935d38e09695db7b48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86f693ade72bd593bc16021fe4b7564f1914ec2f1cda7039474ac2fb13de10f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86f693ade72bd593bc16021fe4b7564f1914ec2f1cda7039474ac2fb13de10f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7d7674252f67bde0ded2ec8ad1d925a92a26d3e3f741dee0061b735cb0b7a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7d7674252f67bde0ded2ec8ad1d925a92a26d3e3f741dee0061b735cb0b7a02\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kh8g9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:08Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:08 crc kubenswrapper[4824]: I1211 02:02:08.831207 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6bkc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8022f696-d0e7-437b-848a-3cd25bd1f364\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f40719f59881e077e7edbab3adb9efd101359bd461e5434c365e119e8cdbddb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3e28f6778ec0dba6ea3e6b69e5a2c1c73c0d5db1abeead241fb58efdbfaefe4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T02:02:03Z\\\",\\\"message\\\":\\\"2025-12-11T02:01:18+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_00fc4257-9dfb-4131-a927-0fc54bfa1304\\\\n2025-12-11T02:01:18+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_00fc4257-9dfb-4131-a927-0fc54bfa1304 to /host/opt/cni/bin/\\\\n2025-12-11T02:01:18Z [verbose] multus-daemon started\\\\n2025-12-11T02:01:18Z [verbose] Readiness Indicator file check\\\\n2025-12-11T02:02:03Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:02:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hh5kn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6bkc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:08Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:08 crc kubenswrapper[4824]: I1211 02:02:08.847642 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d14986666febd651039afb52ddf4db8969aca55e88eda9bbb20a67e35ba722ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:08Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:08 crc kubenswrapper[4824]: I1211 02:02:08.854381 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:08 crc kubenswrapper[4824]: I1211 02:02:08.854472 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:08 crc kubenswrapper[4824]: I1211 02:02:08.854533 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:08 crc kubenswrapper[4824]: I1211 02:02:08.854589 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:08 crc kubenswrapper[4824]: I1211 02:02:08.854649 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:08Z","lastTransitionTime":"2025-12-11T02:02:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:08 crc kubenswrapper[4824]: I1211 02:02:08.859997 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:08Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:08 crc kubenswrapper[4824]: I1211 02:02:08.874195 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44b156e4-64a4-4d45-aa5e-9b10a862faed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7368d92144b188e95a7796fabf497b0f8d9c0b240293bdc7b4aec1ff75497bc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w62dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcf289ac27dbda4f24aabbde3fb5119329e7b4fc7d6e018168da9a69661a4c57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w62dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gx6xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:08Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:08 crc kubenswrapper[4824]: I1211 02:02:08.887671 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:08Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:08 crc kubenswrapper[4824]: I1211 02:02:08.900511 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"253d4291-2b2f-487c-8862-3d305f0bdc55\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44b29bb7a6eda7cb890d36f86375b5f3f197ae5cf390f16e71139b0904873e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://922af7f67b99a0ce08776f3f75c0c377e6ea69af1720b6178a6d1e2f9b4b8ba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://136c68020325db8e6930d0f40b0d0d7e013c14c4dd5e7d1449d01d6d458104d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c89afaf56b4a9597f6543ec2e3b176c7a6dab7dc341941fc3bc8ec85cc4edaf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:00:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:08Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:08 crc kubenswrapper[4824]: I1211 02:02:08.912418 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98027cfa-5dde-4297-9f74-c4d662c7f82b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f410073059c5898660c50f775a270bebe51836f73e2a2f02db38a2aabf15931d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://335a774b8a3aff18695de4aba0b522d346aecc8ed8d96a80150301f5debce32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://339d619250c1c0a12f988a9b0ebc03874db3ac41a9d3b3f8573795467578cd0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e27af6c438009a44288f8bc10f8ebf2b61d9309f87e8d955e8d255a7190174f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e27af6c438009a44288f8bc10f8ebf2b61d9309f87e8d955e8d255a7190174f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:00:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:08Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:08 crc kubenswrapper[4824]: I1211 02:02:08.957916 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:08 crc kubenswrapper[4824]: I1211 02:02:08.957978 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:08 crc kubenswrapper[4824]: I1211 02:02:08.957996 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:08 crc kubenswrapper[4824]: I1211 02:02:08.958021 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:08 crc kubenswrapper[4824]: I1211 02:02:08.958036 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:08Z","lastTransitionTime":"2025-12-11T02:02:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:09 crc kubenswrapper[4824]: I1211 02:02:09.059905 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:09 crc kubenswrapper[4824]: I1211 02:02:09.059949 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:09 crc kubenswrapper[4824]: I1211 02:02:09.059959 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:09 crc kubenswrapper[4824]: I1211 02:02:09.059973 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:09 crc kubenswrapper[4824]: I1211 02:02:09.059982 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:09Z","lastTransitionTime":"2025-12-11T02:02:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:09 crc kubenswrapper[4824]: I1211 02:02:09.162063 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:09 crc kubenswrapper[4824]: I1211 02:02:09.162087 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:09 crc kubenswrapper[4824]: I1211 02:02:09.162095 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:09 crc kubenswrapper[4824]: I1211 02:02:09.162128 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:09 crc kubenswrapper[4824]: I1211 02:02:09.162137 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:09Z","lastTransitionTime":"2025-12-11T02:02:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:09 crc kubenswrapper[4824]: I1211 02:02:09.264505 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:09 crc kubenswrapper[4824]: I1211 02:02:09.264555 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:09 crc kubenswrapper[4824]: I1211 02:02:09.264570 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:09 crc kubenswrapper[4824]: I1211 02:02:09.264589 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:09 crc kubenswrapper[4824]: I1211 02:02:09.264603 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:09Z","lastTransitionTime":"2025-12-11T02:02:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:09 crc kubenswrapper[4824]: I1211 02:02:09.367312 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:09 crc kubenswrapper[4824]: I1211 02:02:09.367387 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:09 crc kubenswrapper[4824]: I1211 02:02:09.367401 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:09 crc kubenswrapper[4824]: I1211 02:02:09.367426 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:09 crc kubenswrapper[4824]: I1211 02:02:09.367439 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:09Z","lastTransitionTime":"2025-12-11T02:02:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:09 crc kubenswrapper[4824]: I1211 02:02:09.470399 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:09 crc kubenswrapper[4824]: I1211 02:02:09.470430 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:09 crc kubenswrapper[4824]: I1211 02:02:09.470464 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:09 crc kubenswrapper[4824]: I1211 02:02:09.470478 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:09 crc kubenswrapper[4824]: I1211 02:02:09.470487 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:09Z","lastTransitionTime":"2025-12-11T02:02:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:09 crc kubenswrapper[4824]: I1211 02:02:09.573194 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:09 crc kubenswrapper[4824]: I1211 02:02:09.573301 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:09 crc kubenswrapper[4824]: I1211 02:02:09.573327 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:09 crc kubenswrapper[4824]: I1211 02:02:09.573352 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:09 crc kubenswrapper[4824]: I1211 02:02:09.573369 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:09Z","lastTransitionTime":"2025-12-11T02:02:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:09 crc kubenswrapper[4824]: I1211 02:02:09.632398 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 02:02:09 crc kubenswrapper[4824]: E1211 02:02:09.632569 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 02:02:09 crc kubenswrapper[4824]: I1211 02:02:09.632645 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 02:02:09 crc kubenswrapper[4824]: I1211 02:02:09.632722 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 02:02:09 crc kubenswrapper[4824]: E1211 02:02:09.632767 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 02:02:09 crc kubenswrapper[4824]: I1211 02:02:09.633095 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-grnpt" Dec 11 02:02:09 crc kubenswrapper[4824]: E1211 02:02:09.633178 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 02:02:09 crc kubenswrapper[4824]: E1211 02:02:09.633306 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-grnpt" podUID="09a52948-0404-45ff-85b1-51479268bf71" Dec 11 02:02:09 crc kubenswrapper[4824]: I1211 02:02:09.633449 4824 scope.go:117] "RemoveContainer" containerID="09bb8b4c85e1fd5fa8d62db23ebee77a25a0faa274da3bb22ab679987597f26b" Dec 11 02:02:09 crc kubenswrapper[4824]: I1211 02:02:09.675628 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:09 crc kubenswrapper[4824]: I1211 02:02:09.675663 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:09 crc kubenswrapper[4824]: I1211 02:02:09.675677 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:09 crc kubenswrapper[4824]: I1211 02:02:09.675693 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:09 crc kubenswrapper[4824]: I1211 02:02:09.675705 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:09Z","lastTransitionTime":"2025-12-11T02:02:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:09 crc kubenswrapper[4824]: I1211 02:02:09.778348 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:09 crc kubenswrapper[4824]: I1211 02:02:09.778383 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:09 crc kubenswrapper[4824]: I1211 02:02:09.778392 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:09 crc kubenswrapper[4824]: I1211 02:02:09.778409 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:09 crc kubenswrapper[4824]: I1211 02:02:09.778420 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:09Z","lastTransitionTime":"2025-12-11T02:02:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:09 crc kubenswrapper[4824]: I1211 02:02:09.884616 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:09 crc kubenswrapper[4824]: I1211 02:02:09.884674 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:09 crc kubenswrapper[4824]: I1211 02:02:09.884691 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:09 crc kubenswrapper[4824]: I1211 02:02:09.884716 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:09 crc kubenswrapper[4824]: I1211 02:02:09.884733 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:09Z","lastTransitionTime":"2025-12-11T02:02:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:09 crc kubenswrapper[4824]: I1211 02:02:09.987053 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:09 crc kubenswrapper[4824]: I1211 02:02:09.987102 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:09 crc kubenswrapper[4824]: I1211 02:02:09.987133 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:09 crc kubenswrapper[4824]: I1211 02:02:09.987152 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:09 crc kubenswrapper[4824]: I1211 02:02:09.987163 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:09Z","lastTransitionTime":"2025-12-11T02:02:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:10 crc kubenswrapper[4824]: I1211 02:02:10.089619 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:10 crc kubenswrapper[4824]: I1211 02:02:10.089655 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:10 crc kubenswrapper[4824]: I1211 02:02:10.089666 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:10 crc kubenswrapper[4824]: I1211 02:02:10.089680 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:10 crc kubenswrapper[4824]: I1211 02:02:10.089690 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:10Z","lastTransitionTime":"2025-12-11T02:02:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:10 crc kubenswrapper[4824]: I1211 02:02:10.186863 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f6pgc_5b193bd2-b536-4056-92f1-94c9836ab2eb/ovnkube-controller/2.log" Dec 11 02:02:10 crc kubenswrapper[4824]: I1211 02:02:10.189506 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" event={"ID":"5b193bd2-b536-4056-92f1-94c9836ab2eb","Type":"ContainerStarted","Data":"154d82fd6acee749bfd37220547e52647dc7f9d86a73c28ce7d8577c9cadef6e"} Dec 11 02:02:10 crc kubenswrapper[4824]: I1211 02:02:10.189930 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" Dec 11 02:02:10 crc kubenswrapper[4824]: I1211 02:02:10.191638 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:10 crc kubenswrapper[4824]: I1211 02:02:10.191664 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:10 crc kubenswrapper[4824]: I1211 02:02:10.191673 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:10 crc kubenswrapper[4824]: I1211 02:02:10.191688 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:10 crc kubenswrapper[4824]: I1211 02:02:10.191698 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:10Z","lastTransitionTime":"2025-12-11T02:02:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:10 crc kubenswrapper[4824]: I1211 02:02:10.210682 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"beab386b-6c65-4bc3-9ef0-dc3f68b4403f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c26f06d2ed5e6e98b61a51f073130675f2c49acda106b8aedcbb2a6d4f7580c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20af9b10a67c8b21a7abcd23e42bc412798d159c8422e7c8ff58f5b997ca19a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9bbfc6edfe050b3fa9dbeee5f98d416bcb2a1533a38608a881b75396fd0b059\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7c766f8de73ca762ffde0c18b2d1c46bb152b2b37535a611c5a0e454a21b7d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d45819f0611c91b20c83eeee1092031a3441fb69f3015241632ad419880227d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 02:01:11.070250 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 02:01:11.071673 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2752882565/tls.crt::/tmp/serving-cert-2752882565/tls.key\\\\\\\"\\\\nI1211 02:01:17.042610 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 02:01:17.048046 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 02:01:17.048080 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 02:01:17.048172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 02:01:17.048188 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 02:01:17.058248 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 02:01:17.058294 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 02:01:17.058303 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 02:01:17.058315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 02:01:17.058322 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 02:01:17.058328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 02:01:17.058335 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 02:01:17.058808 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 02:01:17.060225 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14a940a72b2404a6cbd52a69e150b03266bcd7f5c1d3f4d07b51c0c69e51786\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:00:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:10Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:10 crc kubenswrapper[4824]: I1211 02:02:10.221743 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ded146c045085be2b58821749052413a36850ede00ae1024fe2dae9fa5ac84b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:10Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:10 crc kubenswrapper[4824]: I1211 02:02:10.243519 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76b10bbd1e891b8c682f4b2e17d023fd787a8ed44ed94fbf7d97314d641ccf36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f32838f33832b9b67e7dd9c53340680e17721cf1c7bdf0170230c5f8feafe64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:10Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:10 crc kubenswrapper[4824]: I1211 02:02:10.256476 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kptws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68568dba-83a5-46c1-8d01-337359679d91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f965a6d6db619995132f61b5d51b66f0b2b3fdbef3fbec63e231b9a62ccd0a9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd7nf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kptws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:10Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:10 crc kubenswrapper[4824]: I1211 02:02:10.269833 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:10Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:10 crc kubenswrapper[4824]: I1211 02:02:10.283421 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kh8g9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"50c69d41-a65b-463e-a829-ea349bd2fe46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dca6be41bd7a3ce12ecd4567c868f696b65e8eb1bad436feda6e4f9bfb8e525\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69af1f33fd0bafc929b94e702b4f46c6c7d7a362d5552b121729c477892d91d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69af1f33fd0bafc929b94e702b4f46c6c7d7a362d5552b121729c477892d91d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd9675d0afded587bc660a3862fd0c5dd6885961afacda957aa9ee8501d68a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd9675d0afded587bc660a3862fd0c5dd6885961afacda957aa9ee8501d68a69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aac94149458bd6321be90b26b48ee26428d273e7f61407d95f9c9d6b3aeae4b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aac94149458bd6321be90b26b48ee26428d273e7f61407d95f9c9d6b3aeae4b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbee11bd7cb65e031415990869b52d2b8f81bf8b56e6cc935d38e09695db7b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbee11bd7cb65e031415990869b52d2b8f81bf8b56e6cc935d38e09695db7b48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86f693ade72bd593bc16021fe4b7564f1914ec2f1cda7039474ac2fb13de10f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86f693ade72bd593bc16021fe4b7564f1914ec2f1cda7039474ac2fb13de10f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7d7674252f67bde0ded2ec8ad1d925a92a26d3e3f741dee0061b735cb0b7a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7d7674252f67bde0ded2ec8ad1d925a92a26d3e3f741dee0061b735cb0b7a02\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kh8g9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:10Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:10 crc kubenswrapper[4824]: I1211 02:02:10.295554 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:10 crc kubenswrapper[4824]: I1211 02:02:10.295592 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:10 crc kubenswrapper[4824]: I1211 02:02:10.295602 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:10 crc kubenswrapper[4824]: I1211 02:02:10.295616 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:10 crc kubenswrapper[4824]: I1211 02:02:10.295630 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:10Z","lastTransitionTime":"2025-12-11T02:02:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:10 crc kubenswrapper[4824]: I1211 02:02:10.299270 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6bkc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8022f696-d0e7-437b-848a-3cd25bd1f364\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f40719f59881e077e7edbab3adb9efd101359bd461e5434c365e119e8cdbddb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3e28f6778ec0dba6ea3e6b69e5a2c1c73c0d5db1abeead241fb58efdbfaefe4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T02:02:03Z\\\",\\\"message\\\":\\\"2025-12-11T02:01:18+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_00fc4257-9dfb-4131-a927-0fc54bfa1304\\\\n2025-12-11T02:01:18+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_00fc4257-9dfb-4131-a927-0fc54bfa1304 to /host/opt/cni/bin/\\\\n2025-12-11T02:01:18Z [verbose] multus-daemon started\\\\n2025-12-11T02:01:18Z [verbose] Readiness Indicator file check\\\\n2025-12-11T02:02:03Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:02:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hh5kn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6bkc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:10Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:10 crc kubenswrapper[4824]: I1211 02:02:10.314357 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b193bd2-b536-4056-92f1-94c9836ab2eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51f3c41d86ddc44f7eca118029876ad353317f753d0ad173ed4d30c2a15f10e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c345850891744a0240781277e9a89a59f6899c4c944995f62e44650f807440a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed2ae40b1fe28111efc76d9d72345edecb5f3ea34522b238f3730812b29f2bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce7aa078800e6f299cfe03770fa7e3b64fa3742fe0c0f383c87d84cf48569ddf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6aa9fc1f1bff192c69bd59d32a05d4a72d09641d9ce4b265bc4517147a7e15dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b4a3bc50eb80777188206adab774e8acad8b2e4952a8415f91b1c29ba9a0ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://154d82fd6acee749bfd37220547e52647dc7f9d86a73c28ce7d8577c9cadef6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09bb8b4c85e1fd5fa8d62db23ebee77a25a0faa274da3bb22ab679987597f26b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T02:01:43Z\\\",\\\"message\\\":\\\"ng]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-webhook\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.254\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1211 02:01:43.627342 6417 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-target-xd92c in node crc\\\\nI1211 02:01:43.627315 6417 services_controller.go:443] Built service openshift-machine-config-operator/machine-config-daemon LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.43\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:9001, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}, services.lbConfig{vips:[]string{\\\\\\\"10.217.4.43\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8798, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLo\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:42Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:02:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ead58694813806d307bbbc6a7b3cf41730d1c059b231a3002684bf09c10c7f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f6pgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:10Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:10 crc kubenswrapper[4824]: I1211 02:02:10.323445 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-br4j9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce125194-9c9d-4ea4-8a72-2c30d5722295\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9798f904052e08196a0db1745060dab11e2710308bca01ff160475910f2665df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fc5s8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-br4j9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:10Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:10 crc kubenswrapper[4824]: I1211 02:02:10.333958 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8v4jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99a6bfee-afca-4726-a964-b95a948945a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b23b9d8f18db3eeadb60f83b36b05b1778cd0fa31669dad2680bb335ae004acf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4hk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3acc1940a0534ba0e6096175e41834fdd0fbcbb4c3fb6e9a38e76aedae55d2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4hk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-8v4jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:10Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:10 crc kubenswrapper[4824]: I1211 02:02:10.344456 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-grnpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09a52948-0404-45ff-85b1-51479268bf71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmmzm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmmzm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-grnpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:10Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:10 crc kubenswrapper[4824]: I1211 02:02:10.355696 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d14986666febd651039afb52ddf4db8969aca55e88eda9bbb20a67e35ba722ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:10Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:10 crc kubenswrapper[4824]: I1211 02:02:10.367361 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:10Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:10 crc kubenswrapper[4824]: I1211 02:02:10.378252 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44b156e4-64a4-4d45-aa5e-9b10a862faed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7368d92144b188e95a7796fabf497b0f8d9c0b240293bdc7b4aec1ff75497bc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w62dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcf289ac27dbda4f24aabbde3fb5119329e7b4fc7d6e018168da9a69661a4c57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w62dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gx6xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:10Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:10 crc kubenswrapper[4824]: I1211 02:02:10.389279 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"253d4291-2b2f-487c-8862-3d305f0bdc55\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44b29bb7a6eda7cb890d36f86375b5f3f197ae5cf390f16e71139b0904873e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://922af7f67b99a0ce08776f3f75c0c377e6ea69af1720b6178a6d1e2f9b4b8ba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://136c68020325db8e6930d0f40b0d0d7e013c14c4dd5e7d1449d01d6d458104d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c89afaf56b4a9597f6543ec2e3b176c7a6dab7dc341941fc3bc8ec85cc4edaf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:00:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:10Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:10 crc kubenswrapper[4824]: I1211 02:02:10.398363 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:10 crc kubenswrapper[4824]: I1211 02:02:10.398406 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:10 crc kubenswrapper[4824]: I1211 02:02:10.398420 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:10 crc kubenswrapper[4824]: I1211 02:02:10.398436 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:10 crc kubenswrapper[4824]: I1211 02:02:10.398447 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:10Z","lastTransitionTime":"2025-12-11T02:02:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:10 crc kubenswrapper[4824]: I1211 02:02:10.404273 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98027cfa-5dde-4297-9f74-c4d662c7f82b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f410073059c5898660c50f775a270bebe51836f73e2a2f02db38a2aabf15931d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://335a774b8a3aff18695de4aba0b522d346aecc8ed8d96a80150301f5debce32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://339d619250c1c0a12f988a9b0ebc03874db3ac41a9d3b3f8573795467578cd0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e27af6c438009a44288f8bc10f8ebf2b61d9309f87e8d955e8d255a7190174f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e27af6c438009a44288f8bc10f8ebf2b61d9309f87e8d955e8d255a7190174f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:00:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:10Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:10 crc kubenswrapper[4824]: I1211 02:02:10.417785 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:10Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:10 crc kubenswrapper[4824]: I1211 02:02:10.501154 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:10 crc kubenswrapper[4824]: I1211 02:02:10.501210 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:10 crc kubenswrapper[4824]: I1211 02:02:10.501228 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:10 crc kubenswrapper[4824]: I1211 02:02:10.501252 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:10 crc kubenswrapper[4824]: I1211 02:02:10.501268 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:10Z","lastTransitionTime":"2025-12-11T02:02:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:10 crc kubenswrapper[4824]: I1211 02:02:10.603885 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:10 crc kubenswrapper[4824]: I1211 02:02:10.603924 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:10 crc kubenswrapper[4824]: I1211 02:02:10.603933 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:10 crc kubenswrapper[4824]: I1211 02:02:10.603948 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:10 crc kubenswrapper[4824]: I1211 02:02:10.603959 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:10Z","lastTransitionTime":"2025-12-11T02:02:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:10 crc kubenswrapper[4824]: I1211 02:02:10.706731 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:10 crc kubenswrapper[4824]: I1211 02:02:10.706775 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:10 crc kubenswrapper[4824]: I1211 02:02:10.706786 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:10 crc kubenswrapper[4824]: I1211 02:02:10.706803 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:10 crc kubenswrapper[4824]: I1211 02:02:10.706816 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:10Z","lastTransitionTime":"2025-12-11T02:02:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:10 crc kubenswrapper[4824]: I1211 02:02:10.809258 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:10 crc kubenswrapper[4824]: I1211 02:02:10.809305 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:10 crc kubenswrapper[4824]: I1211 02:02:10.809316 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:10 crc kubenswrapper[4824]: I1211 02:02:10.809334 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:10 crc kubenswrapper[4824]: I1211 02:02:10.809346 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:10Z","lastTransitionTime":"2025-12-11T02:02:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:10 crc kubenswrapper[4824]: I1211 02:02:10.911377 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:10 crc kubenswrapper[4824]: I1211 02:02:10.911418 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:10 crc kubenswrapper[4824]: I1211 02:02:10.911430 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:10 crc kubenswrapper[4824]: I1211 02:02:10.911445 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:10 crc kubenswrapper[4824]: I1211 02:02:10.911458 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:10Z","lastTransitionTime":"2025-12-11T02:02:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.013470 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.013509 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.013518 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.013535 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.013546 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:11Z","lastTransitionTime":"2025-12-11T02:02:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.116155 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.116209 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.116226 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.116250 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.116268 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:11Z","lastTransitionTime":"2025-12-11T02:02:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.195427 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f6pgc_5b193bd2-b536-4056-92f1-94c9836ab2eb/ovnkube-controller/3.log" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.196519 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f6pgc_5b193bd2-b536-4056-92f1-94c9836ab2eb/ovnkube-controller/2.log" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.199759 4824 generic.go:334] "Generic (PLEG): container finished" podID="5b193bd2-b536-4056-92f1-94c9836ab2eb" containerID="154d82fd6acee749bfd37220547e52647dc7f9d86a73c28ce7d8577c9cadef6e" exitCode=1 Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.199793 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" event={"ID":"5b193bd2-b536-4056-92f1-94c9836ab2eb","Type":"ContainerDied","Data":"154d82fd6acee749bfd37220547e52647dc7f9d86a73c28ce7d8577c9cadef6e"} Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.199845 4824 scope.go:117] "RemoveContainer" containerID="09bb8b4c85e1fd5fa8d62db23ebee77a25a0faa274da3bb22ab679987597f26b" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.205048 4824 scope.go:117] "RemoveContainer" containerID="154d82fd6acee749bfd37220547e52647dc7f9d86a73c28ce7d8577c9cadef6e" Dec 11 02:02:11 crc kubenswrapper[4824]: E1211 02:02:11.206162 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-f6pgc_openshift-ovn-kubernetes(5b193bd2-b536-4056-92f1-94c9836ab2eb)\"" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" podUID="5b193bd2-b536-4056-92f1-94c9836ab2eb" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.217831 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.217852 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.217859 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.217872 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.217881 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:11Z","lastTransitionTime":"2025-12-11T02:02:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.222488 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d14986666febd651039afb52ddf4db8969aca55e88eda9bbb20a67e35ba722ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:11Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.239885 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:11Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.255486 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44b156e4-64a4-4d45-aa5e-9b10a862faed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7368d92144b188e95a7796fabf497b0f8d9c0b240293bdc7b4aec1ff75497bc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w62dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcf289ac27dbda4f24aabbde3fb5119329e7b4fc7d6e018168da9a69661a4c57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w62dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gx6xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:11Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.271889 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"253d4291-2b2f-487c-8862-3d305f0bdc55\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44b29bb7a6eda7cb890d36f86375b5f3f197ae5cf390f16e71139b0904873e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://922af7f67b99a0ce08776f3f75c0c377e6ea69af1720b6178a6d1e2f9b4b8ba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://136c68020325db8e6930d0f40b0d0d7e013c14c4dd5e7d1449d01d6d458104d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c89afaf56b4a9597f6543ec2e3b176c7a6dab7dc341941fc3bc8ec85cc4edaf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:00:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:11Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.283630 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98027cfa-5dde-4297-9f74-c4d662c7f82b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f410073059c5898660c50f775a270bebe51836f73e2a2f02db38a2aabf15931d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://335a774b8a3aff18695de4aba0b522d346aecc8ed8d96a80150301f5debce32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://339d619250c1c0a12f988a9b0ebc03874db3ac41a9d3b3f8573795467578cd0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e27af6c438009a44288f8bc10f8ebf2b61d9309f87e8d955e8d255a7190174f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e27af6c438009a44288f8bc10f8ebf2b61d9309f87e8d955e8d255a7190174f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:00:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:11Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.299352 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:11Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.318795 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"beab386b-6c65-4bc3-9ef0-dc3f68b4403f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c26f06d2ed5e6e98b61a51f073130675f2c49acda106b8aedcbb2a6d4f7580c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20af9b10a67c8b21a7abcd23e42bc412798d159c8422e7c8ff58f5b997ca19a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9bbfc6edfe050b3fa9dbeee5f98d416bcb2a1533a38608a881b75396fd0b059\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7c766f8de73ca762ffde0c18b2d1c46bb152b2b37535a611c5a0e454a21b7d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d45819f0611c91b20c83eeee1092031a3441fb69f3015241632ad419880227d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 02:01:11.070250 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 02:01:11.071673 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2752882565/tls.crt::/tmp/serving-cert-2752882565/tls.key\\\\\\\"\\\\nI1211 02:01:17.042610 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 02:01:17.048046 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 02:01:17.048080 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 02:01:17.048172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 02:01:17.048188 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 02:01:17.058248 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 02:01:17.058294 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 02:01:17.058303 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 02:01:17.058315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 02:01:17.058322 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 02:01:17.058328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 02:01:17.058335 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 02:01:17.058808 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 02:01:17.060225 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14a940a72b2404a6cbd52a69e150b03266bcd7f5c1d3f4d07b51c0c69e51786\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:00:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:11Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.320838 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.320904 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.320926 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.320949 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.320966 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:11Z","lastTransitionTime":"2025-12-11T02:02:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.337717 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ded146c045085be2b58821749052413a36850ede00ae1024fe2dae9fa5ac84b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:11Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.355469 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76b10bbd1e891b8c682f4b2e17d023fd787a8ed44ed94fbf7d97314d641ccf36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f32838f33832b9b67e7dd9c53340680e17721cf1c7bdf0170230c5f8feafe64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:11Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.369435 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kptws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68568dba-83a5-46c1-8d01-337359679d91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f965a6d6db619995132f61b5d51b66f0b2b3fdbef3fbec63e231b9a62ccd0a9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd7nf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kptws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:11Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.387572 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8v4jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99a6bfee-afca-4726-a964-b95a948945a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b23b9d8f18db3eeadb60f83b36b05b1778cd0fa31669dad2680bb335ae004acf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4hk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3acc1940a0534ba0e6096175e41834fdd0fbcbb4c3fb6e9a38e76aedae55d2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4hk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-8v4jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:11Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.402606 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-grnpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09a52948-0404-45ff-85b1-51479268bf71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmmzm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmmzm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-grnpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:11Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.419498 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:11Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.423699 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.423756 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.423774 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.423797 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.423816 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:11Z","lastTransitionTime":"2025-12-11T02:02:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.444030 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kh8g9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"50c69d41-a65b-463e-a829-ea349bd2fe46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dca6be41bd7a3ce12ecd4567c868f696b65e8eb1bad436feda6e4f9bfb8e525\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69af1f33fd0bafc929b94e702b4f46c6c7d7a362d5552b121729c477892d91d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69af1f33fd0bafc929b94e702b4f46c6c7d7a362d5552b121729c477892d91d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd9675d0afded587bc660a3862fd0c5dd6885961afacda957aa9ee8501d68a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd9675d0afded587bc660a3862fd0c5dd6885961afacda957aa9ee8501d68a69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aac94149458bd6321be90b26b48ee26428d273e7f61407d95f9c9d6b3aeae4b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aac94149458bd6321be90b26b48ee26428d273e7f61407d95f9c9d6b3aeae4b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbee11bd7cb65e031415990869b52d2b8f81bf8b56e6cc935d38e09695db7b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbee11bd7cb65e031415990869b52d2b8f81bf8b56e6cc935d38e09695db7b48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86f693ade72bd593bc16021fe4b7564f1914ec2f1cda7039474ac2fb13de10f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86f693ade72bd593bc16021fe4b7564f1914ec2f1cda7039474ac2fb13de10f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7d7674252f67bde0ded2ec8ad1d925a92a26d3e3f741dee0061b735cb0b7a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7d7674252f67bde0ded2ec8ad1d925a92a26d3e3f741dee0061b735cb0b7a02\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kh8g9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:11Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.461366 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6bkc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8022f696-d0e7-437b-848a-3cd25bd1f364\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f40719f59881e077e7edbab3adb9efd101359bd461e5434c365e119e8cdbddb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3e28f6778ec0dba6ea3e6b69e5a2c1c73c0d5db1abeead241fb58efdbfaefe4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T02:02:03Z\\\",\\\"message\\\":\\\"2025-12-11T02:01:18+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_00fc4257-9dfb-4131-a927-0fc54bfa1304\\\\n2025-12-11T02:01:18+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_00fc4257-9dfb-4131-a927-0fc54bfa1304 to /host/opt/cni/bin/\\\\n2025-12-11T02:01:18Z [verbose] multus-daemon started\\\\n2025-12-11T02:01:18Z [verbose] Readiness Indicator file check\\\\n2025-12-11T02:02:03Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:02:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hh5kn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6bkc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:11Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.492973 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b193bd2-b536-4056-92f1-94c9836ab2eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51f3c41d86ddc44f7eca118029876ad353317f753d0ad173ed4d30c2a15f10e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c345850891744a0240781277e9a89a59f6899c4c944995f62e44650f807440a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed2ae40b1fe28111efc76d9d72345edecb5f3ea34522b238f3730812b29f2bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce7aa078800e6f299cfe03770fa7e3b64fa3742fe0c0f383c87d84cf48569ddf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6aa9fc1f1bff192c69bd59d32a05d4a72d09641d9ce4b265bc4517147a7e15dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b4a3bc50eb80777188206adab774e8acad8b2e4952a8415f91b1c29ba9a0ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://154d82fd6acee749bfd37220547e52647dc7f9d86a73c28ce7d8577c9cadef6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09bb8b4c85e1fd5fa8d62db23ebee77a25a0faa274da3bb22ab679987597f26b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T02:01:43Z\\\",\\\"message\\\":\\\"ng]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-webhook\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.254\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1211 02:01:43.627342 6417 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-target-xd92c in node crc\\\\nI1211 02:01:43.627315 6417 services_controller.go:443] Built service openshift-machine-config-operator/machine-config-daemon LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.43\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:9001, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}, services.lbConfig{vips:[]string{\\\\\\\"10.217.4.43\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8798, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLo\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:42Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://154d82fd6acee749bfd37220547e52647dc7f9d86a73c28ce7d8577c9cadef6e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T02:02:10Z\\\",\\\"message\\\":\\\"t handler 9 for removal\\\\nI1211 02:02:10.678557 6803 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1211 02:02:10.678594 6803 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1211 02:02:10.678724 6803 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1211 02:02:10.678871 6803 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1211 02:02:10.678731 6803 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1211 02:02:10.678906 6803 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1211 02:02:10.678920 6803 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1211 02:02:10.678935 6803 handler.go:208] Removed *v1.Node event handler 7\\\\nI1211 02:02:10.678961 6803 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1211 02:02:10.678945 6803 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1211 02:02:10.678980 6803 handler.go:208] Removed *v1.Node event handler 2\\\\nI1211 02:02:10.678982 6803 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1211 02:02:10.679043 6803 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1211 02:02:10.679161 6803 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1211 02:02:10.679264 6803 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T02:02:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ead58694813806d307bbbc6a7b3cf41730d1c059b231a3002684bf09c10c7f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f6pgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:11Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.502245 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-br4j9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce125194-9c9d-4ea4-8a72-2c30d5722295\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9798f904052e08196a0db1745060dab11e2710308bca01ff160475910f2665df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fc5s8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-br4j9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:11Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.526566 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.526605 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.526613 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.526628 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.526637 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:11Z","lastTransitionTime":"2025-12-11T02:02:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.629102 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.629208 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.629228 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.629254 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.629272 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:11Z","lastTransitionTime":"2025-12-11T02:02:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.632420 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-grnpt" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.632467 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 02:02:11 crc kubenswrapper[4824]: E1211 02:02:11.632563 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-grnpt" podUID="09a52948-0404-45ff-85b1-51479268bf71" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.632617 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.632638 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 02:02:11 crc kubenswrapper[4824]: E1211 02:02:11.632782 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 02:02:11 crc kubenswrapper[4824]: E1211 02:02:11.632905 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 02:02:11 crc kubenswrapper[4824]: E1211 02:02:11.633183 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.732584 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.732653 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.732665 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.732688 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.732702 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:11Z","lastTransitionTime":"2025-12-11T02:02:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.835751 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.835793 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.835805 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.835821 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.835832 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:11Z","lastTransitionTime":"2025-12-11T02:02:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.836833 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.836894 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.836913 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.836936 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.836953 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:11Z","lastTransitionTime":"2025-12-11T02:02:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:11 crc kubenswrapper[4824]: E1211 02:02:11.850677 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:02:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:02:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:02:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:02:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d57bece6-6fc6-4a84-b8a8-e1678f9e50db\\\",\\\"systemUUID\\\":\\\"f76b32e4-2764-4219-bbf4-f7e56d76df3b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:11Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.854505 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.854549 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.854560 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.854576 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.854588 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:11Z","lastTransitionTime":"2025-12-11T02:02:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:11 crc kubenswrapper[4824]: E1211 02:02:11.867802 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:02:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:02:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:02:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:02:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d57bece6-6fc6-4a84-b8a8-e1678f9e50db\\\",\\\"systemUUID\\\":\\\"f76b32e4-2764-4219-bbf4-f7e56d76df3b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:11Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.871656 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.871703 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.871719 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.871738 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.871752 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:11Z","lastTransitionTime":"2025-12-11T02:02:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:11 crc kubenswrapper[4824]: E1211 02:02:11.884927 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:02:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:02:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:02:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:02:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d57bece6-6fc6-4a84-b8a8-e1678f9e50db\\\",\\\"systemUUID\\\":\\\"f76b32e4-2764-4219-bbf4-f7e56d76df3b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:11Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.888077 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.888141 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.888155 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.888174 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.888186 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:11Z","lastTransitionTime":"2025-12-11T02:02:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:11 crc kubenswrapper[4824]: E1211 02:02:11.901175 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:02:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:02:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:02:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:02:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d57bece6-6fc6-4a84-b8a8-e1678f9e50db\\\",\\\"systemUUID\\\":\\\"f76b32e4-2764-4219-bbf4-f7e56d76df3b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:11Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.905539 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.905564 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.905589 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.905604 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.905613 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:11Z","lastTransitionTime":"2025-12-11T02:02:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:11 crc kubenswrapper[4824]: E1211 02:02:11.917778 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:02:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:02:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:02:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:02:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d57bece6-6fc6-4a84-b8a8-e1678f9e50db\\\",\\\"systemUUID\\\":\\\"f76b32e4-2764-4219-bbf4-f7e56d76df3b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:11Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:11 crc kubenswrapper[4824]: E1211 02:02:11.917995 4824 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.938302 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.938363 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.938382 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.938407 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:11 crc kubenswrapper[4824]: I1211 02:02:11.938426 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:11Z","lastTransitionTime":"2025-12-11T02:02:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:12 crc kubenswrapper[4824]: I1211 02:02:12.041300 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:12 crc kubenswrapper[4824]: I1211 02:02:12.041385 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:12 crc kubenswrapper[4824]: I1211 02:02:12.041409 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:12 crc kubenswrapper[4824]: I1211 02:02:12.041445 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:12 crc kubenswrapper[4824]: I1211 02:02:12.041470 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:12Z","lastTransitionTime":"2025-12-11T02:02:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:12 crc kubenswrapper[4824]: I1211 02:02:12.144213 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:12 crc kubenswrapper[4824]: I1211 02:02:12.144278 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:12 crc kubenswrapper[4824]: I1211 02:02:12.144294 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:12 crc kubenswrapper[4824]: I1211 02:02:12.144316 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:12 crc kubenswrapper[4824]: I1211 02:02:12.144332 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:12Z","lastTransitionTime":"2025-12-11T02:02:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:12 crc kubenswrapper[4824]: I1211 02:02:12.205586 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f6pgc_5b193bd2-b536-4056-92f1-94c9836ab2eb/ovnkube-controller/3.log" Dec 11 02:02:12 crc kubenswrapper[4824]: I1211 02:02:12.210264 4824 scope.go:117] "RemoveContainer" containerID="154d82fd6acee749bfd37220547e52647dc7f9d86a73c28ce7d8577c9cadef6e" Dec 11 02:02:12 crc kubenswrapper[4824]: E1211 02:02:12.210552 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-f6pgc_openshift-ovn-kubernetes(5b193bd2-b536-4056-92f1-94c9836ab2eb)\"" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" podUID="5b193bd2-b536-4056-92f1-94c9836ab2eb" Dec 11 02:02:12 crc kubenswrapper[4824]: I1211 02:02:12.227376 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d14986666febd651039afb52ddf4db8969aca55e88eda9bbb20a67e35ba722ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:12Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:12 crc kubenswrapper[4824]: I1211 02:02:12.245636 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:12Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:12 crc kubenswrapper[4824]: I1211 02:02:12.248198 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:12 crc kubenswrapper[4824]: I1211 02:02:12.248248 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:12 crc kubenswrapper[4824]: I1211 02:02:12.248265 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:12 crc kubenswrapper[4824]: I1211 02:02:12.248291 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:12 crc kubenswrapper[4824]: I1211 02:02:12.248309 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:12Z","lastTransitionTime":"2025-12-11T02:02:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:12 crc kubenswrapper[4824]: I1211 02:02:12.260994 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44b156e4-64a4-4d45-aa5e-9b10a862faed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7368d92144b188e95a7796fabf497b0f8d9c0b240293bdc7b4aec1ff75497bc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w62dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcf289ac27dbda4f24aabbde3fb5119329e7b4fc7d6e018168da9a69661a4c57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w62dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gx6xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:12Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:12 crc kubenswrapper[4824]: I1211 02:02:12.278904 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"253d4291-2b2f-487c-8862-3d305f0bdc55\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44b29bb7a6eda7cb890d36f86375b5f3f197ae5cf390f16e71139b0904873e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://922af7f67b99a0ce08776f3f75c0c377e6ea69af1720b6178a6d1e2f9b4b8ba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://136c68020325db8e6930d0f40b0d0d7e013c14c4dd5e7d1449d01d6d458104d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c89afaf56b4a9597f6543ec2e3b176c7a6dab7dc341941fc3bc8ec85cc4edaf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:00:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:12Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:12 crc kubenswrapper[4824]: I1211 02:02:12.294689 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98027cfa-5dde-4297-9f74-c4d662c7f82b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f410073059c5898660c50f775a270bebe51836f73e2a2f02db38a2aabf15931d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://335a774b8a3aff18695de4aba0b522d346aecc8ed8d96a80150301f5debce32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://339d619250c1c0a12f988a9b0ebc03874db3ac41a9d3b3f8573795467578cd0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e27af6c438009a44288f8bc10f8ebf2b61d9309f87e8d955e8d255a7190174f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e27af6c438009a44288f8bc10f8ebf2b61d9309f87e8d955e8d255a7190174f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:00:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:12Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:12 crc kubenswrapper[4824]: I1211 02:02:12.308270 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:12Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:12 crc kubenswrapper[4824]: I1211 02:02:12.327988 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"beab386b-6c65-4bc3-9ef0-dc3f68b4403f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c26f06d2ed5e6e98b61a51f073130675f2c49acda106b8aedcbb2a6d4f7580c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20af9b10a67c8b21a7abcd23e42bc412798d159c8422e7c8ff58f5b997ca19a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9bbfc6edfe050b3fa9dbeee5f98d416bcb2a1533a38608a881b75396fd0b059\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7c766f8de73ca762ffde0c18b2d1c46bb152b2b37535a611c5a0e454a21b7d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d45819f0611c91b20c83eeee1092031a3441fb69f3015241632ad419880227d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 02:01:11.070250 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 02:01:11.071673 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2752882565/tls.crt::/tmp/serving-cert-2752882565/tls.key\\\\\\\"\\\\nI1211 02:01:17.042610 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 02:01:17.048046 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 02:01:17.048080 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 02:01:17.048172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 02:01:17.048188 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 02:01:17.058248 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 02:01:17.058294 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 02:01:17.058303 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 02:01:17.058315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 02:01:17.058322 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 02:01:17.058328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 02:01:17.058335 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 02:01:17.058808 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 02:01:17.060225 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14a940a72b2404a6cbd52a69e150b03266bcd7f5c1d3f4d07b51c0c69e51786\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:00:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:12Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:12 crc kubenswrapper[4824]: I1211 02:02:12.339690 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ded146c045085be2b58821749052413a36850ede00ae1024fe2dae9fa5ac84b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:12Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:12 crc kubenswrapper[4824]: I1211 02:02:12.351370 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:12 crc kubenswrapper[4824]: I1211 02:02:12.351411 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:12 crc kubenswrapper[4824]: I1211 02:02:12.351420 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:12 crc kubenswrapper[4824]: I1211 02:02:12.351438 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:12 crc kubenswrapper[4824]: I1211 02:02:12.351447 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:12Z","lastTransitionTime":"2025-12-11T02:02:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:12 crc kubenswrapper[4824]: I1211 02:02:12.354820 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76b10bbd1e891b8c682f4b2e17d023fd787a8ed44ed94fbf7d97314d641ccf36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f32838f33832b9b67e7dd9c53340680e17721cf1c7bdf0170230c5f8feafe64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:12Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:12 crc kubenswrapper[4824]: I1211 02:02:12.364948 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kptws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68568dba-83a5-46c1-8d01-337359679d91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f965a6d6db619995132f61b5d51b66f0b2b3fdbef3fbec63e231b9a62ccd0a9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd7nf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kptws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:12Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:12 crc kubenswrapper[4824]: I1211 02:02:12.375485 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-grnpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09a52948-0404-45ff-85b1-51479268bf71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmmzm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmmzm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-grnpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:12Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:12 crc kubenswrapper[4824]: I1211 02:02:12.389777 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:12Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:12 crc kubenswrapper[4824]: I1211 02:02:12.412169 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kh8g9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"50c69d41-a65b-463e-a829-ea349bd2fe46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dca6be41bd7a3ce12ecd4567c868f696b65e8eb1bad436feda6e4f9bfb8e525\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69af1f33fd0bafc929b94e702b4f46c6c7d7a362d5552b121729c477892d91d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69af1f33fd0bafc929b94e702b4f46c6c7d7a362d5552b121729c477892d91d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd9675d0afded587bc660a3862fd0c5dd6885961afacda957aa9ee8501d68a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd9675d0afded587bc660a3862fd0c5dd6885961afacda957aa9ee8501d68a69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aac94149458bd6321be90b26b48ee26428d273e7f61407d95f9c9d6b3aeae4b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aac94149458bd6321be90b26b48ee26428d273e7f61407d95f9c9d6b3aeae4b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbee11bd7cb65e031415990869b52d2b8f81bf8b56e6cc935d38e09695db7b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbee11bd7cb65e031415990869b52d2b8f81bf8b56e6cc935d38e09695db7b48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86f693ade72bd593bc16021fe4b7564f1914ec2f1cda7039474ac2fb13de10f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86f693ade72bd593bc16021fe4b7564f1914ec2f1cda7039474ac2fb13de10f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7d7674252f67bde0ded2ec8ad1d925a92a26d3e3f741dee0061b735cb0b7a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7d7674252f67bde0ded2ec8ad1d925a92a26d3e3f741dee0061b735cb0b7a02\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kh8g9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:12Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:12 crc kubenswrapper[4824]: I1211 02:02:12.427826 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6bkc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8022f696-d0e7-437b-848a-3cd25bd1f364\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f40719f59881e077e7edbab3adb9efd101359bd461e5434c365e119e8cdbddb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3e28f6778ec0dba6ea3e6b69e5a2c1c73c0d5db1abeead241fb58efdbfaefe4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T02:02:03Z\\\",\\\"message\\\":\\\"2025-12-11T02:01:18+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_00fc4257-9dfb-4131-a927-0fc54bfa1304\\\\n2025-12-11T02:01:18+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_00fc4257-9dfb-4131-a927-0fc54bfa1304 to /host/opt/cni/bin/\\\\n2025-12-11T02:01:18Z [verbose] multus-daemon started\\\\n2025-12-11T02:01:18Z [verbose] Readiness Indicator file check\\\\n2025-12-11T02:02:03Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:02:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hh5kn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6bkc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:12Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:12 crc kubenswrapper[4824]: I1211 02:02:12.454571 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:12 crc kubenswrapper[4824]: I1211 02:02:12.454640 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:12 crc kubenswrapper[4824]: I1211 02:02:12.454663 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:12 crc kubenswrapper[4824]: I1211 02:02:12.454695 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:12 crc kubenswrapper[4824]: I1211 02:02:12.454718 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:12Z","lastTransitionTime":"2025-12-11T02:02:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:12 crc kubenswrapper[4824]: I1211 02:02:12.462566 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b193bd2-b536-4056-92f1-94c9836ab2eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51f3c41d86ddc44f7eca118029876ad353317f753d0ad173ed4d30c2a15f10e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c345850891744a0240781277e9a89a59f6899c4c944995f62e44650f807440a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed2ae40b1fe28111efc76d9d72345edecb5f3ea34522b238f3730812b29f2bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce7aa078800e6f299cfe03770fa7e3b64fa3742fe0c0f383c87d84cf48569ddf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6aa9fc1f1bff192c69bd59d32a05d4a72d09641d9ce4b265bc4517147a7e15dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b4a3bc50eb80777188206adab774e8acad8b2e4952a8415f91b1c29ba9a0ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://154d82fd6acee749bfd37220547e52647dc7f9d86a73c28ce7d8577c9cadef6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://154d82fd6acee749bfd37220547e52647dc7f9d86a73c28ce7d8577c9cadef6e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T02:02:10Z\\\",\\\"message\\\":\\\"t handler 9 for removal\\\\nI1211 02:02:10.678557 6803 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1211 02:02:10.678594 6803 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1211 02:02:10.678724 6803 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1211 02:02:10.678871 6803 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1211 02:02:10.678731 6803 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1211 02:02:10.678906 6803 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1211 02:02:10.678920 6803 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1211 02:02:10.678935 6803 handler.go:208] Removed *v1.Node event handler 7\\\\nI1211 02:02:10.678961 6803 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1211 02:02:10.678945 6803 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1211 02:02:10.678980 6803 handler.go:208] Removed *v1.Node event handler 2\\\\nI1211 02:02:10.678982 6803 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1211 02:02:10.679043 6803 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1211 02:02:10.679161 6803 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1211 02:02:10.679264 6803 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T02:02:09Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-f6pgc_openshift-ovn-kubernetes(5b193bd2-b536-4056-92f1-94c9836ab2eb)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ead58694813806d307bbbc6a7b3cf41730d1c059b231a3002684bf09c10c7f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f6pgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:12Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:12 crc kubenswrapper[4824]: I1211 02:02:12.479782 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-br4j9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce125194-9c9d-4ea4-8a72-2c30d5722295\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9798f904052e08196a0db1745060dab11e2710308bca01ff160475910f2665df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fc5s8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-br4j9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:12Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:12 crc kubenswrapper[4824]: I1211 02:02:12.496831 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8v4jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99a6bfee-afca-4726-a964-b95a948945a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b23b9d8f18db3eeadb60f83b36b05b1778cd0fa31669dad2680bb335ae004acf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4hk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3acc1940a0534ba0e6096175e41834fdd0fbcbb4c3fb6e9a38e76aedae55d2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4hk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-8v4jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:12Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:12 crc kubenswrapper[4824]: I1211 02:02:12.557316 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:12 crc kubenswrapper[4824]: I1211 02:02:12.557374 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:12 crc kubenswrapper[4824]: I1211 02:02:12.557392 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:12 crc kubenswrapper[4824]: I1211 02:02:12.557415 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:12 crc kubenswrapper[4824]: I1211 02:02:12.557433 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:12Z","lastTransitionTime":"2025-12-11T02:02:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:12 crc kubenswrapper[4824]: I1211 02:02:12.660094 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:12 crc kubenswrapper[4824]: I1211 02:02:12.660154 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:12 crc kubenswrapper[4824]: I1211 02:02:12.660168 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:12 crc kubenswrapper[4824]: I1211 02:02:12.660185 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:12 crc kubenswrapper[4824]: I1211 02:02:12.660197 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:12Z","lastTransitionTime":"2025-12-11T02:02:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:12 crc kubenswrapper[4824]: I1211 02:02:12.762913 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:12 crc kubenswrapper[4824]: I1211 02:02:12.762996 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:12 crc kubenswrapper[4824]: I1211 02:02:12.763024 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:12 crc kubenswrapper[4824]: I1211 02:02:12.763056 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:12 crc kubenswrapper[4824]: I1211 02:02:12.763075 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:12Z","lastTransitionTime":"2025-12-11T02:02:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:12 crc kubenswrapper[4824]: I1211 02:02:12.866568 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:12 crc kubenswrapper[4824]: I1211 02:02:12.866661 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:12 crc kubenswrapper[4824]: I1211 02:02:12.866688 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:12 crc kubenswrapper[4824]: I1211 02:02:12.866720 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:12 crc kubenswrapper[4824]: I1211 02:02:12.866738 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:12Z","lastTransitionTime":"2025-12-11T02:02:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:12 crc kubenswrapper[4824]: I1211 02:02:12.969631 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:12 crc kubenswrapper[4824]: I1211 02:02:12.969679 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:12 crc kubenswrapper[4824]: I1211 02:02:12.969697 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:12 crc kubenswrapper[4824]: I1211 02:02:12.969720 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:12 crc kubenswrapper[4824]: I1211 02:02:12.969737 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:12Z","lastTransitionTime":"2025-12-11T02:02:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:13 crc kubenswrapper[4824]: I1211 02:02:13.072479 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:13 crc kubenswrapper[4824]: I1211 02:02:13.072526 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:13 crc kubenswrapper[4824]: I1211 02:02:13.072544 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:13 crc kubenswrapper[4824]: I1211 02:02:13.072566 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:13 crc kubenswrapper[4824]: I1211 02:02:13.072582 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:13Z","lastTransitionTime":"2025-12-11T02:02:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:13 crc kubenswrapper[4824]: I1211 02:02:13.176556 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:13 crc kubenswrapper[4824]: I1211 02:02:13.176631 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:13 crc kubenswrapper[4824]: I1211 02:02:13.176744 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:13 crc kubenswrapper[4824]: I1211 02:02:13.176776 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:13 crc kubenswrapper[4824]: I1211 02:02:13.176803 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:13Z","lastTransitionTime":"2025-12-11T02:02:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:13 crc kubenswrapper[4824]: I1211 02:02:13.279556 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:13 crc kubenswrapper[4824]: I1211 02:02:13.279619 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:13 crc kubenswrapper[4824]: I1211 02:02:13.279636 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:13 crc kubenswrapper[4824]: I1211 02:02:13.279661 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:13 crc kubenswrapper[4824]: I1211 02:02:13.279683 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:13Z","lastTransitionTime":"2025-12-11T02:02:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:13 crc kubenswrapper[4824]: I1211 02:02:13.382381 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:13 crc kubenswrapper[4824]: I1211 02:02:13.382439 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:13 crc kubenswrapper[4824]: I1211 02:02:13.382456 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:13 crc kubenswrapper[4824]: I1211 02:02:13.382480 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:13 crc kubenswrapper[4824]: I1211 02:02:13.382499 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:13Z","lastTransitionTime":"2025-12-11T02:02:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:13 crc kubenswrapper[4824]: I1211 02:02:13.485249 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:13 crc kubenswrapper[4824]: I1211 02:02:13.485300 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:13 crc kubenswrapper[4824]: I1211 02:02:13.485317 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:13 crc kubenswrapper[4824]: I1211 02:02:13.485344 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:13 crc kubenswrapper[4824]: I1211 02:02:13.485363 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:13Z","lastTransitionTime":"2025-12-11T02:02:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:13 crc kubenswrapper[4824]: I1211 02:02:13.588672 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:13 crc kubenswrapper[4824]: I1211 02:02:13.588754 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:13 crc kubenswrapper[4824]: I1211 02:02:13.588778 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:13 crc kubenswrapper[4824]: I1211 02:02:13.588808 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:13 crc kubenswrapper[4824]: I1211 02:02:13.588832 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:13Z","lastTransitionTime":"2025-12-11T02:02:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:13 crc kubenswrapper[4824]: I1211 02:02:13.632461 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 02:02:13 crc kubenswrapper[4824]: I1211 02:02:13.632496 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-grnpt" Dec 11 02:02:13 crc kubenswrapper[4824]: I1211 02:02:13.632554 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 02:02:13 crc kubenswrapper[4824]: I1211 02:02:13.632609 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 02:02:13 crc kubenswrapper[4824]: E1211 02:02:13.632787 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 02:02:13 crc kubenswrapper[4824]: E1211 02:02:13.632945 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 02:02:13 crc kubenswrapper[4824]: E1211 02:02:13.633046 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-grnpt" podUID="09a52948-0404-45ff-85b1-51479268bf71" Dec 11 02:02:13 crc kubenswrapper[4824]: E1211 02:02:13.633208 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 02:02:13 crc kubenswrapper[4824]: I1211 02:02:13.691867 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:13 crc kubenswrapper[4824]: I1211 02:02:13.691935 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:13 crc kubenswrapper[4824]: I1211 02:02:13.691952 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:13 crc kubenswrapper[4824]: I1211 02:02:13.691978 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:13 crc kubenswrapper[4824]: I1211 02:02:13.691996 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:13Z","lastTransitionTime":"2025-12-11T02:02:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:13 crc kubenswrapper[4824]: I1211 02:02:13.794400 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:13 crc kubenswrapper[4824]: I1211 02:02:13.794479 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:13 crc kubenswrapper[4824]: I1211 02:02:13.794506 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:13 crc kubenswrapper[4824]: I1211 02:02:13.794538 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:13 crc kubenswrapper[4824]: I1211 02:02:13.794564 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:13Z","lastTransitionTime":"2025-12-11T02:02:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:13 crc kubenswrapper[4824]: I1211 02:02:13.897883 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:13 crc kubenswrapper[4824]: I1211 02:02:13.897942 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:13 crc kubenswrapper[4824]: I1211 02:02:13.897959 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:13 crc kubenswrapper[4824]: I1211 02:02:13.897984 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:13 crc kubenswrapper[4824]: I1211 02:02:13.898002 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:13Z","lastTransitionTime":"2025-12-11T02:02:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:14 crc kubenswrapper[4824]: I1211 02:02:14.000460 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:14 crc kubenswrapper[4824]: I1211 02:02:14.000500 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:14 crc kubenswrapper[4824]: I1211 02:02:14.000510 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:14 crc kubenswrapper[4824]: I1211 02:02:14.000524 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:14 crc kubenswrapper[4824]: I1211 02:02:14.000535 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:14Z","lastTransitionTime":"2025-12-11T02:02:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:14 crc kubenswrapper[4824]: I1211 02:02:14.103286 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:14 crc kubenswrapper[4824]: I1211 02:02:14.103327 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:14 crc kubenswrapper[4824]: I1211 02:02:14.103336 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:14 crc kubenswrapper[4824]: I1211 02:02:14.103351 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:14 crc kubenswrapper[4824]: I1211 02:02:14.103363 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:14Z","lastTransitionTime":"2025-12-11T02:02:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:14 crc kubenswrapper[4824]: I1211 02:02:14.206847 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:14 crc kubenswrapper[4824]: I1211 02:02:14.206921 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:14 crc kubenswrapper[4824]: I1211 02:02:14.206938 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:14 crc kubenswrapper[4824]: I1211 02:02:14.206963 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:14 crc kubenswrapper[4824]: I1211 02:02:14.206980 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:14Z","lastTransitionTime":"2025-12-11T02:02:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:14 crc kubenswrapper[4824]: I1211 02:02:14.308792 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:14 crc kubenswrapper[4824]: I1211 02:02:14.308851 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:14 crc kubenswrapper[4824]: I1211 02:02:14.308868 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:14 crc kubenswrapper[4824]: I1211 02:02:14.308891 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:14 crc kubenswrapper[4824]: I1211 02:02:14.308909 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:14Z","lastTransitionTime":"2025-12-11T02:02:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:14 crc kubenswrapper[4824]: I1211 02:02:14.410971 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:14 crc kubenswrapper[4824]: I1211 02:02:14.411044 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:14 crc kubenswrapper[4824]: I1211 02:02:14.411059 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:14 crc kubenswrapper[4824]: I1211 02:02:14.411084 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:14 crc kubenswrapper[4824]: I1211 02:02:14.411101 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:14Z","lastTransitionTime":"2025-12-11T02:02:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:14 crc kubenswrapper[4824]: I1211 02:02:14.513922 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:14 crc kubenswrapper[4824]: I1211 02:02:14.513997 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:14 crc kubenswrapper[4824]: I1211 02:02:14.514024 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:14 crc kubenswrapper[4824]: I1211 02:02:14.514056 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:14 crc kubenswrapper[4824]: I1211 02:02:14.514079 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:14Z","lastTransitionTime":"2025-12-11T02:02:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:14 crc kubenswrapper[4824]: I1211 02:02:14.616268 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:14 crc kubenswrapper[4824]: I1211 02:02:14.616336 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:14 crc kubenswrapper[4824]: I1211 02:02:14.616356 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:14 crc kubenswrapper[4824]: I1211 02:02:14.616381 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:14 crc kubenswrapper[4824]: I1211 02:02:14.616399 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:14Z","lastTransitionTime":"2025-12-11T02:02:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:14 crc kubenswrapper[4824]: I1211 02:02:14.720186 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:14 crc kubenswrapper[4824]: I1211 02:02:14.720261 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:14 crc kubenswrapper[4824]: I1211 02:02:14.720278 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:14 crc kubenswrapper[4824]: I1211 02:02:14.720304 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:14 crc kubenswrapper[4824]: I1211 02:02:14.720329 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:14Z","lastTransitionTime":"2025-12-11T02:02:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:14 crc kubenswrapper[4824]: I1211 02:02:14.823771 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:14 crc kubenswrapper[4824]: I1211 02:02:14.823840 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:14 crc kubenswrapper[4824]: I1211 02:02:14.823857 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:14 crc kubenswrapper[4824]: I1211 02:02:14.823885 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:14 crc kubenswrapper[4824]: I1211 02:02:14.823902 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:14Z","lastTransitionTime":"2025-12-11T02:02:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:14 crc kubenswrapper[4824]: I1211 02:02:14.930872 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:14 crc kubenswrapper[4824]: I1211 02:02:14.930925 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:14 crc kubenswrapper[4824]: I1211 02:02:14.930942 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:14 crc kubenswrapper[4824]: I1211 02:02:14.930966 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:14 crc kubenswrapper[4824]: I1211 02:02:14.930982 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:14Z","lastTransitionTime":"2025-12-11T02:02:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:15 crc kubenswrapper[4824]: I1211 02:02:15.034615 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:15 crc kubenswrapper[4824]: I1211 02:02:15.034679 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:15 crc kubenswrapper[4824]: I1211 02:02:15.034700 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:15 crc kubenswrapper[4824]: I1211 02:02:15.034728 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:15 crc kubenswrapper[4824]: I1211 02:02:15.034746 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:15Z","lastTransitionTime":"2025-12-11T02:02:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:15 crc kubenswrapper[4824]: I1211 02:02:15.137593 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:15 crc kubenswrapper[4824]: I1211 02:02:15.137653 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:15 crc kubenswrapper[4824]: I1211 02:02:15.137692 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:15 crc kubenswrapper[4824]: I1211 02:02:15.137722 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:15 crc kubenswrapper[4824]: I1211 02:02:15.137749 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:15Z","lastTransitionTime":"2025-12-11T02:02:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:15 crc kubenswrapper[4824]: I1211 02:02:15.240742 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:15 crc kubenswrapper[4824]: I1211 02:02:15.240790 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:15 crc kubenswrapper[4824]: I1211 02:02:15.240807 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:15 crc kubenswrapper[4824]: I1211 02:02:15.240828 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:15 crc kubenswrapper[4824]: I1211 02:02:15.240844 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:15Z","lastTransitionTime":"2025-12-11T02:02:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:15 crc kubenswrapper[4824]: I1211 02:02:15.343994 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:15 crc kubenswrapper[4824]: I1211 02:02:15.344055 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:15 crc kubenswrapper[4824]: I1211 02:02:15.344080 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:15 crc kubenswrapper[4824]: I1211 02:02:15.344145 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:15 crc kubenswrapper[4824]: I1211 02:02:15.344175 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:15Z","lastTransitionTime":"2025-12-11T02:02:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:15 crc kubenswrapper[4824]: I1211 02:02:15.447781 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:15 crc kubenswrapper[4824]: I1211 02:02:15.448041 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:15 crc kubenswrapper[4824]: I1211 02:02:15.448061 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:15 crc kubenswrapper[4824]: I1211 02:02:15.448084 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:15 crc kubenswrapper[4824]: I1211 02:02:15.448101 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:15Z","lastTransitionTime":"2025-12-11T02:02:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:15 crc kubenswrapper[4824]: I1211 02:02:15.550997 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:15 crc kubenswrapper[4824]: I1211 02:02:15.551066 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:15 crc kubenswrapper[4824]: I1211 02:02:15.551090 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:15 crc kubenswrapper[4824]: I1211 02:02:15.551154 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:15 crc kubenswrapper[4824]: I1211 02:02:15.551182 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:15Z","lastTransitionTime":"2025-12-11T02:02:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:15 crc kubenswrapper[4824]: I1211 02:02:15.631765 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-grnpt" Dec 11 02:02:15 crc kubenswrapper[4824]: I1211 02:02:15.631848 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 02:02:15 crc kubenswrapper[4824]: I1211 02:02:15.631845 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 02:02:15 crc kubenswrapper[4824]: I1211 02:02:15.631788 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 02:02:15 crc kubenswrapper[4824]: E1211 02:02:15.631985 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-grnpt" podUID="09a52948-0404-45ff-85b1-51479268bf71" Dec 11 02:02:15 crc kubenswrapper[4824]: E1211 02:02:15.632157 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 02:02:15 crc kubenswrapper[4824]: E1211 02:02:15.632334 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 02:02:15 crc kubenswrapper[4824]: E1211 02:02:15.632526 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 02:02:15 crc kubenswrapper[4824]: I1211 02:02:15.654418 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:15 crc kubenswrapper[4824]: I1211 02:02:15.654459 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:15 crc kubenswrapper[4824]: I1211 02:02:15.654474 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:15 crc kubenswrapper[4824]: I1211 02:02:15.654494 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:15 crc kubenswrapper[4824]: I1211 02:02:15.654511 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:15Z","lastTransitionTime":"2025-12-11T02:02:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:15 crc kubenswrapper[4824]: I1211 02:02:15.757647 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:15 crc kubenswrapper[4824]: I1211 02:02:15.757703 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:15 crc kubenswrapper[4824]: I1211 02:02:15.757720 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:15 crc kubenswrapper[4824]: I1211 02:02:15.757742 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:15 crc kubenswrapper[4824]: I1211 02:02:15.757761 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:15Z","lastTransitionTime":"2025-12-11T02:02:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:15 crc kubenswrapper[4824]: I1211 02:02:15.860669 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:15 crc kubenswrapper[4824]: I1211 02:02:15.860766 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:15 crc kubenswrapper[4824]: I1211 02:02:15.860788 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:15 crc kubenswrapper[4824]: I1211 02:02:15.860816 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:15 crc kubenswrapper[4824]: I1211 02:02:15.860834 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:15Z","lastTransitionTime":"2025-12-11T02:02:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:15 crc kubenswrapper[4824]: I1211 02:02:15.962975 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:15 crc kubenswrapper[4824]: I1211 02:02:15.963036 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:15 crc kubenswrapper[4824]: I1211 02:02:15.963048 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:15 crc kubenswrapper[4824]: I1211 02:02:15.963063 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:15 crc kubenswrapper[4824]: I1211 02:02:15.963075 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:15Z","lastTransitionTime":"2025-12-11T02:02:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:16 crc kubenswrapper[4824]: I1211 02:02:16.067076 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:16 crc kubenswrapper[4824]: I1211 02:02:16.067163 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:16 crc kubenswrapper[4824]: I1211 02:02:16.067182 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:16 crc kubenswrapper[4824]: I1211 02:02:16.067208 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:16 crc kubenswrapper[4824]: I1211 02:02:16.067225 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:16Z","lastTransitionTime":"2025-12-11T02:02:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:16 crc kubenswrapper[4824]: I1211 02:02:16.170659 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:16 crc kubenswrapper[4824]: I1211 02:02:16.170714 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:16 crc kubenswrapper[4824]: I1211 02:02:16.170730 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:16 crc kubenswrapper[4824]: I1211 02:02:16.170754 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:16 crc kubenswrapper[4824]: I1211 02:02:16.170771 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:16Z","lastTransitionTime":"2025-12-11T02:02:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:16 crc kubenswrapper[4824]: I1211 02:02:16.273716 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:16 crc kubenswrapper[4824]: I1211 02:02:16.273850 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:16 crc kubenswrapper[4824]: I1211 02:02:16.273874 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:16 crc kubenswrapper[4824]: I1211 02:02:16.273899 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:16 crc kubenswrapper[4824]: I1211 02:02:16.273917 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:16Z","lastTransitionTime":"2025-12-11T02:02:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:16 crc kubenswrapper[4824]: I1211 02:02:16.377211 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:16 crc kubenswrapper[4824]: I1211 02:02:16.377285 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:16 crc kubenswrapper[4824]: I1211 02:02:16.377301 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:16 crc kubenswrapper[4824]: I1211 02:02:16.377326 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:16 crc kubenswrapper[4824]: I1211 02:02:16.377343 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:16Z","lastTransitionTime":"2025-12-11T02:02:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:16 crc kubenswrapper[4824]: I1211 02:02:16.480484 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:16 crc kubenswrapper[4824]: I1211 02:02:16.480536 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:16 crc kubenswrapper[4824]: I1211 02:02:16.480548 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:16 crc kubenswrapper[4824]: I1211 02:02:16.480564 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:16 crc kubenswrapper[4824]: I1211 02:02:16.480576 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:16Z","lastTransitionTime":"2025-12-11T02:02:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:16 crc kubenswrapper[4824]: I1211 02:02:16.584599 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:16 crc kubenswrapper[4824]: I1211 02:02:16.584660 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:16 crc kubenswrapper[4824]: I1211 02:02:16.584671 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:16 crc kubenswrapper[4824]: I1211 02:02:16.584689 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:16 crc kubenswrapper[4824]: I1211 02:02:16.584702 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:16Z","lastTransitionTime":"2025-12-11T02:02:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:16 crc kubenswrapper[4824]: I1211 02:02:16.687789 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:16 crc kubenswrapper[4824]: I1211 02:02:16.687848 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:16 crc kubenswrapper[4824]: I1211 02:02:16.687865 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:16 crc kubenswrapper[4824]: I1211 02:02:16.687887 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:16 crc kubenswrapper[4824]: I1211 02:02:16.687905 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:16Z","lastTransitionTime":"2025-12-11T02:02:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:16 crc kubenswrapper[4824]: I1211 02:02:16.791389 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:16 crc kubenswrapper[4824]: I1211 02:02:16.791455 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:16 crc kubenswrapper[4824]: I1211 02:02:16.791472 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:16 crc kubenswrapper[4824]: I1211 02:02:16.791494 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:16 crc kubenswrapper[4824]: I1211 02:02:16.791512 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:16Z","lastTransitionTime":"2025-12-11T02:02:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:16 crc kubenswrapper[4824]: I1211 02:02:16.895530 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:16 crc kubenswrapper[4824]: I1211 02:02:16.895602 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:16 crc kubenswrapper[4824]: I1211 02:02:16.895626 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:16 crc kubenswrapper[4824]: I1211 02:02:16.895658 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:16 crc kubenswrapper[4824]: I1211 02:02:16.895680 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:16Z","lastTransitionTime":"2025-12-11T02:02:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:16 crc kubenswrapper[4824]: I1211 02:02:16.999490 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:16 crc kubenswrapper[4824]: I1211 02:02:16.999553 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:16 crc kubenswrapper[4824]: I1211 02:02:16.999569 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:16 crc kubenswrapper[4824]: I1211 02:02:16.999591 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:16 crc kubenswrapper[4824]: I1211 02:02:16.999611 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:16Z","lastTransitionTime":"2025-12-11T02:02:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:17 crc kubenswrapper[4824]: I1211 02:02:17.102727 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:17 crc kubenswrapper[4824]: I1211 02:02:17.102790 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:17 crc kubenswrapper[4824]: I1211 02:02:17.102808 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:17 crc kubenswrapper[4824]: I1211 02:02:17.102832 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:17 crc kubenswrapper[4824]: I1211 02:02:17.102848 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:17Z","lastTransitionTime":"2025-12-11T02:02:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:17 crc kubenswrapper[4824]: I1211 02:02:17.206446 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:17 crc kubenswrapper[4824]: I1211 02:02:17.206500 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:17 crc kubenswrapper[4824]: I1211 02:02:17.206517 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:17 crc kubenswrapper[4824]: I1211 02:02:17.206538 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:17 crc kubenswrapper[4824]: I1211 02:02:17.206555 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:17Z","lastTransitionTime":"2025-12-11T02:02:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:17 crc kubenswrapper[4824]: I1211 02:02:17.309194 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:17 crc kubenswrapper[4824]: I1211 02:02:17.309231 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:17 crc kubenswrapper[4824]: I1211 02:02:17.309241 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:17 crc kubenswrapper[4824]: I1211 02:02:17.309283 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:17 crc kubenswrapper[4824]: I1211 02:02:17.309297 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:17Z","lastTransitionTime":"2025-12-11T02:02:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:17 crc kubenswrapper[4824]: I1211 02:02:17.411659 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:17 crc kubenswrapper[4824]: I1211 02:02:17.411727 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:17 crc kubenswrapper[4824]: I1211 02:02:17.411748 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:17 crc kubenswrapper[4824]: I1211 02:02:17.411771 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:17 crc kubenswrapper[4824]: I1211 02:02:17.411788 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:17Z","lastTransitionTime":"2025-12-11T02:02:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:17 crc kubenswrapper[4824]: I1211 02:02:17.514726 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:17 crc kubenswrapper[4824]: I1211 02:02:17.514772 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:17 crc kubenswrapper[4824]: I1211 02:02:17.514781 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:17 crc kubenswrapper[4824]: I1211 02:02:17.514795 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:17 crc kubenswrapper[4824]: I1211 02:02:17.514804 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:17Z","lastTransitionTime":"2025-12-11T02:02:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:17 crc kubenswrapper[4824]: I1211 02:02:17.617907 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:17 crc kubenswrapper[4824]: I1211 02:02:17.617993 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:17 crc kubenswrapper[4824]: I1211 02:02:17.618021 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:17 crc kubenswrapper[4824]: I1211 02:02:17.618052 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:17 crc kubenswrapper[4824]: I1211 02:02:17.618076 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:17Z","lastTransitionTime":"2025-12-11T02:02:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:17 crc kubenswrapper[4824]: I1211 02:02:17.631482 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 02:02:17 crc kubenswrapper[4824]: I1211 02:02:17.631505 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 02:02:17 crc kubenswrapper[4824]: I1211 02:02:17.631667 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-grnpt" Dec 11 02:02:17 crc kubenswrapper[4824]: E1211 02:02:17.631656 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 02:02:17 crc kubenswrapper[4824]: I1211 02:02:17.632280 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 02:02:17 crc kubenswrapper[4824]: E1211 02:02:17.632385 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 02:02:17 crc kubenswrapper[4824]: E1211 02:02:17.632496 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 02:02:17 crc kubenswrapper[4824]: E1211 02:02:17.632543 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-grnpt" podUID="09a52948-0404-45ff-85b1-51479268bf71" Dec 11 02:02:17 crc kubenswrapper[4824]: I1211 02:02:17.649450 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Dec 11 02:02:17 crc kubenswrapper[4824]: I1211 02:02:17.720756 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:17 crc kubenswrapper[4824]: I1211 02:02:17.720798 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:17 crc kubenswrapper[4824]: I1211 02:02:17.720834 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:17 crc kubenswrapper[4824]: I1211 02:02:17.720852 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:17 crc kubenswrapper[4824]: I1211 02:02:17.720863 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:17Z","lastTransitionTime":"2025-12-11T02:02:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:17 crc kubenswrapper[4824]: I1211 02:02:17.823024 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:17 crc kubenswrapper[4824]: I1211 02:02:17.823063 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:17 crc kubenswrapper[4824]: I1211 02:02:17.823094 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:17 crc kubenswrapper[4824]: I1211 02:02:17.823131 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:17 crc kubenswrapper[4824]: I1211 02:02:17.823144 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:17Z","lastTransitionTime":"2025-12-11T02:02:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:17 crc kubenswrapper[4824]: I1211 02:02:17.926205 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:17 crc kubenswrapper[4824]: I1211 02:02:17.926257 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:17 crc kubenswrapper[4824]: I1211 02:02:17.926274 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:17 crc kubenswrapper[4824]: I1211 02:02:17.926299 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:17 crc kubenswrapper[4824]: I1211 02:02:17.926316 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:17Z","lastTransitionTime":"2025-12-11T02:02:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:18 crc kubenswrapper[4824]: I1211 02:02:18.028420 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:18 crc kubenswrapper[4824]: I1211 02:02:18.028543 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:18 crc kubenswrapper[4824]: I1211 02:02:18.028558 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:18 crc kubenswrapper[4824]: I1211 02:02:18.028573 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:18 crc kubenswrapper[4824]: I1211 02:02:18.028584 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:18Z","lastTransitionTime":"2025-12-11T02:02:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:18 crc kubenswrapper[4824]: I1211 02:02:18.131560 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:18 crc kubenswrapper[4824]: I1211 02:02:18.131607 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:18 crc kubenswrapper[4824]: I1211 02:02:18.131623 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:18 crc kubenswrapper[4824]: I1211 02:02:18.131645 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:18 crc kubenswrapper[4824]: I1211 02:02:18.131665 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:18Z","lastTransitionTime":"2025-12-11T02:02:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:18 crc kubenswrapper[4824]: I1211 02:02:18.233294 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:18 crc kubenswrapper[4824]: I1211 02:02:18.233326 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:18 crc kubenswrapper[4824]: I1211 02:02:18.233335 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:18 crc kubenswrapper[4824]: I1211 02:02:18.233350 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:18 crc kubenswrapper[4824]: I1211 02:02:18.233360 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:18Z","lastTransitionTime":"2025-12-11T02:02:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:18 crc kubenswrapper[4824]: I1211 02:02:18.337039 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:18 crc kubenswrapper[4824]: I1211 02:02:18.337430 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:18 crc kubenswrapper[4824]: I1211 02:02:18.337443 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:18 crc kubenswrapper[4824]: I1211 02:02:18.337463 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:18 crc kubenswrapper[4824]: I1211 02:02:18.337477 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:18Z","lastTransitionTime":"2025-12-11T02:02:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:18 crc kubenswrapper[4824]: I1211 02:02:18.440573 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:18 crc kubenswrapper[4824]: I1211 02:02:18.440635 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:18 crc kubenswrapper[4824]: I1211 02:02:18.440651 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:18 crc kubenswrapper[4824]: I1211 02:02:18.440676 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:18 crc kubenswrapper[4824]: I1211 02:02:18.440693 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:18Z","lastTransitionTime":"2025-12-11T02:02:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:18 crc kubenswrapper[4824]: I1211 02:02:18.543022 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:18 crc kubenswrapper[4824]: I1211 02:02:18.543083 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:18 crc kubenswrapper[4824]: I1211 02:02:18.543103 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:18 crc kubenswrapper[4824]: I1211 02:02:18.543161 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:18 crc kubenswrapper[4824]: I1211 02:02:18.543181 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:18Z","lastTransitionTime":"2025-12-11T02:02:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:18 crc kubenswrapper[4824]: I1211 02:02:18.645602 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:18 crc kubenswrapper[4824]: I1211 02:02:18.645710 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:18 crc kubenswrapper[4824]: I1211 02:02:18.645735 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:18 crc kubenswrapper[4824]: I1211 02:02:18.645806 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:18 crc kubenswrapper[4824]: I1211 02:02:18.645827 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:18Z","lastTransitionTime":"2025-12-11T02:02:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:18 crc kubenswrapper[4824]: I1211 02:02:18.657139 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kh8g9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"50c69d41-a65b-463e-a829-ea349bd2fe46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dca6be41bd7a3ce12ecd4567c868f696b65e8eb1bad436feda6e4f9bfb8e525\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69af1f33fd0bafc929b94e702b4f46c6c7d7a362d5552b121729c477892d91d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69af1f33fd0bafc929b94e702b4f46c6c7d7a362d5552b121729c477892d91d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd9675d0afded587bc660a3862fd0c5dd6885961afacda957aa9ee8501d68a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd9675d0afded587bc660a3862fd0c5dd6885961afacda957aa9ee8501d68a69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aac94149458bd6321be90b26b48ee26428d273e7f61407d95f9c9d6b3aeae4b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aac94149458bd6321be90b26b48ee26428d273e7f61407d95f9c9d6b3aeae4b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbee11bd7cb65e031415990869b52d2b8f81bf8b56e6cc935d38e09695db7b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbee11bd7cb65e031415990869b52d2b8f81bf8b56e6cc935d38e09695db7b48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86f693ade72bd593bc16021fe4b7564f1914ec2f1cda7039474ac2fb13de10f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86f693ade72bd593bc16021fe4b7564f1914ec2f1cda7039474ac2fb13de10f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7d7674252f67bde0ded2ec8ad1d925a92a26d3e3f741dee0061b735cb0b7a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7d7674252f67bde0ded2ec8ad1d925a92a26d3e3f741dee0061b735cb0b7a02\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kh8g9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:18Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:18 crc kubenswrapper[4824]: I1211 02:02:18.681028 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6bkc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8022f696-d0e7-437b-848a-3cd25bd1f364\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f40719f59881e077e7edbab3adb9efd101359bd461e5434c365e119e8cdbddb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3e28f6778ec0dba6ea3e6b69e5a2c1c73c0d5db1abeead241fb58efdbfaefe4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T02:02:03Z\\\",\\\"message\\\":\\\"2025-12-11T02:01:18+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_00fc4257-9dfb-4131-a927-0fc54bfa1304\\\\n2025-12-11T02:01:18+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_00fc4257-9dfb-4131-a927-0fc54bfa1304 to /host/opt/cni/bin/\\\\n2025-12-11T02:01:18Z [verbose] multus-daemon started\\\\n2025-12-11T02:01:18Z [verbose] Readiness Indicator file check\\\\n2025-12-11T02:02:03Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:02:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hh5kn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6bkc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:18Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:18 crc kubenswrapper[4824]: I1211 02:02:18.717844 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b193bd2-b536-4056-92f1-94c9836ab2eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51f3c41d86ddc44f7eca118029876ad353317f753d0ad173ed4d30c2a15f10e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c345850891744a0240781277e9a89a59f6899c4c944995f62e44650f807440a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed2ae40b1fe28111efc76d9d72345edecb5f3ea34522b238f3730812b29f2bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce7aa078800e6f299cfe03770fa7e3b64fa3742fe0c0f383c87d84cf48569ddf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6aa9fc1f1bff192c69bd59d32a05d4a72d09641d9ce4b265bc4517147a7e15dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b4a3bc50eb80777188206adab774e8acad8b2e4952a8415f91b1c29ba9a0ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://154d82fd6acee749bfd37220547e52647dc7f9d86a73c28ce7d8577c9cadef6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://154d82fd6acee749bfd37220547e52647dc7f9d86a73c28ce7d8577c9cadef6e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T02:02:10Z\\\",\\\"message\\\":\\\"t handler 9 for removal\\\\nI1211 02:02:10.678557 6803 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1211 02:02:10.678594 6803 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1211 02:02:10.678724 6803 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1211 02:02:10.678871 6803 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1211 02:02:10.678731 6803 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1211 02:02:10.678906 6803 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1211 02:02:10.678920 6803 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1211 02:02:10.678935 6803 handler.go:208] Removed *v1.Node event handler 7\\\\nI1211 02:02:10.678961 6803 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1211 02:02:10.678945 6803 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1211 02:02:10.678980 6803 handler.go:208] Removed *v1.Node event handler 2\\\\nI1211 02:02:10.678982 6803 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1211 02:02:10.679043 6803 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1211 02:02:10.679161 6803 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1211 02:02:10.679264 6803 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T02:02:09Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-f6pgc_openshift-ovn-kubernetes(5b193bd2-b536-4056-92f1-94c9836ab2eb)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ead58694813806d307bbbc6a7b3cf41730d1c059b231a3002684bf09c10c7f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f6pgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:18Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:18 crc kubenswrapper[4824]: I1211 02:02:18.737218 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-br4j9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce125194-9c9d-4ea4-8a72-2c30d5722295\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9798f904052e08196a0db1745060dab11e2710308bca01ff160475910f2665df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fc5s8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-br4j9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:18Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:18 crc kubenswrapper[4824]: I1211 02:02:18.751227 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:18 crc kubenswrapper[4824]: I1211 02:02:18.751289 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:18 crc kubenswrapper[4824]: I1211 02:02:18.751309 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:18 crc kubenswrapper[4824]: I1211 02:02:18.751337 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:18 crc kubenswrapper[4824]: I1211 02:02:18.751361 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:18Z","lastTransitionTime":"2025-12-11T02:02:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:18 crc kubenswrapper[4824]: I1211 02:02:18.758829 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8v4jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99a6bfee-afca-4726-a964-b95a948945a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b23b9d8f18db3eeadb60f83b36b05b1778cd0fa31669dad2680bb335ae004acf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4hk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3acc1940a0534ba0e6096175e41834fdd0fbcbb4c3fb6e9a38e76aedae55d2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4hk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-8v4jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:18Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:18 crc kubenswrapper[4824]: I1211 02:02:18.784156 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-grnpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09a52948-0404-45ff-85b1-51479268bf71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmmzm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmmzm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-grnpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:18Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:18 crc kubenswrapper[4824]: I1211 02:02:18.804387 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:18Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:18 crc kubenswrapper[4824]: I1211 02:02:18.826877 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:18Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:18 crc kubenswrapper[4824]: I1211 02:02:18.844943 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44b156e4-64a4-4d45-aa5e-9b10a862faed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7368d92144b188e95a7796fabf497b0f8d9c0b240293bdc7b4aec1ff75497bc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w62dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcf289ac27dbda4f24aabbde3fb5119329e7b4fc7d6e018168da9a69661a4c57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w62dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gx6xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:18Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:18 crc kubenswrapper[4824]: I1211 02:02:18.854308 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:18 crc kubenswrapper[4824]: I1211 02:02:18.854360 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:18 crc kubenswrapper[4824]: I1211 02:02:18.854379 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:18 crc kubenswrapper[4824]: I1211 02:02:18.854401 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:18 crc kubenswrapper[4824]: I1211 02:02:18.854418 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:18Z","lastTransitionTime":"2025-12-11T02:02:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:18 crc kubenswrapper[4824]: I1211 02:02:18.867137 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d14986666febd651039afb52ddf4db8969aca55e88eda9bbb20a67e35ba722ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:18Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:18 crc kubenswrapper[4824]: I1211 02:02:18.890593 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"253d4291-2b2f-487c-8862-3d305f0bdc55\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44b29bb7a6eda7cb890d36f86375b5f3f197ae5cf390f16e71139b0904873e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://922af7f67b99a0ce08776f3f75c0c377e6ea69af1720b6178a6d1e2f9b4b8ba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://136c68020325db8e6930d0f40b0d0d7e013c14c4dd5e7d1449d01d6d458104d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c89afaf56b4a9597f6543ec2e3b176c7a6dab7dc341941fc3bc8ec85cc4edaf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:00:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:18Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:18 crc kubenswrapper[4824]: I1211 02:02:18.910323 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98027cfa-5dde-4297-9f74-c4d662c7f82b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f410073059c5898660c50f775a270bebe51836f73e2a2f02db38a2aabf15931d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://335a774b8a3aff18695de4aba0b522d346aecc8ed8d96a80150301f5debce32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://339d619250c1c0a12f988a9b0ebc03874db3ac41a9d3b3f8573795467578cd0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e27af6c438009a44288f8bc10f8ebf2b61d9309f87e8d955e8d255a7190174f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e27af6c438009a44288f8bc10f8ebf2b61d9309f87e8d955e8d255a7190174f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:00:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:18Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:18 crc kubenswrapper[4824]: I1211 02:02:18.932576 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:18Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:18 crc kubenswrapper[4824]: I1211 02:02:18.955751 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"beab386b-6c65-4bc3-9ef0-dc3f68b4403f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c26f06d2ed5e6e98b61a51f073130675f2c49acda106b8aedcbb2a6d4f7580c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20af9b10a67c8b21a7abcd23e42bc412798d159c8422e7c8ff58f5b997ca19a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9bbfc6edfe050b3fa9dbeee5f98d416bcb2a1533a38608a881b75396fd0b059\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7c766f8de73ca762ffde0c18b2d1c46bb152b2b37535a611c5a0e454a21b7d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d45819f0611c91b20c83eeee1092031a3441fb69f3015241632ad419880227d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 02:01:11.070250 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 02:01:11.071673 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2752882565/tls.crt::/tmp/serving-cert-2752882565/tls.key\\\\\\\"\\\\nI1211 02:01:17.042610 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 02:01:17.048046 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 02:01:17.048080 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 02:01:17.048172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 02:01:17.048188 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 02:01:17.058248 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 02:01:17.058294 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 02:01:17.058303 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 02:01:17.058315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 02:01:17.058322 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 02:01:17.058328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 02:01:17.058335 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 02:01:17.058808 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 02:01:17.060225 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14a940a72b2404a6cbd52a69e150b03266bcd7f5c1d3f4d07b51c0c69e51786\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:00:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:18Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:18 crc kubenswrapper[4824]: I1211 02:02:18.962413 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:18 crc kubenswrapper[4824]: I1211 02:02:18.962485 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:18 crc kubenswrapper[4824]: I1211 02:02:18.962504 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:18 crc kubenswrapper[4824]: I1211 02:02:18.962528 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:18 crc kubenswrapper[4824]: I1211 02:02:18.962545 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:18Z","lastTransitionTime":"2025-12-11T02:02:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:18 crc kubenswrapper[4824]: I1211 02:02:18.982503 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ded146c045085be2b58821749052413a36850ede00ae1024fe2dae9fa5ac84b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:18Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:19 crc kubenswrapper[4824]: I1211 02:02:19.005643 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76b10bbd1e891b8c682f4b2e17d023fd787a8ed44ed94fbf7d97314d641ccf36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f32838f33832b9b67e7dd9c53340680e17721cf1c7bdf0170230c5f8feafe64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:19Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:19 crc kubenswrapper[4824]: I1211 02:02:19.020641 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kptws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68568dba-83a5-46c1-8d01-337359679d91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f965a6d6db619995132f61b5d51b66f0b2b3fdbef3fbec63e231b9a62ccd0a9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd7nf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kptws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:19Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:19 crc kubenswrapper[4824]: I1211 02:02:19.055354 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afca157c-af3d-4197-8a60-39fc607efd69\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bae1b2b71f76385e3144a18a4335f125adaf2b40a59e9f8c353ab2fd5efd5549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09587408482a9771161017d1246b051cb9ad2961bd949414210b3c0ad58ae3fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e42434f5fe6b847211ad32b3fee7e6a6a6adbe9c5caa30235ac07a25507f7871\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af95eed3f8b150973c8f151d41b7a8ce2c41d90ede17bf0f027b721f4f3858d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0426825294b0be03de99d39e8a1584f7717017b99d70a12af858ff8ea0699302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18a2507c91dd7ab306ef38abf724e07f468190caba6cb110bf3224448f631aad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18a2507c91dd7ab306ef38abf724e07f468190caba6cb110bf3224448f631aad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be8f3f2460ba45c67ac7c6e7eac2d6ada631e90f4c7de645a41c134539437f82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be8f3f2460ba45c67ac7c6e7eac2d6ada631e90f4c7de645a41c134539437f82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://016b733da5b0e8f0e3830ee9746c4ab5decfcac7573f42aec257a842a393c60d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://016b733da5b0e8f0e3830ee9746c4ab5decfcac7573f42aec257a842a393c60d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:00:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:19Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:19 crc kubenswrapper[4824]: I1211 02:02:19.065381 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:19 crc kubenswrapper[4824]: I1211 02:02:19.065436 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:19 crc kubenswrapper[4824]: I1211 02:02:19.065452 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:19 crc kubenswrapper[4824]: I1211 02:02:19.065478 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:19 crc kubenswrapper[4824]: I1211 02:02:19.065495 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:19Z","lastTransitionTime":"2025-12-11T02:02:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:19 crc kubenswrapper[4824]: I1211 02:02:19.168355 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:19 crc kubenswrapper[4824]: I1211 02:02:19.168428 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:19 crc kubenswrapper[4824]: I1211 02:02:19.168447 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:19 crc kubenswrapper[4824]: I1211 02:02:19.168472 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:19 crc kubenswrapper[4824]: I1211 02:02:19.168489 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:19Z","lastTransitionTime":"2025-12-11T02:02:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:19 crc kubenswrapper[4824]: I1211 02:02:19.270952 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:19 crc kubenswrapper[4824]: I1211 02:02:19.271012 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:19 crc kubenswrapper[4824]: I1211 02:02:19.271029 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:19 crc kubenswrapper[4824]: I1211 02:02:19.271055 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:19 crc kubenswrapper[4824]: I1211 02:02:19.271073 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:19Z","lastTransitionTime":"2025-12-11T02:02:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:19 crc kubenswrapper[4824]: I1211 02:02:19.373723 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:19 crc kubenswrapper[4824]: I1211 02:02:19.373785 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:19 crc kubenswrapper[4824]: I1211 02:02:19.373807 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:19 crc kubenswrapper[4824]: I1211 02:02:19.373836 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:19 crc kubenswrapper[4824]: I1211 02:02:19.373858 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:19Z","lastTransitionTime":"2025-12-11T02:02:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:19 crc kubenswrapper[4824]: I1211 02:02:19.475963 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:19 crc kubenswrapper[4824]: I1211 02:02:19.476032 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:19 crc kubenswrapper[4824]: I1211 02:02:19.476054 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:19 crc kubenswrapper[4824]: I1211 02:02:19.476085 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:19 crc kubenswrapper[4824]: I1211 02:02:19.476140 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:19Z","lastTransitionTime":"2025-12-11T02:02:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:19 crc kubenswrapper[4824]: I1211 02:02:19.578424 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:19 crc kubenswrapper[4824]: I1211 02:02:19.578505 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:19 crc kubenswrapper[4824]: I1211 02:02:19.578529 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:19 crc kubenswrapper[4824]: I1211 02:02:19.578558 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:19 crc kubenswrapper[4824]: I1211 02:02:19.578581 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:19Z","lastTransitionTime":"2025-12-11T02:02:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:19 crc kubenswrapper[4824]: I1211 02:02:19.632333 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 02:02:19 crc kubenswrapper[4824]: I1211 02:02:19.632429 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 02:02:19 crc kubenswrapper[4824]: E1211 02:02:19.632598 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 02:02:19 crc kubenswrapper[4824]: I1211 02:02:19.632639 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 02:02:19 crc kubenswrapper[4824]: E1211 02:02:19.632724 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 02:02:19 crc kubenswrapper[4824]: I1211 02:02:19.632637 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-grnpt" Dec 11 02:02:19 crc kubenswrapper[4824]: E1211 02:02:19.632812 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 02:02:19 crc kubenswrapper[4824]: E1211 02:02:19.633026 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-grnpt" podUID="09a52948-0404-45ff-85b1-51479268bf71" Dec 11 02:02:19 crc kubenswrapper[4824]: I1211 02:02:19.681983 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:19 crc kubenswrapper[4824]: I1211 02:02:19.682061 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:19 crc kubenswrapper[4824]: I1211 02:02:19.682080 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:19 crc kubenswrapper[4824]: I1211 02:02:19.682107 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:19 crc kubenswrapper[4824]: I1211 02:02:19.682156 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:19Z","lastTransitionTime":"2025-12-11T02:02:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:19 crc kubenswrapper[4824]: I1211 02:02:19.784722 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:19 crc kubenswrapper[4824]: I1211 02:02:19.784791 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:19 crc kubenswrapper[4824]: I1211 02:02:19.784813 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:19 crc kubenswrapper[4824]: I1211 02:02:19.784844 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:19 crc kubenswrapper[4824]: I1211 02:02:19.784869 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:19Z","lastTransitionTime":"2025-12-11T02:02:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:19 crc kubenswrapper[4824]: I1211 02:02:19.887635 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:19 crc kubenswrapper[4824]: I1211 02:02:19.887696 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:19 crc kubenswrapper[4824]: I1211 02:02:19.887718 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:19 crc kubenswrapper[4824]: I1211 02:02:19.887740 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:19 crc kubenswrapper[4824]: I1211 02:02:19.887756 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:19Z","lastTransitionTime":"2025-12-11T02:02:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:19 crc kubenswrapper[4824]: I1211 02:02:19.990526 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:19 crc kubenswrapper[4824]: I1211 02:02:19.990595 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:19 crc kubenswrapper[4824]: I1211 02:02:19.990612 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:19 crc kubenswrapper[4824]: I1211 02:02:19.990635 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:19 crc kubenswrapper[4824]: I1211 02:02:19.990653 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:19Z","lastTransitionTime":"2025-12-11T02:02:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:20 crc kubenswrapper[4824]: I1211 02:02:20.093282 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:20 crc kubenswrapper[4824]: I1211 02:02:20.093339 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:20 crc kubenswrapper[4824]: I1211 02:02:20.093359 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:20 crc kubenswrapper[4824]: I1211 02:02:20.093382 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:20 crc kubenswrapper[4824]: I1211 02:02:20.093398 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:20Z","lastTransitionTime":"2025-12-11T02:02:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:20 crc kubenswrapper[4824]: I1211 02:02:20.196282 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:20 crc kubenswrapper[4824]: I1211 02:02:20.196326 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:20 crc kubenswrapper[4824]: I1211 02:02:20.196337 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:20 crc kubenswrapper[4824]: I1211 02:02:20.196352 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:20 crc kubenswrapper[4824]: I1211 02:02:20.196362 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:20Z","lastTransitionTime":"2025-12-11T02:02:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:20 crc kubenswrapper[4824]: I1211 02:02:20.299538 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:20 crc kubenswrapper[4824]: I1211 02:02:20.299661 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:20 crc kubenswrapper[4824]: I1211 02:02:20.299723 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:20 crc kubenswrapper[4824]: I1211 02:02:20.299753 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:20 crc kubenswrapper[4824]: I1211 02:02:20.299770 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:20Z","lastTransitionTime":"2025-12-11T02:02:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:20 crc kubenswrapper[4824]: I1211 02:02:20.401754 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:20 crc kubenswrapper[4824]: I1211 02:02:20.401815 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:20 crc kubenswrapper[4824]: I1211 02:02:20.401834 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:20 crc kubenswrapper[4824]: I1211 02:02:20.401860 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:20 crc kubenswrapper[4824]: I1211 02:02:20.401880 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:20Z","lastTransitionTime":"2025-12-11T02:02:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:20 crc kubenswrapper[4824]: I1211 02:02:20.510009 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:20 crc kubenswrapper[4824]: I1211 02:02:20.510156 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:20 crc kubenswrapper[4824]: I1211 02:02:20.510177 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:20 crc kubenswrapper[4824]: I1211 02:02:20.510201 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:20 crc kubenswrapper[4824]: I1211 02:02:20.510219 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:20Z","lastTransitionTime":"2025-12-11T02:02:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:20 crc kubenswrapper[4824]: I1211 02:02:20.613648 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:20 crc kubenswrapper[4824]: I1211 02:02:20.613714 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:20 crc kubenswrapper[4824]: I1211 02:02:20.613732 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:20 crc kubenswrapper[4824]: I1211 02:02:20.613756 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:20 crc kubenswrapper[4824]: I1211 02:02:20.613774 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:20Z","lastTransitionTime":"2025-12-11T02:02:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:20 crc kubenswrapper[4824]: I1211 02:02:20.647698 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Dec 11 02:02:20 crc kubenswrapper[4824]: I1211 02:02:20.716529 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:20 crc kubenswrapper[4824]: I1211 02:02:20.716589 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:20 crc kubenswrapper[4824]: I1211 02:02:20.716607 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:20 crc kubenswrapper[4824]: I1211 02:02:20.716631 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:20 crc kubenswrapper[4824]: I1211 02:02:20.716650 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:20Z","lastTransitionTime":"2025-12-11T02:02:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:20 crc kubenswrapper[4824]: I1211 02:02:20.822431 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:20 crc kubenswrapper[4824]: I1211 02:02:20.822489 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:20 crc kubenswrapper[4824]: I1211 02:02:20.822502 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:20 crc kubenswrapper[4824]: I1211 02:02:20.822521 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:20 crc kubenswrapper[4824]: I1211 02:02:20.822534 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:20Z","lastTransitionTime":"2025-12-11T02:02:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:20 crc kubenswrapper[4824]: I1211 02:02:20.925736 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:20 crc kubenswrapper[4824]: I1211 02:02:20.925829 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:20 crc kubenswrapper[4824]: I1211 02:02:20.925861 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:20 crc kubenswrapper[4824]: I1211 02:02:20.925897 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:20 crc kubenswrapper[4824]: I1211 02:02:20.925920 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:20Z","lastTransitionTime":"2025-12-11T02:02:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:21 crc kubenswrapper[4824]: I1211 02:02:21.028931 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:21 crc kubenswrapper[4824]: I1211 02:02:21.029383 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:21 crc kubenswrapper[4824]: I1211 02:02:21.029410 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:21 crc kubenswrapper[4824]: I1211 02:02:21.029446 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:21 crc kubenswrapper[4824]: I1211 02:02:21.029465 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:21Z","lastTransitionTime":"2025-12-11T02:02:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:21 crc kubenswrapper[4824]: I1211 02:02:21.132459 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:21 crc kubenswrapper[4824]: I1211 02:02:21.132531 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:21 crc kubenswrapper[4824]: I1211 02:02:21.132548 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:21 crc kubenswrapper[4824]: I1211 02:02:21.132575 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:21 crc kubenswrapper[4824]: I1211 02:02:21.132595 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:21Z","lastTransitionTime":"2025-12-11T02:02:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:21 crc kubenswrapper[4824]: I1211 02:02:21.236357 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:21 crc kubenswrapper[4824]: I1211 02:02:21.236425 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:21 crc kubenswrapper[4824]: I1211 02:02:21.236446 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:21 crc kubenswrapper[4824]: I1211 02:02:21.236470 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:21 crc kubenswrapper[4824]: I1211 02:02:21.236490 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:21Z","lastTransitionTime":"2025-12-11T02:02:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:21 crc kubenswrapper[4824]: I1211 02:02:21.339439 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:21 crc kubenswrapper[4824]: I1211 02:02:21.339513 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:21 crc kubenswrapper[4824]: I1211 02:02:21.339534 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:21 crc kubenswrapper[4824]: I1211 02:02:21.339562 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:21 crc kubenswrapper[4824]: I1211 02:02:21.339584 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:21Z","lastTransitionTime":"2025-12-11T02:02:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:21 crc kubenswrapper[4824]: I1211 02:02:21.441860 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:21 crc kubenswrapper[4824]: I1211 02:02:21.441930 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:21 crc kubenswrapper[4824]: I1211 02:02:21.441948 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:21 crc kubenswrapper[4824]: I1211 02:02:21.441974 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:21 crc kubenswrapper[4824]: I1211 02:02:21.441992 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:21Z","lastTransitionTime":"2025-12-11T02:02:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:21 crc kubenswrapper[4824]: I1211 02:02:21.545196 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:21 crc kubenswrapper[4824]: I1211 02:02:21.545293 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:21 crc kubenswrapper[4824]: I1211 02:02:21.545313 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:21 crc kubenswrapper[4824]: I1211 02:02:21.545336 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:21 crc kubenswrapper[4824]: I1211 02:02:21.545353 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:21Z","lastTransitionTime":"2025-12-11T02:02:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:21 crc kubenswrapper[4824]: I1211 02:02:21.580576 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 02:02:21 crc kubenswrapper[4824]: E1211 02:02:21.580798 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 02:03:25.580775237 +0000 UTC m=+147.269812646 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:02:21 crc kubenswrapper[4824]: I1211 02:02:21.631845 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 02:02:21 crc kubenswrapper[4824]: I1211 02:02:21.631869 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 02:02:21 crc kubenswrapper[4824]: I1211 02:02:21.631910 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 02:02:21 crc kubenswrapper[4824]: I1211 02:02:21.631960 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-grnpt" Dec 11 02:02:21 crc kubenswrapper[4824]: E1211 02:02:21.632028 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 02:02:21 crc kubenswrapper[4824]: E1211 02:02:21.632189 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 02:02:21 crc kubenswrapper[4824]: E1211 02:02:21.632407 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-grnpt" podUID="09a52948-0404-45ff-85b1-51479268bf71" Dec 11 02:02:21 crc kubenswrapper[4824]: E1211 02:02:21.632597 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 02:02:21 crc kubenswrapper[4824]: I1211 02:02:21.649148 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:21 crc kubenswrapper[4824]: I1211 02:02:21.649203 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:21 crc kubenswrapper[4824]: I1211 02:02:21.649220 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:21 crc kubenswrapper[4824]: I1211 02:02:21.649243 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:21 crc kubenswrapper[4824]: I1211 02:02:21.649262 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:21Z","lastTransitionTime":"2025-12-11T02:02:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:21 crc kubenswrapper[4824]: I1211 02:02:21.681887 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 02:02:21 crc kubenswrapper[4824]: I1211 02:02:21.681946 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 02:02:21 crc kubenswrapper[4824]: I1211 02:02:21.681987 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 02:02:21 crc kubenswrapper[4824]: E1211 02:02:21.682038 4824 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 11 02:02:21 crc kubenswrapper[4824]: E1211 02:02:21.682210 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 11 02:02:21 crc kubenswrapper[4824]: I1211 02:02:21.682056 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 02:02:21 crc kubenswrapper[4824]: E1211 02:02:21.682248 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 11 02:02:21 crc kubenswrapper[4824]: E1211 02:02:21.682268 4824 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 02:02:21 crc kubenswrapper[4824]: E1211 02:02:21.682216 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-11 02:03:25.682178373 +0000 UTC m=+147.371215842 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 11 02:02:21 crc kubenswrapper[4824]: E1211 02:02:21.682358 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-11 02:03:25.682335287 +0000 UTC m=+147.371372706 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 02:02:21 crc kubenswrapper[4824]: E1211 02:02:21.682362 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 11 02:02:21 crc kubenswrapper[4824]: E1211 02:02:21.682403 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 11 02:02:21 crc kubenswrapper[4824]: E1211 02:02:21.682423 4824 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 02:02:21 crc kubenswrapper[4824]: E1211 02:02:21.682489 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-11 02:03:25.682464551 +0000 UTC m=+147.371501970 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 02:02:21 crc kubenswrapper[4824]: E1211 02:02:21.682214 4824 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 11 02:02:21 crc kubenswrapper[4824]: E1211 02:02:21.682613 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-11 02:03:25.682587514 +0000 UTC m=+147.371624933 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 11 02:02:21 crc kubenswrapper[4824]: I1211 02:02:21.752199 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:21 crc kubenswrapper[4824]: I1211 02:02:21.752260 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:21 crc kubenswrapper[4824]: I1211 02:02:21.752278 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:21 crc kubenswrapper[4824]: I1211 02:02:21.752302 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:21 crc kubenswrapper[4824]: I1211 02:02:21.752319 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:21Z","lastTransitionTime":"2025-12-11T02:02:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:21 crc kubenswrapper[4824]: I1211 02:02:21.855390 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:21 crc kubenswrapper[4824]: I1211 02:02:21.855466 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:21 crc kubenswrapper[4824]: I1211 02:02:21.855490 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:21 crc kubenswrapper[4824]: I1211 02:02:21.855523 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:21 crc kubenswrapper[4824]: I1211 02:02:21.855547 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:21Z","lastTransitionTime":"2025-12-11T02:02:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:21 crc kubenswrapper[4824]: I1211 02:02:21.958547 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:21 crc kubenswrapper[4824]: I1211 02:02:21.958616 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:21 crc kubenswrapper[4824]: I1211 02:02:21.958634 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:21 crc kubenswrapper[4824]: I1211 02:02:21.958660 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:21 crc kubenswrapper[4824]: I1211 02:02:21.958678 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:21Z","lastTransitionTime":"2025-12-11T02:02:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:22 crc kubenswrapper[4824]: I1211 02:02:22.061919 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:22 crc kubenswrapper[4824]: I1211 02:02:22.061987 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:22 crc kubenswrapper[4824]: I1211 02:02:22.062006 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:22 crc kubenswrapper[4824]: I1211 02:02:22.062030 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:22 crc kubenswrapper[4824]: I1211 02:02:22.062048 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:22Z","lastTransitionTime":"2025-12-11T02:02:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:22 crc kubenswrapper[4824]: I1211 02:02:22.165484 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:22 crc kubenswrapper[4824]: I1211 02:02:22.165525 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:22 crc kubenswrapper[4824]: I1211 02:02:22.165535 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:22 crc kubenswrapper[4824]: I1211 02:02:22.165552 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:22 crc kubenswrapper[4824]: I1211 02:02:22.165562 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:22Z","lastTransitionTime":"2025-12-11T02:02:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:22 crc kubenswrapper[4824]: I1211 02:02:22.231441 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:22 crc kubenswrapper[4824]: I1211 02:02:22.231496 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:22 crc kubenswrapper[4824]: I1211 02:02:22.231514 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:22 crc kubenswrapper[4824]: I1211 02:02:22.231585 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:22 crc kubenswrapper[4824]: I1211 02:02:22.231607 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:22Z","lastTransitionTime":"2025-12-11T02:02:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:22 crc kubenswrapper[4824]: E1211 02:02:22.253574 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:02:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:02:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:02:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:02:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d57bece6-6fc6-4a84-b8a8-e1678f9e50db\\\",\\\"systemUUID\\\":\\\"f76b32e4-2764-4219-bbf4-f7e56d76df3b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:22Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:22 crc kubenswrapper[4824]: I1211 02:02:22.258313 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:22 crc kubenswrapper[4824]: I1211 02:02:22.258367 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:22 crc kubenswrapper[4824]: I1211 02:02:22.258385 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:22 crc kubenswrapper[4824]: I1211 02:02:22.258410 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:22 crc kubenswrapper[4824]: I1211 02:02:22.258428 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:22Z","lastTransitionTime":"2025-12-11T02:02:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:22 crc kubenswrapper[4824]: E1211 02:02:22.280188 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:02:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:02:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:02:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:02:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d57bece6-6fc6-4a84-b8a8-e1678f9e50db\\\",\\\"systemUUID\\\":\\\"f76b32e4-2764-4219-bbf4-f7e56d76df3b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:22Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:22 crc kubenswrapper[4824]: I1211 02:02:22.284394 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:22 crc kubenswrapper[4824]: I1211 02:02:22.284444 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:22 crc kubenswrapper[4824]: I1211 02:02:22.284460 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:22 crc kubenswrapper[4824]: I1211 02:02:22.284483 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:22 crc kubenswrapper[4824]: I1211 02:02:22.284500 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:22Z","lastTransitionTime":"2025-12-11T02:02:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:22 crc kubenswrapper[4824]: E1211 02:02:22.304725 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:02:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:02:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:02:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:02:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d57bece6-6fc6-4a84-b8a8-e1678f9e50db\\\",\\\"systemUUID\\\":\\\"f76b32e4-2764-4219-bbf4-f7e56d76df3b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:22Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:22 crc kubenswrapper[4824]: I1211 02:02:22.310838 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:22 crc kubenswrapper[4824]: I1211 02:02:22.310886 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:22 crc kubenswrapper[4824]: I1211 02:02:22.310902 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:22 crc kubenswrapper[4824]: I1211 02:02:22.310932 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:22 crc kubenswrapper[4824]: I1211 02:02:22.310951 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:22Z","lastTransitionTime":"2025-12-11T02:02:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:22 crc kubenswrapper[4824]: E1211 02:02:22.328216 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:02:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:02:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:02:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:02:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d57bece6-6fc6-4a84-b8a8-e1678f9e50db\\\",\\\"systemUUID\\\":\\\"f76b32e4-2764-4219-bbf4-f7e56d76df3b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:22Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:22 crc kubenswrapper[4824]: I1211 02:02:22.332408 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:22 crc kubenswrapper[4824]: I1211 02:02:22.332447 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:22 crc kubenswrapper[4824]: I1211 02:02:22.332464 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:22 crc kubenswrapper[4824]: I1211 02:02:22.332486 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:22 crc kubenswrapper[4824]: I1211 02:02:22.332502 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:22Z","lastTransitionTime":"2025-12-11T02:02:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:22 crc kubenswrapper[4824]: E1211 02:02:22.351754 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:02:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:02:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:02:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:02:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d57bece6-6fc6-4a84-b8a8-e1678f9e50db\\\",\\\"systemUUID\\\":\\\"f76b32e4-2764-4219-bbf4-f7e56d76df3b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:22Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:22 crc kubenswrapper[4824]: E1211 02:02:22.352037 4824 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 11 02:02:22 crc kubenswrapper[4824]: I1211 02:02:22.355541 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:22 crc kubenswrapper[4824]: I1211 02:02:22.355625 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:22 crc kubenswrapper[4824]: I1211 02:02:22.355644 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:22 crc kubenswrapper[4824]: I1211 02:02:22.355673 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:22 crc kubenswrapper[4824]: I1211 02:02:22.355694 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:22Z","lastTransitionTime":"2025-12-11T02:02:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:22 crc kubenswrapper[4824]: I1211 02:02:22.459782 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:22 crc kubenswrapper[4824]: I1211 02:02:22.459845 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:22 crc kubenswrapper[4824]: I1211 02:02:22.459864 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:22 crc kubenswrapper[4824]: I1211 02:02:22.459891 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:22 crc kubenswrapper[4824]: I1211 02:02:22.459910 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:22Z","lastTransitionTime":"2025-12-11T02:02:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:22 crc kubenswrapper[4824]: I1211 02:02:22.562238 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:22 crc kubenswrapper[4824]: I1211 02:02:22.562291 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:22 crc kubenswrapper[4824]: I1211 02:02:22.562307 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:22 crc kubenswrapper[4824]: I1211 02:02:22.562330 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:22 crc kubenswrapper[4824]: I1211 02:02:22.562347 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:22Z","lastTransitionTime":"2025-12-11T02:02:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:22 crc kubenswrapper[4824]: I1211 02:02:22.664526 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:22 crc kubenswrapper[4824]: I1211 02:02:22.664926 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:22 crc kubenswrapper[4824]: I1211 02:02:22.665073 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:22 crc kubenswrapper[4824]: I1211 02:02:22.665275 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:22 crc kubenswrapper[4824]: I1211 02:02:22.665431 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:22Z","lastTransitionTime":"2025-12-11T02:02:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:22 crc kubenswrapper[4824]: I1211 02:02:22.768543 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:22 crc kubenswrapper[4824]: I1211 02:02:22.768614 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:22 crc kubenswrapper[4824]: I1211 02:02:22.768639 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:22 crc kubenswrapper[4824]: I1211 02:02:22.768668 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:22 crc kubenswrapper[4824]: I1211 02:02:22.768685 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:22Z","lastTransitionTime":"2025-12-11T02:02:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:22 crc kubenswrapper[4824]: I1211 02:02:22.872979 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:22 crc kubenswrapper[4824]: I1211 02:02:22.873448 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:22 crc kubenswrapper[4824]: I1211 02:02:22.873680 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:22 crc kubenswrapper[4824]: I1211 02:02:22.873882 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:22 crc kubenswrapper[4824]: I1211 02:02:22.874046 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:22Z","lastTransitionTime":"2025-12-11T02:02:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:22 crc kubenswrapper[4824]: I1211 02:02:22.976249 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:22 crc kubenswrapper[4824]: I1211 02:02:22.976291 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:22 crc kubenswrapper[4824]: I1211 02:02:22.976302 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:22 crc kubenswrapper[4824]: I1211 02:02:22.976317 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:22 crc kubenswrapper[4824]: I1211 02:02:22.976329 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:22Z","lastTransitionTime":"2025-12-11T02:02:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:23 crc kubenswrapper[4824]: I1211 02:02:23.079569 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:23 crc kubenswrapper[4824]: I1211 02:02:23.080179 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:23 crc kubenswrapper[4824]: I1211 02:02:23.080198 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:23 crc kubenswrapper[4824]: I1211 02:02:23.080214 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:23 crc kubenswrapper[4824]: I1211 02:02:23.080225 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:23Z","lastTransitionTime":"2025-12-11T02:02:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:23 crc kubenswrapper[4824]: I1211 02:02:23.182977 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:23 crc kubenswrapper[4824]: I1211 02:02:23.183048 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:23 crc kubenswrapper[4824]: I1211 02:02:23.183073 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:23 crc kubenswrapper[4824]: I1211 02:02:23.183105 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:23 crc kubenswrapper[4824]: I1211 02:02:23.183202 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:23Z","lastTransitionTime":"2025-12-11T02:02:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:23 crc kubenswrapper[4824]: I1211 02:02:23.286047 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:23 crc kubenswrapper[4824]: I1211 02:02:23.286081 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:23 crc kubenswrapper[4824]: I1211 02:02:23.286089 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:23 crc kubenswrapper[4824]: I1211 02:02:23.286143 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:23 crc kubenswrapper[4824]: I1211 02:02:23.286152 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:23Z","lastTransitionTime":"2025-12-11T02:02:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:23 crc kubenswrapper[4824]: I1211 02:02:23.388561 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:23 crc kubenswrapper[4824]: I1211 02:02:23.388608 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:23 crc kubenswrapper[4824]: I1211 02:02:23.388617 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:23 crc kubenswrapper[4824]: I1211 02:02:23.388630 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:23 crc kubenswrapper[4824]: I1211 02:02:23.388639 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:23Z","lastTransitionTime":"2025-12-11T02:02:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:23 crc kubenswrapper[4824]: I1211 02:02:23.491073 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:23 crc kubenswrapper[4824]: I1211 02:02:23.491137 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:23 crc kubenswrapper[4824]: I1211 02:02:23.491160 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:23 crc kubenswrapper[4824]: I1211 02:02:23.491180 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:23 crc kubenswrapper[4824]: I1211 02:02:23.491194 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:23Z","lastTransitionTime":"2025-12-11T02:02:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:23 crc kubenswrapper[4824]: I1211 02:02:23.593032 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:23 crc kubenswrapper[4824]: I1211 02:02:23.593087 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:23 crc kubenswrapper[4824]: I1211 02:02:23.593099 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:23 crc kubenswrapper[4824]: I1211 02:02:23.593138 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:23 crc kubenswrapper[4824]: I1211 02:02:23.593151 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:23Z","lastTransitionTime":"2025-12-11T02:02:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:23 crc kubenswrapper[4824]: I1211 02:02:23.632101 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-grnpt" Dec 11 02:02:23 crc kubenswrapper[4824]: I1211 02:02:23.632221 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 02:02:23 crc kubenswrapper[4824]: E1211 02:02:23.632254 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-grnpt" podUID="09a52948-0404-45ff-85b1-51479268bf71" Dec 11 02:02:23 crc kubenswrapper[4824]: I1211 02:02:23.632299 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 02:02:23 crc kubenswrapper[4824]: I1211 02:02:23.632356 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 02:02:23 crc kubenswrapper[4824]: E1211 02:02:23.632360 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 02:02:23 crc kubenswrapper[4824]: E1211 02:02:23.632438 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 02:02:23 crc kubenswrapper[4824]: E1211 02:02:23.632484 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 02:02:23 crc kubenswrapper[4824]: I1211 02:02:23.695425 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:23 crc kubenswrapper[4824]: I1211 02:02:23.695475 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:23 crc kubenswrapper[4824]: I1211 02:02:23.695488 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:23 crc kubenswrapper[4824]: I1211 02:02:23.695504 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:23 crc kubenswrapper[4824]: I1211 02:02:23.695516 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:23Z","lastTransitionTime":"2025-12-11T02:02:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:23 crc kubenswrapper[4824]: I1211 02:02:23.797358 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:23 crc kubenswrapper[4824]: I1211 02:02:23.797402 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:23 crc kubenswrapper[4824]: I1211 02:02:23.797412 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:23 crc kubenswrapper[4824]: I1211 02:02:23.797427 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:23 crc kubenswrapper[4824]: I1211 02:02:23.797438 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:23Z","lastTransitionTime":"2025-12-11T02:02:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:23 crc kubenswrapper[4824]: I1211 02:02:23.899519 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:23 crc kubenswrapper[4824]: I1211 02:02:23.899568 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:23 crc kubenswrapper[4824]: I1211 02:02:23.899583 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:23 crc kubenswrapper[4824]: I1211 02:02:23.899604 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:23 crc kubenswrapper[4824]: I1211 02:02:23.899618 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:23Z","lastTransitionTime":"2025-12-11T02:02:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:24 crc kubenswrapper[4824]: I1211 02:02:24.002261 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:24 crc kubenswrapper[4824]: I1211 02:02:24.002315 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:24 crc kubenswrapper[4824]: I1211 02:02:24.002327 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:24 crc kubenswrapper[4824]: I1211 02:02:24.002347 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:24 crc kubenswrapper[4824]: I1211 02:02:24.002371 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:24Z","lastTransitionTime":"2025-12-11T02:02:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:24 crc kubenswrapper[4824]: I1211 02:02:24.104406 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:24 crc kubenswrapper[4824]: I1211 02:02:24.104450 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:24 crc kubenswrapper[4824]: I1211 02:02:24.104464 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:24 crc kubenswrapper[4824]: I1211 02:02:24.104483 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:24 crc kubenswrapper[4824]: I1211 02:02:24.104498 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:24Z","lastTransitionTime":"2025-12-11T02:02:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:24 crc kubenswrapper[4824]: I1211 02:02:24.207377 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:24 crc kubenswrapper[4824]: I1211 02:02:24.207444 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:24 crc kubenswrapper[4824]: I1211 02:02:24.207466 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:24 crc kubenswrapper[4824]: I1211 02:02:24.207495 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:24 crc kubenswrapper[4824]: I1211 02:02:24.207516 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:24Z","lastTransitionTime":"2025-12-11T02:02:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:24 crc kubenswrapper[4824]: I1211 02:02:24.309597 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:24 crc kubenswrapper[4824]: I1211 02:02:24.309665 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:24 crc kubenswrapper[4824]: I1211 02:02:24.309684 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:24 crc kubenswrapper[4824]: I1211 02:02:24.309709 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:24 crc kubenswrapper[4824]: I1211 02:02:24.309729 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:24Z","lastTransitionTime":"2025-12-11T02:02:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:24 crc kubenswrapper[4824]: I1211 02:02:24.411999 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:24 crc kubenswrapper[4824]: I1211 02:02:24.412072 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:24 crc kubenswrapper[4824]: I1211 02:02:24.412092 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:24 crc kubenswrapper[4824]: I1211 02:02:24.412146 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:24 crc kubenswrapper[4824]: I1211 02:02:24.412166 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:24Z","lastTransitionTime":"2025-12-11T02:02:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:24 crc kubenswrapper[4824]: I1211 02:02:24.515195 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:24 crc kubenswrapper[4824]: I1211 02:02:24.515243 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:24 crc kubenswrapper[4824]: I1211 02:02:24.515257 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:24 crc kubenswrapper[4824]: I1211 02:02:24.515280 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:24 crc kubenswrapper[4824]: I1211 02:02:24.515294 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:24Z","lastTransitionTime":"2025-12-11T02:02:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:24 crc kubenswrapper[4824]: I1211 02:02:24.618343 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:24 crc kubenswrapper[4824]: I1211 02:02:24.618402 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:24 crc kubenswrapper[4824]: I1211 02:02:24.618423 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:24 crc kubenswrapper[4824]: I1211 02:02:24.618446 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:24 crc kubenswrapper[4824]: I1211 02:02:24.618463 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:24Z","lastTransitionTime":"2025-12-11T02:02:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:24 crc kubenswrapper[4824]: I1211 02:02:24.632382 4824 scope.go:117] "RemoveContainer" containerID="154d82fd6acee749bfd37220547e52647dc7f9d86a73c28ce7d8577c9cadef6e" Dec 11 02:02:24 crc kubenswrapper[4824]: E1211 02:02:24.632571 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-f6pgc_openshift-ovn-kubernetes(5b193bd2-b536-4056-92f1-94c9836ab2eb)\"" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" podUID="5b193bd2-b536-4056-92f1-94c9836ab2eb" Dec 11 02:02:24 crc kubenswrapper[4824]: I1211 02:02:24.721030 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:24 crc kubenswrapper[4824]: I1211 02:02:24.721084 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:24 crc kubenswrapper[4824]: I1211 02:02:24.721100 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:24 crc kubenswrapper[4824]: I1211 02:02:24.721162 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:24 crc kubenswrapper[4824]: I1211 02:02:24.721183 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:24Z","lastTransitionTime":"2025-12-11T02:02:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:24 crc kubenswrapper[4824]: I1211 02:02:24.823827 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:24 crc kubenswrapper[4824]: I1211 02:02:24.823878 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:24 crc kubenswrapper[4824]: I1211 02:02:24.823889 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:24 crc kubenswrapper[4824]: I1211 02:02:24.823905 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:24 crc kubenswrapper[4824]: I1211 02:02:24.823917 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:24Z","lastTransitionTime":"2025-12-11T02:02:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:24 crc kubenswrapper[4824]: I1211 02:02:24.926308 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:24 crc kubenswrapper[4824]: I1211 02:02:24.926348 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:24 crc kubenswrapper[4824]: I1211 02:02:24.926359 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:24 crc kubenswrapper[4824]: I1211 02:02:24.926375 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:24 crc kubenswrapper[4824]: I1211 02:02:24.926386 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:24Z","lastTransitionTime":"2025-12-11T02:02:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:25 crc kubenswrapper[4824]: I1211 02:02:25.029193 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:25 crc kubenswrapper[4824]: I1211 02:02:25.029242 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:25 crc kubenswrapper[4824]: I1211 02:02:25.029257 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:25 crc kubenswrapper[4824]: I1211 02:02:25.029275 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:25 crc kubenswrapper[4824]: I1211 02:02:25.029290 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:25Z","lastTransitionTime":"2025-12-11T02:02:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:25 crc kubenswrapper[4824]: I1211 02:02:25.131739 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:25 crc kubenswrapper[4824]: I1211 02:02:25.131781 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:25 crc kubenswrapper[4824]: I1211 02:02:25.131793 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:25 crc kubenswrapper[4824]: I1211 02:02:25.131807 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:25 crc kubenswrapper[4824]: I1211 02:02:25.131818 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:25Z","lastTransitionTime":"2025-12-11T02:02:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:25 crc kubenswrapper[4824]: I1211 02:02:25.235000 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:25 crc kubenswrapper[4824]: I1211 02:02:25.235081 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:25 crc kubenswrapper[4824]: I1211 02:02:25.235150 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:25 crc kubenswrapper[4824]: I1211 02:02:25.235183 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:25 crc kubenswrapper[4824]: I1211 02:02:25.235207 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:25Z","lastTransitionTime":"2025-12-11T02:02:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:25 crc kubenswrapper[4824]: I1211 02:02:25.338167 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:25 crc kubenswrapper[4824]: I1211 02:02:25.338232 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:25 crc kubenswrapper[4824]: I1211 02:02:25.338250 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:25 crc kubenswrapper[4824]: I1211 02:02:25.338273 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:25 crc kubenswrapper[4824]: I1211 02:02:25.338290 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:25Z","lastTransitionTime":"2025-12-11T02:02:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:25 crc kubenswrapper[4824]: I1211 02:02:25.441836 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:25 crc kubenswrapper[4824]: I1211 02:02:25.441895 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:25 crc kubenswrapper[4824]: I1211 02:02:25.441915 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:25 crc kubenswrapper[4824]: I1211 02:02:25.441946 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:25 crc kubenswrapper[4824]: I1211 02:02:25.441971 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:25Z","lastTransitionTime":"2025-12-11T02:02:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:25 crc kubenswrapper[4824]: I1211 02:02:25.544520 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:25 crc kubenswrapper[4824]: I1211 02:02:25.544624 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:25 crc kubenswrapper[4824]: I1211 02:02:25.544650 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:25 crc kubenswrapper[4824]: I1211 02:02:25.544680 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:25 crc kubenswrapper[4824]: I1211 02:02:25.544705 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:25Z","lastTransitionTime":"2025-12-11T02:02:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:25 crc kubenswrapper[4824]: I1211 02:02:25.631723 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 02:02:25 crc kubenswrapper[4824]: I1211 02:02:25.631762 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 02:02:25 crc kubenswrapper[4824]: I1211 02:02:25.631854 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 02:02:25 crc kubenswrapper[4824]: I1211 02:02:25.632003 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-grnpt" Dec 11 02:02:25 crc kubenswrapper[4824]: E1211 02:02:25.631994 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 02:02:25 crc kubenswrapper[4824]: E1211 02:02:25.632059 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 02:02:25 crc kubenswrapper[4824]: E1211 02:02:25.632138 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-grnpt" podUID="09a52948-0404-45ff-85b1-51479268bf71" Dec 11 02:02:25 crc kubenswrapper[4824]: E1211 02:02:25.632189 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 02:02:25 crc kubenswrapper[4824]: I1211 02:02:25.647397 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:25 crc kubenswrapper[4824]: I1211 02:02:25.647468 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:25 crc kubenswrapper[4824]: I1211 02:02:25.647485 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:25 crc kubenswrapper[4824]: I1211 02:02:25.647510 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:25 crc kubenswrapper[4824]: I1211 02:02:25.647529 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:25Z","lastTransitionTime":"2025-12-11T02:02:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:25 crc kubenswrapper[4824]: I1211 02:02:25.749441 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:25 crc kubenswrapper[4824]: I1211 02:02:25.749485 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:25 crc kubenswrapper[4824]: I1211 02:02:25.749504 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:25 crc kubenswrapper[4824]: I1211 02:02:25.749524 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:25 crc kubenswrapper[4824]: I1211 02:02:25.749541 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:25Z","lastTransitionTime":"2025-12-11T02:02:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:25 crc kubenswrapper[4824]: I1211 02:02:25.852239 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:25 crc kubenswrapper[4824]: I1211 02:02:25.852288 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:25 crc kubenswrapper[4824]: I1211 02:02:25.852302 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:25 crc kubenswrapper[4824]: I1211 02:02:25.852320 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:25 crc kubenswrapper[4824]: I1211 02:02:25.852334 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:25Z","lastTransitionTime":"2025-12-11T02:02:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:25 crc kubenswrapper[4824]: I1211 02:02:25.954152 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:25 crc kubenswrapper[4824]: I1211 02:02:25.954178 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:25 crc kubenswrapper[4824]: I1211 02:02:25.954187 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:25 crc kubenswrapper[4824]: I1211 02:02:25.954197 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:25 crc kubenswrapper[4824]: I1211 02:02:25.954205 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:25Z","lastTransitionTime":"2025-12-11T02:02:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:26 crc kubenswrapper[4824]: I1211 02:02:26.057206 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:26 crc kubenswrapper[4824]: I1211 02:02:26.057259 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:26 crc kubenswrapper[4824]: I1211 02:02:26.057280 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:26 crc kubenswrapper[4824]: I1211 02:02:26.057304 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:26 crc kubenswrapper[4824]: I1211 02:02:26.057321 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:26Z","lastTransitionTime":"2025-12-11T02:02:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:26 crc kubenswrapper[4824]: I1211 02:02:26.161256 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:26 crc kubenswrapper[4824]: I1211 02:02:26.161317 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:26 crc kubenswrapper[4824]: I1211 02:02:26.161351 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:26 crc kubenswrapper[4824]: I1211 02:02:26.161390 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:26 crc kubenswrapper[4824]: I1211 02:02:26.161411 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:26Z","lastTransitionTime":"2025-12-11T02:02:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:26 crc kubenswrapper[4824]: I1211 02:02:26.264520 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:26 crc kubenswrapper[4824]: I1211 02:02:26.264594 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:26 crc kubenswrapper[4824]: I1211 02:02:26.264617 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:26 crc kubenswrapper[4824]: I1211 02:02:26.264760 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:26 crc kubenswrapper[4824]: I1211 02:02:26.264844 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:26Z","lastTransitionTime":"2025-12-11T02:02:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:26 crc kubenswrapper[4824]: I1211 02:02:26.367702 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:26 crc kubenswrapper[4824]: I1211 02:02:26.367765 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:26 crc kubenswrapper[4824]: I1211 02:02:26.367782 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:26 crc kubenswrapper[4824]: I1211 02:02:26.367805 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:26 crc kubenswrapper[4824]: I1211 02:02:26.367824 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:26Z","lastTransitionTime":"2025-12-11T02:02:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:26 crc kubenswrapper[4824]: I1211 02:02:26.470965 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:26 crc kubenswrapper[4824]: I1211 02:02:26.471027 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:26 crc kubenswrapper[4824]: I1211 02:02:26.471047 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:26 crc kubenswrapper[4824]: I1211 02:02:26.471072 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:26 crc kubenswrapper[4824]: I1211 02:02:26.471089 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:26Z","lastTransitionTime":"2025-12-11T02:02:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:26 crc kubenswrapper[4824]: I1211 02:02:26.574935 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:26 crc kubenswrapper[4824]: I1211 02:02:26.575003 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:26 crc kubenswrapper[4824]: I1211 02:02:26.575012 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:26 crc kubenswrapper[4824]: I1211 02:02:26.575034 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:26 crc kubenswrapper[4824]: I1211 02:02:26.575045 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:26Z","lastTransitionTime":"2025-12-11T02:02:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:26 crc kubenswrapper[4824]: I1211 02:02:26.678316 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:26 crc kubenswrapper[4824]: I1211 02:02:26.678374 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:26 crc kubenswrapper[4824]: I1211 02:02:26.678391 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:26 crc kubenswrapper[4824]: I1211 02:02:26.678415 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:26 crc kubenswrapper[4824]: I1211 02:02:26.678433 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:26Z","lastTransitionTime":"2025-12-11T02:02:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:26 crc kubenswrapper[4824]: I1211 02:02:26.781675 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:26 crc kubenswrapper[4824]: I1211 02:02:26.781730 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:26 crc kubenswrapper[4824]: I1211 02:02:26.781746 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:26 crc kubenswrapper[4824]: I1211 02:02:26.781772 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:26 crc kubenswrapper[4824]: I1211 02:02:26.781789 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:26Z","lastTransitionTime":"2025-12-11T02:02:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:26 crc kubenswrapper[4824]: I1211 02:02:26.884843 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:26 crc kubenswrapper[4824]: I1211 02:02:26.884896 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:26 crc kubenswrapper[4824]: I1211 02:02:26.884913 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:26 crc kubenswrapper[4824]: I1211 02:02:26.884938 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:26 crc kubenswrapper[4824]: I1211 02:02:26.884959 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:26Z","lastTransitionTime":"2025-12-11T02:02:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:26 crc kubenswrapper[4824]: I1211 02:02:26.987990 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:26 crc kubenswrapper[4824]: I1211 02:02:26.988080 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:26 crc kubenswrapper[4824]: I1211 02:02:26.988102 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:26 crc kubenswrapper[4824]: I1211 02:02:26.988155 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:26 crc kubenswrapper[4824]: I1211 02:02:26.988173 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:26Z","lastTransitionTime":"2025-12-11T02:02:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:27 crc kubenswrapper[4824]: I1211 02:02:27.091442 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:27 crc kubenswrapper[4824]: I1211 02:02:27.091500 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:27 crc kubenswrapper[4824]: I1211 02:02:27.091568 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:27 crc kubenswrapper[4824]: I1211 02:02:27.091602 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:27 crc kubenswrapper[4824]: I1211 02:02:27.091625 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:27Z","lastTransitionTime":"2025-12-11T02:02:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:27 crc kubenswrapper[4824]: I1211 02:02:27.195026 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:27 crc kubenswrapper[4824]: I1211 02:02:27.195103 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:27 crc kubenswrapper[4824]: I1211 02:02:27.195160 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:27 crc kubenswrapper[4824]: I1211 02:02:27.195188 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:27 crc kubenswrapper[4824]: I1211 02:02:27.195205 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:27Z","lastTransitionTime":"2025-12-11T02:02:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:27 crc kubenswrapper[4824]: I1211 02:02:27.297976 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:27 crc kubenswrapper[4824]: I1211 02:02:27.298425 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:27 crc kubenswrapper[4824]: I1211 02:02:27.298561 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:27 crc kubenswrapper[4824]: I1211 02:02:27.298694 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:27 crc kubenswrapper[4824]: I1211 02:02:27.298837 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:27Z","lastTransitionTime":"2025-12-11T02:02:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:27 crc kubenswrapper[4824]: I1211 02:02:27.402262 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:27 crc kubenswrapper[4824]: I1211 02:02:27.402326 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:27 crc kubenswrapper[4824]: I1211 02:02:27.402344 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:27 crc kubenswrapper[4824]: I1211 02:02:27.402369 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:27 crc kubenswrapper[4824]: I1211 02:02:27.402387 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:27Z","lastTransitionTime":"2025-12-11T02:02:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:27 crc kubenswrapper[4824]: I1211 02:02:27.505105 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:27 crc kubenswrapper[4824]: I1211 02:02:27.505182 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:27 crc kubenswrapper[4824]: I1211 02:02:27.505193 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:27 crc kubenswrapper[4824]: I1211 02:02:27.505209 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:27 crc kubenswrapper[4824]: I1211 02:02:27.505222 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:27Z","lastTransitionTime":"2025-12-11T02:02:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:27 crc kubenswrapper[4824]: I1211 02:02:27.608425 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:27 crc kubenswrapper[4824]: I1211 02:02:27.608514 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:27 crc kubenswrapper[4824]: I1211 02:02:27.608538 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:27 crc kubenswrapper[4824]: I1211 02:02:27.608568 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:27 crc kubenswrapper[4824]: I1211 02:02:27.608592 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:27Z","lastTransitionTime":"2025-12-11T02:02:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:27 crc kubenswrapper[4824]: I1211 02:02:27.632242 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-grnpt" Dec 11 02:02:27 crc kubenswrapper[4824]: I1211 02:02:27.632327 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 02:02:27 crc kubenswrapper[4824]: I1211 02:02:27.632346 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 02:02:27 crc kubenswrapper[4824]: I1211 02:02:27.632254 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 02:02:27 crc kubenswrapper[4824]: E1211 02:02:27.632460 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-grnpt" podUID="09a52948-0404-45ff-85b1-51479268bf71" Dec 11 02:02:27 crc kubenswrapper[4824]: E1211 02:02:27.632648 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 02:02:27 crc kubenswrapper[4824]: E1211 02:02:27.632773 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 02:02:27 crc kubenswrapper[4824]: E1211 02:02:27.632894 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 02:02:27 crc kubenswrapper[4824]: I1211 02:02:27.711307 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:27 crc kubenswrapper[4824]: I1211 02:02:27.711352 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:27 crc kubenswrapper[4824]: I1211 02:02:27.711367 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:27 crc kubenswrapper[4824]: I1211 02:02:27.711432 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:27 crc kubenswrapper[4824]: I1211 02:02:27.711448 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:27Z","lastTransitionTime":"2025-12-11T02:02:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:27 crc kubenswrapper[4824]: I1211 02:02:27.813332 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:27 crc kubenswrapper[4824]: I1211 02:02:27.813384 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:27 crc kubenswrapper[4824]: I1211 02:02:27.813405 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:27 crc kubenswrapper[4824]: I1211 02:02:27.813445 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:27 crc kubenswrapper[4824]: I1211 02:02:27.813457 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:27Z","lastTransitionTime":"2025-12-11T02:02:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:27 crc kubenswrapper[4824]: I1211 02:02:27.915349 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:27 crc kubenswrapper[4824]: I1211 02:02:27.915453 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:27 crc kubenswrapper[4824]: I1211 02:02:27.915472 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:27 crc kubenswrapper[4824]: I1211 02:02:27.915495 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:27 crc kubenswrapper[4824]: I1211 02:02:27.915553 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:27Z","lastTransitionTime":"2025-12-11T02:02:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:28 crc kubenswrapper[4824]: I1211 02:02:28.018956 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:28 crc kubenswrapper[4824]: I1211 02:02:28.019025 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:28 crc kubenswrapper[4824]: I1211 02:02:28.019041 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:28 crc kubenswrapper[4824]: I1211 02:02:28.019067 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:28 crc kubenswrapper[4824]: I1211 02:02:28.019087 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:28Z","lastTransitionTime":"2025-12-11T02:02:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:28 crc kubenswrapper[4824]: I1211 02:02:28.121581 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:28 crc kubenswrapper[4824]: I1211 02:02:28.121640 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:28 crc kubenswrapper[4824]: I1211 02:02:28.121657 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:28 crc kubenswrapper[4824]: I1211 02:02:28.121681 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:28 crc kubenswrapper[4824]: I1211 02:02:28.121696 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:28Z","lastTransitionTime":"2025-12-11T02:02:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:28 crc kubenswrapper[4824]: I1211 02:02:28.225001 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:28 crc kubenswrapper[4824]: I1211 02:02:28.225066 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:28 crc kubenswrapper[4824]: I1211 02:02:28.225099 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:28 crc kubenswrapper[4824]: I1211 02:02:28.225135 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:28 crc kubenswrapper[4824]: I1211 02:02:28.225149 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:28Z","lastTransitionTime":"2025-12-11T02:02:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:28 crc kubenswrapper[4824]: I1211 02:02:28.328217 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:28 crc kubenswrapper[4824]: I1211 02:02:28.328286 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:28 crc kubenswrapper[4824]: I1211 02:02:28.328306 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:28 crc kubenswrapper[4824]: I1211 02:02:28.328331 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:28 crc kubenswrapper[4824]: I1211 02:02:28.328349 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:28Z","lastTransitionTime":"2025-12-11T02:02:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:28 crc kubenswrapper[4824]: I1211 02:02:28.431180 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:28 crc kubenswrapper[4824]: I1211 02:02:28.431239 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:28 crc kubenswrapper[4824]: I1211 02:02:28.431256 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:28 crc kubenswrapper[4824]: I1211 02:02:28.431280 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:28 crc kubenswrapper[4824]: I1211 02:02:28.431303 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:28Z","lastTransitionTime":"2025-12-11T02:02:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:28 crc kubenswrapper[4824]: I1211 02:02:28.534648 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:28 crc kubenswrapper[4824]: I1211 02:02:28.534723 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:28 crc kubenswrapper[4824]: I1211 02:02:28.534751 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:28 crc kubenswrapper[4824]: I1211 02:02:28.534782 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:28 crc kubenswrapper[4824]: I1211 02:02:28.534806 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:28Z","lastTransitionTime":"2025-12-11T02:02:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:28 crc kubenswrapper[4824]: I1211 02:02:28.638393 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:28 crc kubenswrapper[4824]: I1211 02:02:28.638497 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:28 crc kubenswrapper[4824]: I1211 02:02:28.638520 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:28 crc kubenswrapper[4824]: I1211 02:02:28.638545 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:28 crc kubenswrapper[4824]: I1211 02:02:28.638563 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:28Z","lastTransitionTime":"2025-12-11T02:02:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:28 crc kubenswrapper[4824]: I1211 02:02:28.665484 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b193bd2-b536-4056-92f1-94c9836ab2eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51f3c41d86ddc44f7eca118029876ad353317f753d0ad173ed4d30c2a15f10e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c345850891744a0240781277e9a89a59f6899c4c944995f62e44650f807440a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed2ae40b1fe28111efc76d9d72345edecb5f3ea34522b238f3730812b29f2bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce7aa078800e6f299cfe03770fa7e3b64fa3742fe0c0f383c87d84cf48569ddf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6aa9fc1f1bff192c69bd59d32a05d4a72d09641d9ce4b265bc4517147a7e15dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b4a3bc50eb80777188206adab774e8acad8b2e4952a8415f91b1c29ba9a0ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://154d82fd6acee749bfd37220547e52647dc7f9d86a73c28ce7d8577c9cadef6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://154d82fd6acee749bfd37220547e52647dc7f9d86a73c28ce7d8577c9cadef6e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T02:02:10Z\\\",\\\"message\\\":\\\"t handler 9 for removal\\\\nI1211 02:02:10.678557 6803 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1211 02:02:10.678594 6803 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1211 02:02:10.678724 6803 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1211 02:02:10.678871 6803 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1211 02:02:10.678731 6803 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1211 02:02:10.678906 6803 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1211 02:02:10.678920 6803 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1211 02:02:10.678935 6803 handler.go:208] Removed *v1.Node event handler 7\\\\nI1211 02:02:10.678961 6803 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1211 02:02:10.678945 6803 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1211 02:02:10.678980 6803 handler.go:208] Removed *v1.Node event handler 2\\\\nI1211 02:02:10.678982 6803 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1211 02:02:10.679043 6803 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1211 02:02:10.679161 6803 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1211 02:02:10.679264 6803 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T02:02:09Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-f6pgc_openshift-ovn-kubernetes(5b193bd2-b536-4056-92f1-94c9836ab2eb)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ead58694813806d307bbbc6a7b3cf41730d1c059b231a3002684bf09c10c7f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f6pgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:28Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:28 crc kubenswrapper[4824]: I1211 02:02:28.686414 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-br4j9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce125194-9c9d-4ea4-8a72-2c30d5722295\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9798f904052e08196a0db1745060dab11e2710308bca01ff160475910f2665df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fc5s8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-br4j9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:28Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:28 crc kubenswrapper[4824]: I1211 02:02:28.706434 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8v4jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99a6bfee-afca-4726-a964-b95a948945a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b23b9d8f18db3eeadb60f83b36b05b1778cd0fa31669dad2680bb335ae004acf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4hk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3acc1940a0534ba0e6096175e41834fdd0fbcbb4c3fb6e9a38e76aedae55d2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4hk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-8v4jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:28Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:28 crc kubenswrapper[4824]: I1211 02:02:28.724305 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-grnpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09a52948-0404-45ff-85b1-51479268bf71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmmzm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmmzm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-grnpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:28Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:28 crc kubenswrapper[4824]: I1211 02:02:28.745797 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:28 crc kubenswrapper[4824]: I1211 02:02:28.745860 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:28 crc kubenswrapper[4824]: I1211 02:02:28.745878 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:28 crc kubenswrapper[4824]: I1211 02:02:28.745901 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:28 crc kubenswrapper[4824]: I1211 02:02:28.745918 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:28Z","lastTransitionTime":"2025-12-11T02:02:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:28 crc kubenswrapper[4824]: I1211 02:02:28.750019 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:28Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:28 crc kubenswrapper[4824]: I1211 02:02:28.779609 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kh8g9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"50c69d41-a65b-463e-a829-ea349bd2fe46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dca6be41bd7a3ce12ecd4567c868f696b65e8eb1bad436feda6e4f9bfb8e525\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69af1f33fd0bafc929b94e702b4f46c6c7d7a362d5552b121729c477892d91d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69af1f33fd0bafc929b94e702b4f46c6c7d7a362d5552b121729c477892d91d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd9675d0afded587bc660a3862fd0c5dd6885961afacda957aa9ee8501d68a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd9675d0afded587bc660a3862fd0c5dd6885961afacda957aa9ee8501d68a69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aac94149458bd6321be90b26b48ee26428d273e7f61407d95f9c9d6b3aeae4b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aac94149458bd6321be90b26b48ee26428d273e7f61407d95f9c9d6b3aeae4b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbee11bd7cb65e031415990869b52d2b8f81bf8b56e6cc935d38e09695db7b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbee11bd7cb65e031415990869b52d2b8f81bf8b56e6cc935d38e09695db7b48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86f693ade72bd593bc16021fe4b7564f1914ec2f1cda7039474ac2fb13de10f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86f693ade72bd593bc16021fe4b7564f1914ec2f1cda7039474ac2fb13de10f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7d7674252f67bde0ded2ec8ad1d925a92a26d3e3f741dee0061b735cb0b7a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7d7674252f67bde0ded2ec8ad1d925a92a26d3e3f741dee0061b735cb0b7a02\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qdtz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kh8g9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:28Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:28 crc kubenswrapper[4824]: I1211 02:02:28.798289 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6bkc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8022f696-d0e7-437b-848a-3cd25bd1f364\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f40719f59881e077e7edbab3adb9efd101359bd461e5434c365e119e8cdbddb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3e28f6778ec0dba6ea3e6b69e5a2c1c73c0d5db1abeead241fb58efdbfaefe4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T02:02:03Z\\\",\\\"message\\\":\\\"2025-12-11T02:01:18+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_00fc4257-9dfb-4131-a927-0fc54bfa1304\\\\n2025-12-11T02:01:18+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_00fc4257-9dfb-4131-a927-0fc54bfa1304 to /host/opt/cni/bin/\\\\n2025-12-11T02:01:18Z [verbose] multus-daemon started\\\\n2025-12-11T02:01:18Z [verbose] Readiness Indicator file check\\\\n2025-12-11T02:02:03Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:02:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hh5kn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6bkc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:28Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:28 crc kubenswrapper[4824]: I1211 02:02:28.817672 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d14986666febd651039afb52ddf4db8969aca55e88eda9bbb20a67e35ba722ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:28Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:28 crc kubenswrapper[4824]: I1211 02:02:28.835385 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:28Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:28 crc kubenswrapper[4824]: I1211 02:02:28.849632 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:28 crc kubenswrapper[4824]: I1211 02:02:28.849699 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:28 crc kubenswrapper[4824]: I1211 02:02:28.849723 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:28 crc kubenswrapper[4824]: I1211 02:02:28.849749 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:28 crc kubenswrapper[4824]: I1211 02:02:28.849767 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:28Z","lastTransitionTime":"2025-12-11T02:02:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:28 crc kubenswrapper[4824]: I1211 02:02:28.852065 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44b156e4-64a4-4d45-aa5e-9b10a862faed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7368d92144b188e95a7796fabf497b0f8d9c0b240293bdc7b4aec1ff75497bc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w62dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcf289ac27dbda4f24aabbde3fb5119329e7b4fc7d6e018168da9a69661a4c57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w62dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gx6xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:28Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:28 crc kubenswrapper[4824]: I1211 02:02:28.872873 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:28Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:28 crc kubenswrapper[4824]: I1211 02:02:28.886364 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c546a836-56e8-4825-993c-24fd0ed63039\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51533689fc74e35e0db9e2ef1b51a225ae3bbd49669afacce01841b3343cc658\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1c9f0af0451936c800a4b5e764759b7c3b07a957566188e7a7166d1ecf0cf4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f1c9f0af0451936c800a4b5e764759b7c3b07a957566188e7a7166d1ecf0cf4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:00:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:28Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:28 crc kubenswrapper[4824]: I1211 02:02:28.917597 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"253d4291-2b2f-487c-8862-3d305f0bdc55\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44b29bb7a6eda7cb890d36f86375b5f3f197ae5cf390f16e71139b0904873e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://922af7f67b99a0ce08776f3f75c0c377e6ea69af1720b6178a6d1e2f9b4b8ba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://136c68020325db8e6930d0f40b0d0d7e013c14c4dd5e7d1449d01d6d458104d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c89afaf56b4a9597f6543ec2e3b176c7a6dab7dc341941fc3bc8ec85cc4edaf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:00:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:28Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:28 crc kubenswrapper[4824]: I1211 02:02:28.952040 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:28 crc kubenswrapper[4824]: I1211 02:02:28.952089 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:28 crc kubenswrapper[4824]: I1211 02:02:28.952100 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:28 crc kubenswrapper[4824]: I1211 02:02:28.952134 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:28 crc kubenswrapper[4824]: I1211 02:02:28.952149 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:28Z","lastTransitionTime":"2025-12-11T02:02:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:28 crc kubenswrapper[4824]: I1211 02:02:28.965456 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98027cfa-5dde-4297-9f74-c4d662c7f82b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f410073059c5898660c50f775a270bebe51836f73e2a2f02db38a2aabf15931d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://335a774b8a3aff18695de4aba0b522d346aecc8ed8d96a80150301f5debce32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://339d619250c1c0a12f988a9b0ebc03874db3ac41a9d3b3f8573795467578cd0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e27af6c438009a44288f8bc10f8ebf2b61d9309f87e8d955e8d255a7190174f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e27af6c438009a44288f8bc10f8ebf2b61d9309f87e8d955e8d255a7190174f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:00:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:28Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:28 crc kubenswrapper[4824]: I1211 02:02:28.980243 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76b10bbd1e891b8c682f4b2e17d023fd787a8ed44ed94fbf7d97314d641ccf36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f32838f33832b9b67e7dd9c53340680e17721cf1c7bdf0170230c5f8feafe64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:28Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:28 crc kubenswrapper[4824]: I1211 02:02:28.989340 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kptws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68568dba-83a5-46c1-8d01-337359679d91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f965a6d6db619995132f61b5d51b66f0b2b3fdbef3fbec63e231b9a62ccd0a9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd7nf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:01:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kptws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:28Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:29 crc kubenswrapper[4824]: I1211 02:02:29.012529 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afca157c-af3d-4197-8a60-39fc607efd69\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bae1b2b71f76385e3144a18a4335f125adaf2b40a59e9f8c353ab2fd5efd5549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09587408482a9771161017d1246b051cb9ad2961bd949414210b3c0ad58ae3fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e42434f5fe6b847211ad32b3fee7e6a6a6adbe9c5caa30235ac07a25507f7871\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af95eed3f8b150973c8f151d41b7a8ce2c41d90ede17bf0f027b721f4f3858d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0426825294b0be03de99d39e8a1584f7717017b99d70a12af858ff8ea0699302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18a2507c91dd7ab306ef38abf724e07f468190caba6cb110bf3224448f631aad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18a2507c91dd7ab306ef38abf724e07f468190caba6cb110bf3224448f631aad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be8f3f2460ba45c67ac7c6e7eac2d6ada631e90f4c7de645a41c134539437f82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be8f3f2460ba45c67ac7c6e7eac2d6ada631e90f4c7de645a41c134539437f82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://016b733da5b0e8f0e3830ee9746c4ab5decfcac7573f42aec257a842a393c60d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://016b733da5b0e8f0e3830ee9746c4ab5decfcac7573f42aec257a842a393c60d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:01:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:00:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:29Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:29 crc kubenswrapper[4824]: I1211 02:02:29.032461 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"beab386b-6c65-4bc3-9ef0-dc3f68b4403f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T02:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c26f06d2ed5e6e98b61a51f073130675f2c49acda106b8aedcbb2a6d4f7580c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20af9b10a67c8b21a7abcd23e42bc412798d159c8422e7c8ff58f5b997ca19a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9bbfc6edfe050b3fa9dbeee5f98d416bcb2a1533a38608a881b75396fd0b059\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7c766f8de73ca762ffde0c18b2d1c46bb152b2b37535a611c5a0e454a21b7d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d45819f0611c91b20c83eeee1092031a3441fb69f3015241632ad419880227d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T02:01:17Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 02:01:11.070250 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 02:01:11.071673 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2752882565/tls.crt::/tmp/serving-cert-2752882565/tls.key\\\\\\\"\\\\nI1211 02:01:17.042610 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 02:01:17.048046 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 02:01:17.048080 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 02:01:17.048172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 02:01:17.048188 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 02:01:17.058248 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 02:01:17.058294 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 02:01:17.058303 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 02:01:17.058315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 02:01:17.058322 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 02:01:17.058328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 02:01:17.058335 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 02:01:17.058808 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 02:01:17.060225 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14a940a72b2404a6cbd52a69e150b03266bcd7f5c1d3f4d07b51c0c69e51786\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T02:00:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T02:00:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T02:00:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:29Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:29 crc kubenswrapper[4824]: I1211 02:02:29.054796 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T02:01:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ded146c045085be2b58821749052413a36850ede00ae1024fe2dae9fa5ac84b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T02:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:29Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:29 crc kubenswrapper[4824]: I1211 02:02:29.055093 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:29 crc kubenswrapper[4824]: I1211 02:02:29.055152 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:29 crc kubenswrapper[4824]: I1211 02:02:29.055164 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:29 crc kubenswrapper[4824]: I1211 02:02:29.055182 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:29 crc kubenswrapper[4824]: I1211 02:02:29.055193 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:29Z","lastTransitionTime":"2025-12-11T02:02:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:29 crc kubenswrapper[4824]: I1211 02:02:29.158392 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:29 crc kubenswrapper[4824]: I1211 02:02:29.158447 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:29 crc kubenswrapper[4824]: I1211 02:02:29.158466 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:29 crc kubenswrapper[4824]: I1211 02:02:29.158488 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:29 crc kubenswrapper[4824]: I1211 02:02:29.158505 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:29Z","lastTransitionTime":"2025-12-11T02:02:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:29 crc kubenswrapper[4824]: I1211 02:02:29.260437 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:29 crc kubenswrapper[4824]: I1211 02:02:29.260509 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:29 crc kubenswrapper[4824]: I1211 02:02:29.260525 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:29 crc kubenswrapper[4824]: I1211 02:02:29.260553 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:29 crc kubenswrapper[4824]: I1211 02:02:29.260570 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:29Z","lastTransitionTime":"2025-12-11T02:02:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:29 crc kubenswrapper[4824]: I1211 02:02:29.363515 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:29 crc kubenswrapper[4824]: I1211 02:02:29.363575 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:29 crc kubenswrapper[4824]: I1211 02:02:29.363592 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:29 crc kubenswrapper[4824]: I1211 02:02:29.363615 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:29 crc kubenswrapper[4824]: I1211 02:02:29.363634 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:29Z","lastTransitionTime":"2025-12-11T02:02:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:29 crc kubenswrapper[4824]: I1211 02:02:29.466448 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:29 crc kubenswrapper[4824]: I1211 02:02:29.466535 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:29 crc kubenswrapper[4824]: I1211 02:02:29.466560 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:29 crc kubenswrapper[4824]: I1211 02:02:29.466598 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:29 crc kubenswrapper[4824]: I1211 02:02:29.466625 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:29Z","lastTransitionTime":"2025-12-11T02:02:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:29 crc kubenswrapper[4824]: I1211 02:02:29.569510 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:29 crc kubenswrapper[4824]: I1211 02:02:29.569576 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:29 crc kubenswrapper[4824]: I1211 02:02:29.569598 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:29 crc kubenswrapper[4824]: I1211 02:02:29.569628 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:29 crc kubenswrapper[4824]: I1211 02:02:29.569650 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:29Z","lastTransitionTime":"2025-12-11T02:02:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:29 crc kubenswrapper[4824]: I1211 02:02:29.632386 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-grnpt" Dec 11 02:02:29 crc kubenswrapper[4824]: I1211 02:02:29.632485 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 02:02:29 crc kubenswrapper[4824]: I1211 02:02:29.632537 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 02:02:29 crc kubenswrapper[4824]: E1211 02:02:29.632767 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-grnpt" podUID="09a52948-0404-45ff-85b1-51479268bf71" Dec 11 02:02:29 crc kubenswrapper[4824]: I1211 02:02:29.632797 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 02:02:29 crc kubenswrapper[4824]: E1211 02:02:29.633179 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 02:02:29 crc kubenswrapper[4824]: E1211 02:02:29.633427 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 02:02:29 crc kubenswrapper[4824]: E1211 02:02:29.633553 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 02:02:29 crc kubenswrapper[4824]: I1211 02:02:29.673627 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:29 crc kubenswrapper[4824]: I1211 02:02:29.673679 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:29 crc kubenswrapper[4824]: I1211 02:02:29.673696 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:29 crc kubenswrapper[4824]: I1211 02:02:29.673720 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:29 crc kubenswrapper[4824]: I1211 02:02:29.673737 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:29Z","lastTransitionTime":"2025-12-11T02:02:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:29 crc kubenswrapper[4824]: I1211 02:02:29.776097 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:29 crc kubenswrapper[4824]: I1211 02:02:29.776215 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:29 crc kubenswrapper[4824]: I1211 02:02:29.776237 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:29 crc kubenswrapper[4824]: I1211 02:02:29.776260 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:29 crc kubenswrapper[4824]: I1211 02:02:29.776280 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:29Z","lastTransitionTime":"2025-12-11T02:02:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:29 crc kubenswrapper[4824]: I1211 02:02:29.879945 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:29 crc kubenswrapper[4824]: I1211 02:02:29.880051 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:29 crc kubenswrapper[4824]: I1211 02:02:29.880083 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:29 crc kubenswrapper[4824]: I1211 02:02:29.880189 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:29 crc kubenswrapper[4824]: I1211 02:02:29.880228 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:29Z","lastTransitionTime":"2025-12-11T02:02:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:29 crc kubenswrapper[4824]: I1211 02:02:29.983859 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:29 crc kubenswrapper[4824]: I1211 02:02:29.983912 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:29 crc kubenswrapper[4824]: I1211 02:02:29.983929 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:29 crc kubenswrapper[4824]: I1211 02:02:29.983956 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:29 crc kubenswrapper[4824]: I1211 02:02:29.983975 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:29Z","lastTransitionTime":"2025-12-11T02:02:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:30 crc kubenswrapper[4824]: I1211 02:02:30.088416 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:30 crc kubenswrapper[4824]: I1211 02:02:30.088482 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:30 crc kubenswrapper[4824]: I1211 02:02:30.088503 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:30 crc kubenswrapper[4824]: I1211 02:02:30.088528 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:30 crc kubenswrapper[4824]: I1211 02:02:30.088551 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:30Z","lastTransitionTime":"2025-12-11T02:02:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:30 crc kubenswrapper[4824]: I1211 02:02:30.191192 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:30 crc kubenswrapper[4824]: I1211 02:02:30.191256 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:30 crc kubenswrapper[4824]: I1211 02:02:30.191273 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:30 crc kubenswrapper[4824]: I1211 02:02:30.191297 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:30 crc kubenswrapper[4824]: I1211 02:02:30.191321 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:30Z","lastTransitionTime":"2025-12-11T02:02:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:30 crc kubenswrapper[4824]: I1211 02:02:30.294558 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:30 crc kubenswrapper[4824]: I1211 02:02:30.294636 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:30 crc kubenswrapper[4824]: I1211 02:02:30.294660 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:30 crc kubenswrapper[4824]: I1211 02:02:30.294688 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:30 crc kubenswrapper[4824]: I1211 02:02:30.294710 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:30Z","lastTransitionTime":"2025-12-11T02:02:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:30 crc kubenswrapper[4824]: I1211 02:02:30.399193 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:30 crc kubenswrapper[4824]: I1211 02:02:30.399285 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:30 crc kubenswrapper[4824]: I1211 02:02:30.399303 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:30 crc kubenswrapper[4824]: I1211 02:02:30.399326 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:30 crc kubenswrapper[4824]: I1211 02:02:30.399343 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:30Z","lastTransitionTime":"2025-12-11T02:02:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:30 crc kubenswrapper[4824]: I1211 02:02:30.502898 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:30 crc kubenswrapper[4824]: I1211 02:02:30.502967 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:30 crc kubenswrapper[4824]: I1211 02:02:30.502986 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:30 crc kubenswrapper[4824]: I1211 02:02:30.503043 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:30 crc kubenswrapper[4824]: I1211 02:02:30.503064 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:30Z","lastTransitionTime":"2025-12-11T02:02:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:30 crc kubenswrapper[4824]: I1211 02:02:30.617522 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:30 crc kubenswrapper[4824]: I1211 02:02:30.617591 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:30 crc kubenswrapper[4824]: I1211 02:02:30.617609 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:30 crc kubenswrapper[4824]: I1211 02:02:30.617635 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:30 crc kubenswrapper[4824]: I1211 02:02:30.617652 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:30Z","lastTransitionTime":"2025-12-11T02:02:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:30 crc kubenswrapper[4824]: I1211 02:02:30.720878 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:30 crc kubenswrapper[4824]: I1211 02:02:30.720964 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:30 crc kubenswrapper[4824]: I1211 02:02:30.720983 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:30 crc kubenswrapper[4824]: I1211 02:02:30.721012 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:30 crc kubenswrapper[4824]: I1211 02:02:30.721030 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:30Z","lastTransitionTime":"2025-12-11T02:02:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:30 crc kubenswrapper[4824]: I1211 02:02:30.823774 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:30 crc kubenswrapper[4824]: I1211 02:02:30.823833 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:30 crc kubenswrapper[4824]: I1211 02:02:30.823858 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:30 crc kubenswrapper[4824]: I1211 02:02:30.823887 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:30 crc kubenswrapper[4824]: I1211 02:02:30.823909 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:30Z","lastTransitionTime":"2025-12-11T02:02:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:30 crc kubenswrapper[4824]: I1211 02:02:30.927738 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:30 crc kubenswrapper[4824]: I1211 02:02:30.927804 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:30 crc kubenswrapper[4824]: I1211 02:02:30.927831 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:30 crc kubenswrapper[4824]: I1211 02:02:30.927860 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:30 crc kubenswrapper[4824]: I1211 02:02:30.927881 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:30Z","lastTransitionTime":"2025-12-11T02:02:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:31 crc kubenswrapper[4824]: I1211 02:02:31.030435 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:31 crc kubenswrapper[4824]: I1211 02:02:31.030492 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:31 crc kubenswrapper[4824]: I1211 02:02:31.030511 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:31 crc kubenswrapper[4824]: I1211 02:02:31.030535 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:31 crc kubenswrapper[4824]: I1211 02:02:31.030551 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:31Z","lastTransitionTime":"2025-12-11T02:02:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:31 crc kubenswrapper[4824]: I1211 02:02:31.133841 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:31 crc kubenswrapper[4824]: I1211 02:02:31.133873 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:31 crc kubenswrapper[4824]: I1211 02:02:31.133882 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:31 crc kubenswrapper[4824]: I1211 02:02:31.133895 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:31 crc kubenswrapper[4824]: I1211 02:02:31.133905 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:31Z","lastTransitionTime":"2025-12-11T02:02:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:31 crc kubenswrapper[4824]: I1211 02:02:31.237079 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:31 crc kubenswrapper[4824]: I1211 02:02:31.237178 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:31 crc kubenswrapper[4824]: I1211 02:02:31.237195 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:31 crc kubenswrapper[4824]: I1211 02:02:31.237220 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:31 crc kubenswrapper[4824]: I1211 02:02:31.237237 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:31Z","lastTransitionTime":"2025-12-11T02:02:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:31 crc kubenswrapper[4824]: I1211 02:02:31.340474 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:31 crc kubenswrapper[4824]: I1211 02:02:31.340531 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:31 crc kubenswrapper[4824]: I1211 02:02:31.340550 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:31 crc kubenswrapper[4824]: I1211 02:02:31.340574 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:31 crc kubenswrapper[4824]: I1211 02:02:31.340592 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:31Z","lastTransitionTime":"2025-12-11T02:02:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:31 crc kubenswrapper[4824]: I1211 02:02:31.443450 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:31 crc kubenswrapper[4824]: I1211 02:02:31.443517 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:31 crc kubenswrapper[4824]: I1211 02:02:31.443540 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:31 crc kubenswrapper[4824]: I1211 02:02:31.443564 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:31 crc kubenswrapper[4824]: I1211 02:02:31.443581 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:31Z","lastTransitionTime":"2025-12-11T02:02:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:31 crc kubenswrapper[4824]: I1211 02:02:31.546395 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:31 crc kubenswrapper[4824]: I1211 02:02:31.546452 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:31 crc kubenswrapper[4824]: I1211 02:02:31.546469 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:31 crc kubenswrapper[4824]: I1211 02:02:31.546495 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:31 crc kubenswrapper[4824]: I1211 02:02:31.546511 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:31Z","lastTransitionTime":"2025-12-11T02:02:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:31 crc kubenswrapper[4824]: I1211 02:02:31.631537 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 02:02:31 crc kubenswrapper[4824]: I1211 02:02:31.631577 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 02:02:31 crc kubenswrapper[4824]: I1211 02:02:31.631578 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 02:02:31 crc kubenswrapper[4824]: I1211 02:02:31.631719 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-grnpt" Dec 11 02:02:31 crc kubenswrapper[4824]: E1211 02:02:31.631718 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 02:02:31 crc kubenswrapper[4824]: E1211 02:02:31.631957 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-grnpt" podUID="09a52948-0404-45ff-85b1-51479268bf71" Dec 11 02:02:31 crc kubenswrapper[4824]: E1211 02:02:31.632096 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 02:02:31 crc kubenswrapper[4824]: E1211 02:02:31.632219 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 02:02:31 crc kubenswrapper[4824]: I1211 02:02:31.649468 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:31 crc kubenswrapper[4824]: I1211 02:02:31.649519 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:31 crc kubenswrapper[4824]: I1211 02:02:31.649535 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:31 crc kubenswrapper[4824]: I1211 02:02:31.649559 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:31 crc kubenswrapper[4824]: I1211 02:02:31.649575 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:31Z","lastTransitionTime":"2025-12-11T02:02:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:31 crc kubenswrapper[4824]: I1211 02:02:31.752548 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:31 crc kubenswrapper[4824]: I1211 02:02:31.752651 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:31 crc kubenswrapper[4824]: I1211 02:02:31.752674 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:31 crc kubenswrapper[4824]: I1211 02:02:31.752701 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:31 crc kubenswrapper[4824]: I1211 02:02:31.752723 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:31Z","lastTransitionTime":"2025-12-11T02:02:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:31 crc kubenswrapper[4824]: I1211 02:02:31.855975 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:31 crc kubenswrapper[4824]: I1211 02:02:31.856153 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:31 crc kubenswrapper[4824]: I1211 02:02:31.856182 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:31 crc kubenswrapper[4824]: I1211 02:02:31.856207 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:31 crc kubenswrapper[4824]: I1211 02:02:31.856225 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:31Z","lastTransitionTime":"2025-12-11T02:02:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:31 crc kubenswrapper[4824]: I1211 02:02:31.959813 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:31 crc kubenswrapper[4824]: I1211 02:02:31.959891 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:31 crc kubenswrapper[4824]: I1211 02:02:31.959913 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:31 crc kubenswrapper[4824]: I1211 02:02:31.959938 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:31 crc kubenswrapper[4824]: I1211 02:02:31.959965 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:31Z","lastTransitionTime":"2025-12-11T02:02:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:32 crc kubenswrapper[4824]: I1211 02:02:32.063374 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:32 crc kubenswrapper[4824]: I1211 02:02:32.063419 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:32 crc kubenswrapper[4824]: I1211 02:02:32.063435 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:32 crc kubenswrapper[4824]: I1211 02:02:32.063457 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:32 crc kubenswrapper[4824]: I1211 02:02:32.063475 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:32Z","lastTransitionTime":"2025-12-11T02:02:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:32 crc kubenswrapper[4824]: I1211 02:02:32.166572 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:32 crc kubenswrapper[4824]: I1211 02:02:32.166646 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:32 crc kubenswrapper[4824]: I1211 02:02:32.166664 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:32 crc kubenswrapper[4824]: I1211 02:02:32.166689 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:32 crc kubenswrapper[4824]: I1211 02:02:32.166705 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:32Z","lastTransitionTime":"2025-12-11T02:02:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:32 crc kubenswrapper[4824]: I1211 02:02:32.269784 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:32 crc kubenswrapper[4824]: I1211 02:02:32.269863 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:32 crc kubenswrapper[4824]: I1211 02:02:32.269887 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:32 crc kubenswrapper[4824]: I1211 02:02:32.269916 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:32 crc kubenswrapper[4824]: I1211 02:02:32.269940 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:32Z","lastTransitionTime":"2025-12-11T02:02:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:32 crc kubenswrapper[4824]: I1211 02:02:32.373056 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:32 crc kubenswrapper[4824]: I1211 02:02:32.373144 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:32 crc kubenswrapper[4824]: I1211 02:02:32.373163 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:32 crc kubenswrapper[4824]: I1211 02:02:32.373189 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:32 crc kubenswrapper[4824]: I1211 02:02:32.373207 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:32Z","lastTransitionTime":"2025-12-11T02:02:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:32 crc kubenswrapper[4824]: I1211 02:02:32.421529 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:32 crc kubenswrapper[4824]: I1211 02:02:32.421586 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:32 crc kubenswrapper[4824]: I1211 02:02:32.421602 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:32 crc kubenswrapper[4824]: I1211 02:02:32.421627 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:32 crc kubenswrapper[4824]: I1211 02:02:32.421644 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:32Z","lastTransitionTime":"2025-12-11T02:02:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:32 crc kubenswrapper[4824]: E1211 02:02:32.443958 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:02:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:02:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:02:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:02:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d57bece6-6fc6-4a84-b8a8-e1678f9e50db\\\",\\\"systemUUID\\\":\\\"f76b32e4-2764-4219-bbf4-f7e56d76df3b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:32Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:32 crc kubenswrapper[4824]: I1211 02:02:32.450176 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:32 crc kubenswrapper[4824]: I1211 02:02:32.450236 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:32 crc kubenswrapper[4824]: I1211 02:02:32.450255 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:32 crc kubenswrapper[4824]: I1211 02:02:32.450280 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:32 crc kubenswrapper[4824]: I1211 02:02:32.450297 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:32Z","lastTransitionTime":"2025-12-11T02:02:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:32 crc kubenswrapper[4824]: E1211 02:02:32.472466 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:02:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:02:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:02:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:02:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d57bece6-6fc6-4a84-b8a8-e1678f9e50db\\\",\\\"systemUUID\\\":\\\"f76b32e4-2764-4219-bbf4-f7e56d76df3b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:32Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:32 crc kubenswrapper[4824]: I1211 02:02:32.478168 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:32 crc kubenswrapper[4824]: I1211 02:02:32.478214 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:32 crc kubenswrapper[4824]: I1211 02:02:32.478232 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:32 crc kubenswrapper[4824]: I1211 02:02:32.478255 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:32 crc kubenswrapper[4824]: I1211 02:02:32.478271 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:32Z","lastTransitionTime":"2025-12-11T02:02:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:32 crc kubenswrapper[4824]: E1211 02:02:32.499947 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:02:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:02:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:02:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:02:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d57bece6-6fc6-4a84-b8a8-e1678f9e50db\\\",\\\"systemUUID\\\":\\\"f76b32e4-2764-4219-bbf4-f7e56d76df3b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:32Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:32 crc kubenswrapper[4824]: I1211 02:02:32.504826 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:32 crc kubenswrapper[4824]: I1211 02:02:32.504877 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:32 crc kubenswrapper[4824]: I1211 02:02:32.504893 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:32 crc kubenswrapper[4824]: I1211 02:02:32.504917 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:32 crc kubenswrapper[4824]: I1211 02:02:32.504935 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:32Z","lastTransitionTime":"2025-12-11T02:02:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:32 crc kubenswrapper[4824]: E1211 02:02:32.526993 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:02:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:02:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:02:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:02:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d57bece6-6fc6-4a84-b8a8-e1678f9e50db\\\",\\\"systemUUID\\\":\\\"f76b32e4-2764-4219-bbf4-f7e56d76df3b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:32Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:32 crc kubenswrapper[4824]: I1211 02:02:32.532778 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:32 crc kubenswrapper[4824]: I1211 02:02:32.532875 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:32 crc kubenswrapper[4824]: I1211 02:02:32.532902 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:32 crc kubenswrapper[4824]: I1211 02:02:32.532974 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:32 crc kubenswrapper[4824]: I1211 02:02:32.532998 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:32Z","lastTransitionTime":"2025-12-11T02:02:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:32 crc kubenswrapper[4824]: E1211 02:02:32.554788 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:02:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:02:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:02:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T02:02:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T02:02:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d57bece6-6fc6-4a84-b8a8-e1678f9e50db\\\",\\\"systemUUID\\\":\\\"f76b32e4-2764-4219-bbf4-f7e56d76df3b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T02:02:32Z is after 2025-08-24T17:21:41Z" Dec 11 02:02:32 crc kubenswrapper[4824]: E1211 02:02:32.555010 4824 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 11 02:02:32 crc kubenswrapper[4824]: I1211 02:02:32.557256 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:32 crc kubenswrapper[4824]: I1211 02:02:32.557387 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:32 crc kubenswrapper[4824]: I1211 02:02:32.557414 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:32 crc kubenswrapper[4824]: I1211 02:02:32.557443 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:32 crc kubenswrapper[4824]: I1211 02:02:32.557467 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:32Z","lastTransitionTime":"2025-12-11T02:02:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:32 crc kubenswrapper[4824]: I1211 02:02:32.660397 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:32 crc kubenswrapper[4824]: I1211 02:02:32.660480 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:32 crc kubenswrapper[4824]: I1211 02:02:32.660499 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:32 crc kubenswrapper[4824]: I1211 02:02:32.660531 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:32 crc kubenswrapper[4824]: I1211 02:02:32.660547 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:32Z","lastTransitionTime":"2025-12-11T02:02:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:32 crc kubenswrapper[4824]: I1211 02:02:32.763361 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:32 crc kubenswrapper[4824]: I1211 02:02:32.763433 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:32 crc kubenswrapper[4824]: I1211 02:02:32.763456 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:32 crc kubenswrapper[4824]: I1211 02:02:32.763488 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:32 crc kubenswrapper[4824]: I1211 02:02:32.763510 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:32Z","lastTransitionTime":"2025-12-11T02:02:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:32 crc kubenswrapper[4824]: I1211 02:02:32.866963 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:32 crc kubenswrapper[4824]: I1211 02:02:32.867030 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:32 crc kubenswrapper[4824]: I1211 02:02:32.867048 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:32 crc kubenswrapper[4824]: I1211 02:02:32.867071 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:32 crc kubenswrapper[4824]: I1211 02:02:32.867088 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:32Z","lastTransitionTime":"2025-12-11T02:02:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:32 crc kubenswrapper[4824]: I1211 02:02:32.969631 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:32 crc kubenswrapper[4824]: I1211 02:02:32.969677 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:32 crc kubenswrapper[4824]: I1211 02:02:32.969687 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:32 crc kubenswrapper[4824]: I1211 02:02:32.969702 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:32 crc kubenswrapper[4824]: I1211 02:02:32.969713 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:32Z","lastTransitionTime":"2025-12-11T02:02:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:33 crc kubenswrapper[4824]: I1211 02:02:33.072788 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:33 crc kubenswrapper[4824]: I1211 02:02:33.072874 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:33 crc kubenswrapper[4824]: I1211 02:02:33.072897 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:33 crc kubenswrapper[4824]: I1211 02:02:33.072928 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:33 crc kubenswrapper[4824]: I1211 02:02:33.072949 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:33Z","lastTransitionTime":"2025-12-11T02:02:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:33 crc kubenswrapper[4824]: I1211 02:02:33.175641 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:33 crc kubenswrapper[4824]: I1211 02:02:33.175714 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:33 crc kubenswrapper[4824]: I1211 02:02:33.175737 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:33 crc kubenswrapper[4824]: I1211 02:02:33.175767 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:33 crc kubenswrapper[4824]: I1211 02:02:33.175789 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:33Z","lastTransitionTime":"2025-12-11T02:02:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:33 crc kubenswrapper[4824]: I1211 02:02:33.278452 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:33 crc kubenswrapper[4824]: I1211 02:02:33.278504 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:33 crc kubenswrapper[4824]: I1211 02:02:33.278515 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:33 crc kubenswrapper[4824]: I1211 02:02:33.278532 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:33 crc kubenswrapper[4824]: I1211 02:02:33.278543 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:33Z","lastTransitionTime":"2025-12-11T02:02:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:33 crc kubenswrapper[4824]: I1211 02:02:33.381703 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:33 crc kubenswrapper[4824]: I1211 02:02:33.381769 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:33 crc kubenswrapper[4824]: I1211 02:02:33.381791 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:33 crc kubenswrapper[4824]: I1211 02:02:33.381822 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:33 crc kubenswrapper[4824]: I1211 02:02:33.381875 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:33Z","lastTransitionTime":"2025-12-11T02:02:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:33 crc kubenswrapper[4824]: I1211 02:02:33.484669 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:33 crc kubenswrapper[4824]: I1211 02:02:33.484714 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:33 crc kubenswrapper[4824]: I1211 02:02:33.484725 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:33 crc kubenswrapper[4824]: I1211 02:02:33.484741 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:33 crc kubenswrapper[4824]: I1211 02:02:33.484753 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:33Z","lastTransitionTime":"2025-12-11T02:02:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:33 crc kubenswrapper[4824]: I1211 02:02:33.587744 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:33 crc kubenswrapper[4824]: I1211 02:02:33.587831 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:33 crc kubenswrapper[4824]: I1211 02:02:33.587864 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:33 crc kubenswrapper[4824]: I1211 02:02:33.587894 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:33 crc kubenswrapper[4824]: I1211 02:02:33.587917 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:33Z","lastTransitionTime":"2025-12-11T02:02:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:33 crc kubenswrapper[4824]: I1211 02:02:33.631954 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-grnpt" Dec 11 02:02:33 crc kubenswrapper[4824]: I1211 02:02:33.632009 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 02:02:33 crc kubenswrapper[4824]: I1211 02:02:33.632084 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 02:02:33 crc kubenswrapper[4824]: E1211 02:02:33.632236 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 02:02:33 crc kubenswrapper[4824]: E1211 02:02:33.632150 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-grnpt" podUID="09a52948-0404-45ff-85b1-51479268bf71" Dec 11 02:02:33 crc kubenswrapper[4824]: I1211 02:02:33.631954 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 02:02:33 crc kubenswrapper[4824]: E1211 02:02:33.632385 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 02:02:33 crc kubenswrapper[4824]: E1211 02:02:33.632470 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 02:02:33 crc kubenswrapper[4824]: I1211 02:02:33.690546 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:33 crc kubenswrapper[4824]: I1211 02:02:33.690646 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:33 crc kubenswrapper[4824]: I1211 02:02:33.690666 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:33 crc kubenswrapper[4824]: I1211 02:02:33.690704 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:33 crc kubenswrapper[4824]: I1211 02:02:33.690726 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:33Z","lastTransitionTime":"2025-12-11T02:02:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:33 crc kubenswrapper[4824]: I1211 02:02:33.793333 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:33 crc kubenswrapper[4824]: I1211 02:02:33.793381 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:33 crc kubenswrapper[4824]: I1211 02:02:33.793392 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:33 crc kubenswrapper[4824]: I1211 02:02:33.793409 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:33 crc kubenswrapper[4824]: I1211 02:02:33.793421 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:33Z","lastTransitionTime":"2025-12-11T02:02:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:33 crc kubenswrapper[4824]: I1211 02:02:33.906714 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:33 crc kubenswrapper[4824]: I1211 02:02:33.906745 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:33 crc kubenswrapper[4824]: I1211 02:02:33.906754 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:33 crc kubenswrapper[4824]: I1211 02:02:33.906766 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:33 crc kubenswrapper[4824]: I1211 02:02:33.906774 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:33Z","lastTransitionTime":"2025-12-11T02:02:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:34 crc kubenswrapper[4824]: I1211 02:02:34.009979 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:34 crc kubenswrapper[4824]: I1211 02:02:34.010058 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:34 crc kubenswrapper[4824]: I1211 02:02:34.010081 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:34 crc kubenswrapper[4824]: I1211 02:02:34.010140 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:34 crc kubenswrapper[4824]: I1211 02:02:34.010167 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:34Z","lastTransitionTime":"2025-12-11T02:02:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:34 crc kubenswrapper[4824]: I1211 02:02:34.112620 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:34 crc kubenswrapper[4824]: I1211 02:02:34.112682 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:34 crc kubenswrapper[4824]: I1211 02:02:34.112700 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:34 crc kubenswrapper[4824]: I1211 02:02:34.112761 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:34 crc kubenswrapper[4824]: I1211 02:02:34.112784 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:34Z","lastTransitionTime":"2025-12-11T02:02:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:34 crc kubenswrapper[4824]: I1211 02:02:34.215580 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:34 crc kubenswrapper[4824]: I1211 02:02:34.215647 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:34 crc kubenswrapper[4824]: I1211 02:02:34.215665 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:34 crc kubenswrapper[4824]: I1211 02:02:34.215690 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:34 crc kubenswrapper[4824]: I1211 02:02:34.215710 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:34Z","lastTransitionTime":"2025-12-11T02:02:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:34 crc kubenswrapper[4824]: I1211 02:02:34.318414 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:34 crc kubenswrapper[4824]: I1211 02:02:34.318479 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:34 crc kubenswrapper[4824]: I1211 02:02:34.318504 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:34 crc kubenswrapper[4824]: I1211 02:02:34.318533 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:34 crc kubenswrapper[4824]: I1211 02:02:34.318554 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:34Z","lastTransitionTime":"2025-12-11T02:02:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:34 crc kubenswrapper[4824]: I1211 02:02:34.421491 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:34 crc kubenswrapper[4824]: I1211 02:02:34.421545 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:34 crc kubenswrapper[4824]: I1211 02:02:34.421563 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:34 crc kubenswrapper[4824]: I1211 02:02:34.421583 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:34 crc kubenswrapper[4824]: I1211 02:02:34.421599 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:34Z","lastTransitionTime":"2025-12-11T02:02:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:34 crc kubenswrapper[4824]: I1211 02:02:34.523996 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:34 crc kubenswrapper[4824]: I1211 02:02:34.524072 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:34 crc kubenswrapper[4824]: I1211 02:02:34.524092 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:34 crc kubenswrapper[4824]: I1211 02:02:34.524201 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:34 crc kubenswrapper[4824]: I1211 02:02:34.524226 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:34Z","lastTransitionTime":"2025-12-11T02:02:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:34 crc kubenswrapper[4824]: I1211 02:02:34.628256 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:34 crc kubenswrapper[4824]: I1211 02:02:34.628298 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:34 crc kubenswrapper[4824]: I1211 02:02:34.628313 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:34 crc kubenswrapper[4824]: I1211 02:02:34.628334 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:34 crc kubenswrapper[4824]: I1211 02:02:34.628350 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:34Z","lastTransitionTime":"2025-12-11T02:02:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:34 crc kubenswrapper[4824]: I1211 02:02:34.731503 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:34 crc kubenswrapper[4824]: I1211 02:02:34.731549 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:34 crc kubenswrapper[4824]: I1211 02:02:34.731564 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:34 crc kubenswrapper[4824]: I1211 02:02:34.731585 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:34 crc kubenswrapper[4824]: I1211 02:02:34.731599 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:34Z","lastTransitionTime":"2025-12-11T02:02:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:34 crc kubenswrapper[4824]: I1211 02:02:34.834021 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:34 crc kubenswrapper[4824]: I1211 02:02:34.834092 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:34 crc kubenswrapper[4824]: I1211 02:02:34.834144 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:34 crc kubenswrapper[4824]: I1211 02:02:34.834177 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:34 crc kubenswrapper[4824]: I1211 02:02:34.834199 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:34Z","lastTransitionTime":"2025-12-11T02:02:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:34 crc kubenswrapper[4824]: I1211 02:02:34.937598 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:34 crc kubenswrapper[4824]: I1211 02:02:34.937659 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:34 crc kubenswrapper[4824]: I1211 02:02:34.937678 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:34 crc kubenswrapper[4824]: I1211 02:02:34.937702 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:34 crc kubenswrapper[4824]: I1211 02:02:34.937718 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:34Z","lastTransitionTime":"2025-12-11T02:02:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:35 crc kubenswrapper[4824]: I1211 02:02:35.040985 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:35 crc kubenswrapper[4824]: I1211 02:02:35.041047 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:35 crc kubenswrapper[4824]: I1211 02:02:35.041071 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:35 crc kubenswrapper[4824]: I1211 02:02:35.041099 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:35 crc kubenswrapper[4824]: I1211 02:02:35.041161 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:35Z","lastTransitionTime":"2025-12-11T02:02:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:35 crc kubenswrapper[4824]: I1211 02:02:35.144589 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:35 crc kubenswrapper[4824]: I1211 02:02:35.144680 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:35 crc kubenswrapper[4824]: I1211 02:02:35.144697 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:35 crc kubenswrapper[4824]: I1211 02:02:35.144720 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:35 crc kubenswrapper[4824]: I1211 02:02:35.144739 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:35Z","lastTransitionTime":"2025-12-11T02:02:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:35 crc kubenswrapper[4824]: I1211 02:02:35.248527 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:35 crc kubenswrapper[4824]: I1211 02:02:35.248598 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:35 crc kubenswrapper[4824]: I1211 02:02:35.248615 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:35 crc kubenswrapper[4824]: I1211 02:02:35.248639 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:35 crc kubenswrapper[4824]: I1211 02:02:35.248657 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:35Z","lastTransitionTime":"2025-12-11T02:02:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:35 crc kubenswrapper[4824]: I1211 02:02:35.351849 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:35 crc kubenswrapper[4824]: I1211 02:02:35.351891 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:35 crc kubenswrapper[4824]: I1211 02:02:35.351902 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:35 crc kubenswrapper[4824]: I1211 02:02:35.351920 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:35 crc kubenswrapper[4824]: I1211 02:02:35.351934 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:35Z","lastTransitionTime":"2025-12-11T02:02:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:35 crc kubenswrapper[4824]: I1211 02:02:35.454378 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:35 crc kubenswrapper[4824]: I1211 02:02:35.454446 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:35 crc kubenswrapper[4824]: I1211 02:02:35.454462 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:35 crc kubenswrapper[4824]: I1211 02:02:35.454485 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:35 crc kubenswrapper[4824]: I1211 02:02:35.454503 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:35Z","lastTransitionTime":"2025-12-11T02:02:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:35 crc kubenswrapper[4824]: I1211 02:02:35.557091 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:35 crc kubenswrapper[4824]: I1211 02:02:35.557161 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:35 crc kubenswrapper[4824]: I1211 02:02:35.557174 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:35 crc kubenswrapper[4824]: I1211 02:02:35.557192 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:35 crc kubenswrapper[4824]: I1211 02:02:35.557205 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:35Z","lastTransitionTime":"2025-12-11T02:02:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:35 crc kubenswrapper[4824]: I1211 02:02:35.632234 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 02:02:35 crc kubenswrapper[4824]: I1211 02:02:35.632250 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 02:02:35 crc kubenswrapper[4824]: I1211 02:02:35.632321 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-grnpt" Dec 11 02:02:35 crc kubenswrapper[4824]: I1211 02:02:35.632394 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 02:02:35 crc kubenswrapper[4824]: E1211 02:02:35.632555 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 02:02:35 crc kubenswrapper[4824]: E1211 02:02:35.632773 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-grnpt" podUID="09a52948-0404-45ff-85b1-51479268bf71" Dec 11 02:02:35 crc kubenswrapper[4824]: E1211 02:02:35.632851 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 02:02:35 crc kubenswrapper[4824]: E1211 02:02:35.632985 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 02:02:35 crc kubenswrapper[4824]: I1211 02:02:35.660077 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:35 crc kubenswrapper[4824]: I1211 02:02:35.660167 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:35 crc kubenswrapper[4824]: I1211 02:02:35.660185 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:35 crc kubenswrapper[4824]: I1211 02:02:35.660210 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:35 crc kubenswrapper[4824]: I1211 02:02:35.660228 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:35Z","lastTransitionTime":"2025-12-11T02:02:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:35 crc kubenswrapper[4824]: I1211 02:02:35.735878 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/09a52948-0404-45ff-85b1-51479268bf71-metrics-certs\") pod \"network-metrics-daemon-grnpt\" (UID: \"09a52948-0404-45ff-85b1-51479268bf71\") " pod="openshift-multus/network-metrics-daemon-grnpt" Dec 11 02:02:35 crc kubenswrapper[4824]: E1211 02:02:35.736036 4824 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 11 02:02:35 crc kubenswrapper[4824]: E1211 02:02:35.736168 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/09a52948-0404-45ff-85b1-51479268bf71-metrics-certs podName:09a52948-0404-45ff-85b1-51479268bf71 nodeName:}" failed. No retries permitted until 2025-12-11 02:03:39.736143721 +0000 UTC m=+161.425181140 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/09a52948-0404-45ff-85b1-51479268bf71-metrics-certs") pod "network-metrics-daemon-grnpt" (UID: "09a52948-0404-45ff-85b1-51479268bf71") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 11 02:02:35 crc kubenswrapper[4824]: I1211 02:02:35.763245 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:35 crc kubenswrapper[4824]: I1211 02:02:35.763318 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:35 crc kubenswrapper[4824]: I1211 02:02:35.763336 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:35 crc kubenswrapper[4824]: I1211 02:02:35.763368 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:35 crc kubenswrapper[4824]: I1211 02:02:35.763390 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:35Z","lastTransitionTime":"2025-12-11T02:02:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:35 crc kubenswrapper[4824]: I1211 02:02:35.866467 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:35 crc kubenswrapper[4824]: I1211 02:02:35.866558 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:35 crc kubenswrapper[4824]: I1211 02:02:35.866581 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:35 crc kubenswrapper[4824]: I1211 02:02:35.866606 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:35 crc kubenswrapper[4824]: I1211 02:02:35.866624 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:35Z","lastTransitionTime":"2025-12-11T02:02:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:35 crc kubenswrapper[4824]: I1211 02:02:35.969583 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:35 crc kubenswrapper[4824]: I1211 02:02:35.969649 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:35 crc kubenswrapper[4824]: I1211 02:02:35.969666 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:35 crc kubenswrapper[4824]: I1211 02:02:35.969697 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:35 crc kubenswrapper[4824]: I1211 02:02:35.969716 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:35Z","lastTransitionTime":"2025-12-11T02:02:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:36 crc kubenswrapper[4824]: I1211 02:02:36.073634 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:36 crc kubenswrapper[4824]: I1211 02:02:36.073737 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:36 crc kubenswrapper[4824]: I1211 02:02:36.073754 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:36 crc kubenswrapper[4824]: I1211 02:02:36.073780 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:36 crc kubenswrapper[4824]: I1211 02:02:36.073798 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:36Z","lastTransitionTime":"2025-12-11T02:02:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:36 crc kubenswrapper[4824]: I1211 02:02:36.177683 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:36 crc kubenswrapper[4824]: I1211 02:02:36.177750 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:36 crc kubenswrapper[4824]: I1211 02:02:36.177767 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:36 crc kubenswrapper[4824]: I1211 02:02:36.177798 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:36 crc kubenswrapper[4824]: I1211 02:02:36.177816 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:36Z","lastTransitionTime":"2025-12-11T02:02:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:36 crc kubenswrapper[4824]: I1211 02:02:36.281760 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:36 crc kubenswrapper[4824]: I1211 02:02:36.281826 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:36 crc kubenswrapper[4824]: I1211 02:02:36.281843 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:36 crc kubenswrapper[4824]: I1211 02:02:36.281901 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:36 crc kubenswrapper[4824]: I1211 02:02:36.281919 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:36Z","lastTransitionTime":"2025-12-11T02:02:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:36 crc kubenswrapper[4824]: I1211 02:02:36.384362 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:36 crc kubenswrapper[4824]: I1211 02:02:36.384422 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:36 crc kubenswrapper[4824]: I1211 02:02:36.384440 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:36 crc kubenswrapper[4824]: I1211 02:02:36.384465 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:36 crc kubenswrapper[4824]: I1211 02:02:36.384482 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:36Z","lastTransitionTime":"2025-12-11T02:02:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:36 crc kubenswrapper[4824]: I1211 02:02:36.487556 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:36 crc kubenswrapper[4824]: I1211 02:02:36.487623 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:36 crc kubenswrapper[4824]: I1211 02:02:36.487640 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:36 crc kubenswrapper[4824]: I1211 02:02:36.487664 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:36 crc kubenswrapper[4824]: I1211 02:02:36.487680 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:36Z","lastTransitionTime":"2025-12-11T02:02:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:36 crc kubenswrapper[4824]: I1211 02:02:36.590062 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:36 crc kubenswrapper[4824]: I1211 02:02:36.590135 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:36 crc kubenswrapper[4824]: I1211 02:02:36.590150 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:36 crc kubenswrapper[4824]: I1211 02:02:36.590170 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:36 crc kubenswrapper[4824]: I1211 02:02:36.590184 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:36Z","lastTransitionTime":"2025-12-11T02:02:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:36 crc kubenswrapper[4824]: I1211 02:02:36.633546 4824 scope.go:117] "RemoveContainer" containerID="154d82fd6acee749bfd37220547e52647dc7f9d86a73c28ce7d8577c9cadef6e" Dec 11 02:02:36 crc kubenswrapper[4824]: E1211 02:02:36.633807 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-f6pgc_openshift-ovn-kubernetes(5b193bd2-b536-4056-92f1-94c9836ab2eb)\"" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" podUID="5b193bd2-b536-4056-92f1-94c9836ab2eb" Dec 11 02:02:36 crc kubenswrapper[4824]: I1211 02:02:36.693179 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:36 crc kubenswrapper[4824]: I1211 02:02:36.693239 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:36 crc kubenswrapper[4824]: I1211 02:02:36.693256 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:36 crc kubenswrapper[4824]: I1211 02:02:36.693278 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:36 crc kubenswrapper[4824]: I1211 02:02:36.693295 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:36Z","lastTransitionTime":"2025-12-11T02:02:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:36 crc kubenswrapper[4824]: I1211 02:02:36.796861 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:36 crc kubenswrapper[4824]: I1211 02:02:36.796914 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:36 crc kubenswrapper[4824]: I1211 02:02:36.796933 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:36 crc kubenswrapper[4824]: I1211 02:02:36.796957 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:36 crc kubenswrapper[4824]: I1211 02:02:36.796974 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:36Z","lastTransitionTime":"2025-12-11T02:02:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:36 crc kubenswrapper[4824]: I1211 02:02:36.899666 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:36 crc kubenswrapper[4824]: I1211 02:02:36.899710 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:36 crc kubenswrapper[4824]: I1211 02:02:36.899726 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:36 crc kubenswrapper[4824]: I1211 02:02:36.899747 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:36 crc kubenswrapper[4824]: I1211 02:02:36.899765 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:36Z","lastTransitionTime":"2025-12-11T02:02:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:37 crc kubenswrapper[4824]: I1211 02:02:37.002585 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:37 crc kubenswrapper[4824]: I1211 02:02:37.002646 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:37 crc kubenswrapper[4824]: I1211 02:02:37.002663 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:37 crc kubenswrapper[4824]: I1211 02:02:37.002686 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:37 crc kubenswrapper[4824]: I1211 02:02:37.002702 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:37Z","lastTransitionTime":"2025-12-11T02:02:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:37 crc kubenswrapper[4824]: I1211 02:02:37.105601 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:37 crc kubenswrapper[4824]: I1211 02:02:37.105660 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:37 crc kubenswrapper[4824]: I1211 02:02:37.105675 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:37 crc kubenswrapper[4824]: I1211 02:02:37.105700 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:37 crc kubenswrapper[4824]: I1211 02:02:37.105717 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:37Z","lastTransitionTime":"2025-12-11T02:02:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:37 crc kubenswrapper[4824]: I1211 02:02:37.208488 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:37 crc kubenswrapper[4824]: I1211 02:02:37.208574 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:37 crc kubenswrapper[4824]: I1211 02:02:37.208593 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:37 crc kubenswrapper[4824]: I1211 02:02:37.208621 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:37 crc kubenswrapper[4824]: I1211 02:02:37.208640 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:37Z","lastTransitionTime":"2025-12-11T02:02:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:37 crc kubenswrapper[4824]: I1211 02:02:37.311286 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:37 crc kubenswrapper[4824]: I1211 02:02:37.311360 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:37 crc kubenswrapper[4824]: I1211 02:02:37.311384 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:37 crc kubenswrapper[4824]: I1211 02:02:37.311412 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:37 crc kubenswrapper[4824]: I1211 02:02:37.311430 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:37Z","lastTransitionTime":"2025-12-11T02:02:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:37 crc kubenswrapper[4824]: I1211 02:02:37.414944 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:37 crc kubenswrapper[4824]: I1211 02:02:37.415006 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:37 crc kubenswrapper[4824]: I1211 02:02:37.415027 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:37 crc kubenswrapper[4824]: I1211 02:02:37.415056 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:37 crc kubenswrapper[4824]: I1211 02:02:37.415079 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:37Z","lastTransitionTime":"2025-12-11T02:02:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:37 crc kubenswrapper[4824]: I1211 02:02:37.517953 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:37 crc kubenswrapper[4824]: I1211 02:02:37.518078 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:37 crc kubenswrapper[4824]: I1211 02:02:37.518144 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:37 crc kubenswrapper[4824]: I1211 02:02:37.518177 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:37 crc kubenswrapper[4824]: I1211 02:02:37.518198 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:37Z","lastTransitionTime":"2025-12-11T02:02:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:37 crc kubenswrapper[4824]: I1211 02:02:37.621524 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:37 crc kubenswrapper[4824]: I1211 02:02:37.621578 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:37 crc kubenswrapper[4824]: I1211 02:02:37.621594 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:37 crc kubenswrapper[4824]: I1211 02:02:37.621618 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:37 crc kubenswrapper[4824]: I1211 02:02:37.621637 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:37Z","lastTransitionTime":"2025-12-11T02:02:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:37 crc kubenswrapper[4824]: I1211 02:02:37.632147 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-grnpt" Dec 11 02:02:37 crc kubenswrapper[4824]: I1211 02:02:37.632203 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 02:02:37 crc kubenswrapper[4824]: I1211 02:02:37.632249 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 02:02:37 crc kubenswrapper[4824]: I1211 02:02:37.632273 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 02:02:37 crc kubenswrapper[4824]: E1211 02:02:37.632473 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-grnpt" podUID="09a52948-0404-45ff-85b1-51479268bf71" Dec 11 02:02:37 crc kubenswrapper[4824]: E1211 02:02:37.632596 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 02:02:37 crc kubenswrapper[4824]: E1211 02:02:37.632705 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 02:02:37 crc kubenswrapper[4824]: E1211 02:02:37.632835 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 02:02:37 crc kubenswrapper[4824]: I1211 02:02:37.724962 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:37 crc kubenswrapper[4824]: I1211 02:02:37.725034 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:37 crc kubenswrapper[4824]: I1211 02:02:37.725058 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:37 crc kubenswrapper[4824]: I1211 02:02:37.725084 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:37 crc kubenswrapper[4824]: I1211 02:02:37.725105 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:37Z","lastTransitionTime":"2025-12-11T02:02:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:37 crc kubenswrapper[4824]: I1211 02:02:37.828674 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:37 crc kubenswrapper[4824]: I1211 02:02:37.828735 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:37 crc kubenswrapper[4824]: I1211 02:02:37.828752 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:37 crc kubenswrapper[4824]: I1211 02:02:37.828776 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:37 crc kubenswrapper[4824]: I1211 02:02:37.828793 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:37Z","lastTransitionTime":"2025-12-11T02:02:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:37 crc kubenswrapper[4824]: I1211 02:02:37.932143 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:37 crc kubenswrapper[4824]: I1211 02:02:37.932212 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:37 crc kubenswrapper[4824]: I1211 02:02:37.932230 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:37 crc kubenswrapper[4824]: I1211 02:02:37.932253 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:37 crc kubenswrapper[4824]: I1211 02:02:37.932275 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:37Z","lastTransitionTime":"2025-12-11T02:02:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:38 crc kubenswrapper[4824]: I1211 02:02:38.035013 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:38 crc kubenswrapper[4824]: I1211 02:02:38.035178 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:38 crc kubenswrapper[4824]: I1211 02:02:38.035212 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:38 crc kubenswrapper[4824]: I1211 02:02:38.035299 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:38 crc kubenswrapper[4824]: I1211 02:02:38.035326 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:38Z","lastTransitionTime":"2025-12-11T02:02:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:38 crc kubenswrapper[4824]: I1211 02:02:38.138808 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:38 crc kubenswrapper[4824]: I1211 02:02:38.138892 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:38 crc kubenswrapper[4824]: I1211 02:02:38.138916 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:38 crc kubenswrapper[4824]: I1211 02:02:38.138949 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:38 crc kubenswrapper[4824]: I1211 02:02:38.138971 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:38Z","lastTransitionTime":"2025-12-11T02:02:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:38 crc kubenswrapper[4824]: I1211 02:02:38.241826 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:38 crc kubenswrapper[4824]: I1211 02:02:38.241924 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:38 crc kubenswrapper[4824]: I1211 02:02:38.241949 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:38 crc kubenswrapper[4824]: I1211 02:02:38.241969 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:38 crc kubenswrapper[4824]: I1211 02:02:38.241984 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:38Z","lastTransitionTime":"2025-12-11T02:02:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:38 crc kubenswrapper[4824]: I1211 02:02:38.345269 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:38 crc kubenswrapper[4824]: I1211 02:02:38.345340 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:38 crc kubenswrapper[4824]: I1211 02:02:38.345358 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:38 crc kubenswrapper[4824]: I1211 02:02:38.345386 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:38 crc kubenswrapper[4824]: I1211 02:02:38.345406 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:38Z","lastTransitionTime":"2025-12-11T02:02:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:38 crc kubenswrapper[4824]: I1211 02:02:38.448534 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:38 crc kubenswrapper[4824]: I1211 02:02:38.448624 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:38 crc kubenswrapper[4824]: I1211 02:02:38.448644 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:38 crc kubenswrapper[4824]: I1211 02:02:38.448679 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:38 crc kubenswrapper[4824]: I1211 02:02:38.448701 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:38Z","lastTransitionTime":"2025-12-11T02:02:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:38 crc kubenswrapper[4824]: I1211 02:02:38.551669 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:38 crc kubenswrapper[4824]: I1211 02:02:38.551719 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:38 crc kubenswrapper[4824]: I1211 02:02:38.551735 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:38 crc kubenswrapper[4824]: I1211 02:02:38.551760 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:38 crc kubenswrapper[4824]: I1211 02:02:38.551778 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:38Z","lastTransitionTime":"2025-12-11T02:02:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:38 crc kubenswrapper[4824]: I1211 02:02:38.654935 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:38 crc kubenswrapper[4824]: I1211 02:02:38.654985 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:38 crc kubenswrapper[4824]: I1211 02:02:38.655001 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:38 crc kubenswrapper[4824]: I1211 02:02:38.655025 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:38 crc kubenswrapper[4824]: I1211 02:02:38.655043 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:38Z","lastTransitionTime":"2025-12-11T02:02:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:38 crc kubenswrapper[4824]: I1211 02:02:38.691966 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=21.691938152 podStartE2EDuration="21.691938152s" podCreationTimestamp="2025-12-11 02:02:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:02:38.691242432 +0000 UTC m=+100.380279871" watchObservedRunningTime="2025-12-11 02:02:38.691938152 +0000 UTC m=+100.380975571" Dec 11 02:02:38 crc kubenswrapper[4824]: I1211 02:02:38.751078 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=81.751051663 podStartE2EDuration="1m21.751051663s" podCreationTimestamp="2025-12-11 02:01:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:02:38.725358239 +0000 UTC m=+100.414395668" watchObservedRunningTime="2025-12-11 02:02:38.751051663 +0000 UTC m=+100.440089082" Dec 11 02:02:38 crc kubenswrapper[4824]: I1211 02:02:38.758732 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:38 crc kubenswrapper[4824]: I1211 02:02:38.758787 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:38 crc kubenswrapper[4824]: I1211 02:02:38.758803 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:38 crc kubenswrapper[4824]: I1211 02:02:38.758823 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:38 crc kubenswrapper[4824]: I1211 02:02:38.758836 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:38Z","lastTransitionTime":"2025-12-11T02:02:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:38 crc kubenswrapper[4824]: I1211 02:02:38.814634 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-kptws" podStartSLOduration=81.814601346 podStartE2EDuration="1m21.814601346s" podCreationTimestamp="2025-12-11 02:01:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:02:38.796985878 +0000 UTC m=+100.486023267" watchObservedRunningTime="2025-12-11 02:02:38.814601346 +0000 UTC m=+100.503638775" Dec 11 02:02:38 crc kubenswrapper[4824]: I1211 02:02:38.842943 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-kh8g9" podStartSLOduration=81.842921883 podStartE2EDuration="1m21.842921883s" podCreationTimestamp="2025-12-11 02:01:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:02:38.842430179 +0000 UTC m=+100.531467598" watchObservedRunningTime="2025-12-11 02:02:38.842921883 +0000 UTC m=+100.531959262" Dec 11 02:02:38 crc kubenswrapper[4824]: I1211 02:02:38.861054 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:38 crc kubenswrapper[4824]: I1211 02:02:38.861116 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:38 crc kubenswrapper[4824]: I1211 02:02:38.861160 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:38 crc kubenswrapper[4824]: I1211 02:02:38.861183 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:38 crc kubenswrapper[4824]: I1211 02:02:38.861202 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:38Z","lastTransitionTime":"2025-12-11T02:02:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:38 crc kubenswrapper[4824]: I1211 02:02:38.888963 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-6bkc2" podStartSLOduration=81.8889372 podStartE2EDuration="1m21.8889372s" podCreationTimestamp="2025-12-11 02:01:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:02:38.859939486 +0000 UTC m=+100.548976905" watchObservedRunningTime="2025-12-11 02:02:38.8889372 +0000 UTC m=+100.577974639" Dec 11 02:02:38 crc kubenswrapper[4824]: I1211 02:02:38.920912 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-br4j9" podStartSLOduration=81.920890167 podStartE2EDuration="1m21.920890167s" podCreationTimestamp="2025-12-11 02:01:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:02:38.906723644 +0000 UTC m=+100.595761063" watchObservedRunningTime="2025-12-11 02:02:38.920890167 +0000 UTC m=+100.609927556" Dec 11 02:02:38 crc kubenswrapper[4824]: I1211 02:02:38.921045 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8v4jn" podStartSLOduration=81.921038361 podStartE2EDuration="1m21.921038361s" podCreationTimestamp="2025-12-11 02:01:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:02:38.919943091 +0000 UTC m=+100.608980510" watchObservedRunningTime="2025-12-11 02:02:38.921038361 +0000 UTC m=+100.610075760" Dec 11 02:02:38 crc kubenswrapper[4824]: I1211 02:02:38.963552 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:38 crc kubenswrapper[4824]: I1211 02:02:38.963620 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:38 crc kubenswrapper[4824]: I1211 02:02:38.963642 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:38 crc kubenswrapper[4824]: I1211 02:02:38.963672 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:38 crc kubenswrapper[4824]: I1211 02:02:38.963694 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:38Z","lastTransitionTime":"2025-12-11T02:02:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:38 crc kubenswrapper[4824]: I1211 02:02:38.990518 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podStartSLOduration=81.990495169 podStartE2EDuration="1m21.990495169s" podCreationTimestamp="2025-12-11 02:01:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:02:38.990400127 +0000 UTC m=+100.679437526" watchObservedRunningTime="2025-12-11 02:02:38.990495169 +0000 UTC m=+100.679532568" Dec 11 02:02:39 crc kubenswrapper[4824]: I1211 02:02:39.005008 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=19.004994112 podStartE2EDuration="19.004994112s" podCreationTimestamp="2025-12-11 02:02:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:02:39.004424036 +0000 UTC m=+100.693461435" watchObservedRunningTime="2025-12-11 02:02:39.004994112 +0000 UTC m=+100.694031491" Dec 11 02:02:39 crc kubenswrapper[4824]: I1211 02:02:39.025362 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=79.025338497 podStartE2EDuration="1m19.025338497s" podCreationTimestamp="2025-12-11 02:01:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:02:39.024646447 +0000 UTC m=+100.713683856" watchObservedRunningTime="2025-12-11 02:02:39.025338497 +0000 UTC m=+100.714375906" Dec 11 02:02:39 crc kubenswrapper[4824]: I1211 02:02:39.039419 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=48.039394707 podStartE2EDuration="48.039394707s" podCreationTimestamp="2025-12-11 02:01:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:02:39.039014426 +0000 UTC m=+100.728051805" watchObservedRunningTime="2025-12-11 02:02:39.039394707 +0000 UTC m=+100.728432096" Dec 11 02:02:39 crc kubenswrapper[4824]: I1211 02:02:39.066020 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:39 crc kubenswrapper[4824]: I1211 02:02:39.066056 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:39 crc kubenswrapper[4824]: I1211 02:02:39.066068 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:39 crc kubenswrapper[4824]: I1211 02:02:39.066083 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:39 crc kubenswrapper[4824]: I1211 02:02:39.066094 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:39Z","lastTransitionTime":"2025-12-11T02:02:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:39 crc kubenswrapper[4824]: I1211 02:02:39.169270 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:39 crc kubenswrapper[4824]: I1211 02:02:39.169319 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:39 crc kubenswrapper[4824]: I1211 02:02:39.169336 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:39 crc kubenswrapper[4824]: I1211 02:02:39.169357 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:39 crc kubenswrapper[4824]: I1211 02:02:39.169373 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:39Z","lastTransitionTime":"2025-12-11T02:02:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:39 crc kubenswrapper[4824]: I1211 02:02:39.272361 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:39 crc kubenswrapper[4824]: I1211 02:02:39.272415 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:39 crc kubenswrapper[4824]: I1211 02:02:39.272431 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:39 crc kubenswrapper[4824]: I1211 02:02:39.272453 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:39 crc kubenswrapper[4824]: I1211 02:02:39.272470 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:39Z","lastTransitionTime":"2025-12-11T02:02:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:39 crc kubenswrapper[4824]: I1211 02:02:39.376297 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:39 crc kubenswrapper[4824]: I1211 02:02:39.376357 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:39 crc kubenswrapper[4824]: I1211 02:02:39.376368 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:39 crc kubenswrapper[4824]: I1211 02:02:39.376389 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:39 crc kubenswrapper[4824]: I1211 02:02:39.376403 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:39Z","lastTransitionTime":"2025-12-11T02:02:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:39 crc kubenswrapper[4824]: I1211 02:02:39.478999 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:39 crc kubenswrapper[4824]: I1211 02:02:39.479079 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:39 crc kubenswrapper[4824]: I1211 02:02:39.479100 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:39 crc kubenswrapper[4824]: I1211 02:02:39.479167 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:39 crc kubenswrapper[4824]: I1211 02:02:39.479192 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:39Z","lastTransitionTime":"2025-12-11T02:02:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:39 crc kubenswrapper[4824]: I1211 02:02:39.581381 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:39 crc kubenswrapper[4824]: I1211 02:02:39.581522 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:39 crc kubenswrapper[4824]: I1211 02:02:39.581545 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:39 crc kubenswrapper[4824]: I1211 02:02:39.581578 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:39 crc kubenswrapper[4824]: I1211 02:02:39.581599 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:39Z","lastTransitionTime":"2025-12-11T02:02:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:39 crc kubenswrapper[4824]: I1211 02:02:39.632341 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-grnpt" Dec 11 02:02:39 crc kubenswrapper[4824]: I1211 02:02:39.632341 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 02:02:39 crc kubenswrapper[4824]: E1211 02:02:39.632510 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-grnpt" podUID="09a52948-0404-45ff-85b1-51479268bf71" Dec 11 02:02:39 crc kubenswrapper[4824]: I1211 02:02:39.632599 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 02:02:39 crc kubenswrapper[4824]: I1211 02:02:39.632610 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 02:02:39 crc kubenswrapper[4824]: E1211 02:02:39.632810 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 02:02:39 crc kubenswrapper[4824]: E1211 02:02:39.633006 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 02:02:39 crc kubenswrapper[4824]: E1211 02:02:39.633441 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 02:02:39 crc kubenswrapper[4824]: I1211 02:02:39.709343 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:39 crc kubenswrapper[4824]: I1211 02:02:39.709405 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:39 crc kubenswrapper[4824]: I1211 02:02:39.709428 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:39 crc kubenswrapper[4824]: I1211 02:02:39.709458 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:39 crc kubenswrapper[4824]: I1211 02:02:39.709481 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:39Z","lastTransitionTime":"2025-12-11T02:02:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:39 crc kubenswrapper[4824]: I1211 02:02:39.812654 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:39 crc kubenswrapper[4824]: I1211 02:02:39.812724 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:39 crc kubenswrapper[4824]: I1211 02:02:39.812746 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:39 crc kubenswrapper[4824]: I1211 02:02:39.812775 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:39 crc kubenswrapper[4824]: I1211 02:02:39.812796 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:39Z","lastTransitionTime":"2025-12-11T02:02:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:39 crc kubenswrapper[4824]: I1211 02:02:39.916069 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:39 crc kubenswrapper[4824]: I1211 02:02:39.916189 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:39 crc kubenswrapper[4824]: I1211 02:02:39.916214 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:39 crc kubenswrapper[4824]: I1211 02:02:39.916242 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:39 crc kubenswrapper[4824]: I1211 02:02:39.916264 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:39Z","lastTransitionTime":"2025-12-11T02:02:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:40 crc kubenswrapper[4824]: I1211 02:02:40.019326 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:40 crc kubenswrapper[4824]: I1211 02:02:40.019399 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:40 crc kubenswrapper[4824]: I1211 02:02:40.019419 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:40 crc kubenswrapper[4824]: I1211 02:02:40.019443 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:40 crc kubenswrapper[4824]: I1211 02:02:40.019464 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:40Z","lastTransitionTime":"2025-12-11T02:02:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:40 crc kubenswrapper[4824]: I1211 02:02:40.122235 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:40 crc kubenswrapper[4824]: I1211 02:02:40.122295 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:40 crc kubenswrapper[4824]: I1211 02:02:40.122304 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:40 crc kubenswrapper[4824]: I1211 02:02:40.122318 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:40 crc kubenswrapper[4824]: I1211 02:02:40.122328 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:40Z","lastTransitionTime":"2025-12-11T02:02:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:40 crc kubenswrapper[4824]: I1211 02:02:40.226848 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:40 crc kubenswrapper[4824]: I1211 02:02:40.227335 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:40 crc kubenswrapper[4824]: I1211 02:02:40.227523 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:40 crc kubenswrapper[4824]: I1211 02:02:40.227706 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:40 crc kubenswrapper[4824]: I1211 02:02:40.227888 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:40Z","lastTransitionTime":"2025-12-11T02:02:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:40 crc kubenswrapper[4824]: I1211 02:02:40.331527 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:40 crc kubenswrapper[4824]: I1211 02:02:40.331861 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:40 crc kubenswrapper[4824]: I1211 02:02:40.332303 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:40 crc kubenswrapper[4824]: I1211 02:02:40.332471 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:40 crc kubenswrapper[4824]: I1211 02:02:40.332655 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:40Z","lastTransitionTime":"2025-12-11T02:02:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:40 crc kubenswrapper[4824]: I1211 02:02:40.436640 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:40 crc kubenswrapper[4824]: I1211 02:02:40.436716 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:40 crc kubenswrapper[4824]: I1211 02:02:40.436740 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:40 crc kubenswrapper[4824]: I1211 02:02:40.436776 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:40 crc kubenswrapper[4824]: I1211 02:02:40.436799 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:40Z","lastTransitionTime":"2025-12-11T02:02:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:40 crc kubenswrapper[4824]: I1211 02:02:40.540784 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:40 crc kubenswrapper[4824]: I1211 02:02:40.540845 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:40 crc kubenswrapper[4824]: I1211 02:02:40.540865 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:40 crc kubenswrapper[4824]: I1211 02:02:40.540889 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:40 crc kubenswrapper[4824]: I1211 02:02:40.540906 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:40Z","lastTransitionTime":"2025-12-11T02:02:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:40 crc kubenswrapper[4824]: I1211 02:02:40.643035 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:40 crc kubenswrapper[4824]: I1211 02:02:40.643099 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:40 crc kubenswrapper[4824]: I1211 02:02:40.643120 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:40 crc kubenswrapper[4824]: I1211 02:02:40.643175 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:40 crc kubenswrapper[4824]: I1211 02:02:40.643193 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:40Z","lastTransitionTime":"2025-12-11T02:02:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:40 crc kubenswrapper[4824]: I1211 02:02:40.746059 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:40 crc kubenswrapper[4824]: I1211 02:02:40.746108 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:40 crc kubenswrapper[4824]: I1211 02:02:40.746161 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:40 crc kubenswrapper[4824]: I1211 02:02:40.746186 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:40 crc kubenswrapper[4824]: I1211 02:02:40.746203 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:40Z","lastTransitionTime":"2025-12-11T02:02:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:40 crc kubenswrapper[4824]: I1211 02:02:40.849279 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:40 crc kubenswrapper[4824]: I1211 02:02:40.849335 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:40 crc kubenswrapper[4824]: I1211 02:02:40.849351 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:40 crc kubenswrapper[4824]: I1211 02:02:40.849372 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:40 crc kubenswrapper[4824]: I1211 02:02:40.849390 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:40Z","lastTransitionTime":"2025-12-11T02:02:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:40 crc kubenswrapper[4824]: I1211 02:02:40.952410 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:40 crc kubenswrapper[4824]: I1211 02:02:40.952475 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:40 crc kubenswrapper[4824]: I1211 02:02:40.952525 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:40 crc kubenswrapper[4824]: I1211 02:02:40.952555 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:40 crc kubenswrapper[4824]: I1211 02:02:40.952578 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:40Z","lastTransitionTime":"2025-12-11T02:02:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:41 crc kubenswrapper[4824]: I1211 02:02:41.055743 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:41 crc kubenswrapper[4824]: I1211 02:02:41.055810 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:41 crc kubenswrapper[4824]: I1211 02:02:41.055832 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:41 crc kubenswrapper[4824]: I1211 02:02:41.055865 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:41 crc kubenswrapper[4824]: I1211 02:02:41.055885 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:41Z","lastTransitionTime":"2025-12-11T02:02:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:41 crc kubenswrapper[4824]: I1211 02:02:41.159344 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:41 crc kubenswrapper[4824]: I1211 02:02:41.159422 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:41 crc kubenswrapper[4824]: I1211 02:02:41.159446 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:41 crc kubenswrapper[4824]: I1211 02:02:41.159477 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:41 crc kubenswrapper[4824]: I1211 02:02:41.159500 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:41Z","lastTransitionTime":"2025-12-11T02:02:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:41 crc kubenswrapper[4824]: I1211 02:02:41.262790 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:41 crc kubenswrapper[4824]: I1211 02:02:41.262867 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:41 crc kubenswrapper[4824]: I1211 02:02:41.262886 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:41 crc kubenswrapper[4824]: I1211 02:02:41.262913 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:41 crc kubenswrapper[4824]: I1211 02:02:41.262931 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:41Z","lastTransitionTime":"2025-12-11T02:02:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:41 crc kubenswrapper[4824]: I1211 02:02:41.365751 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:41 crc kubenswrapper[4824]: I1211 02:02:41.365803 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:41 crc kubenswrapper[4824]: I1211 02:02:41.365830 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:41 crc kubenswrapper[4824]: I1211 02:02:41.365861 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:41 crc kubenswrapper[4824]: I1211 02:02:41.365879 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:41Z","lastTransitionTime":"2025-12-11T02:02:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:41 crc kubenswrapper[4824]: I1211 02:02:41.468765 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:41 crc kubenswrapper[4824]: I1211 02:02:41.468824 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:41 crc kubenswrapper[4824]: I1211 02:02:41.468840 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:41 crc kubenswrapper[4824]: I1211 02:02:41.468863 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:41 crc kubenswrapper[4824]: I1211 02:02:41.468879 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:41Z","lastTransitionTime":"2025-12-11T02:02:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:41 crc kubenswrapper[4824]: I1211 02:02:41.571491 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:41 crc kubenswrapper[4824]: I1211 02:02:41.571561 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:41 crc kubenswrapper[4824]: I1211 02:02:41.571579 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:41 crc kubenswrapper[4824]: I1211 02:02:41.571603 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:41 crc kubenswrapper[4824]: I1211 02:02:41.571620 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:41Z","lastTransitionTime":"2025-12-11T02:02:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:41 crc kubenswrapper[4824]: I1211 02:02:41.631865 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 02:02:41 crc kubenswrapper[4824]: I1211 02:02:41.631954 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 02:02:41 crc kubenswrapper[4824]: I1211 02:02:41.632034 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-grnpt" Dec 11 02:02:41 crc kubenswrapper[4824]: I1211 02:02:41.632085 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 02:02:41 crc kubenswrapper[4824]: E1211 02:02:41.632374 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 02:02:41 crc kubenswrapper[4824]: E1211 02:02:41.632531 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 02:02:41 crc kubenswrapper[4824]: E1211 02:02:41.632608 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-grnpt" podUID="09a52948-0404-45ff-85b1-51479268bf71" Dec 11 02:02:41 crc kubenswrapper[4824]: E1211 02:02:41.632683 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 02:02:41 crc kubenswrapper[4824]: I1211 02:02:41.674999 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:41 crc kubenswrapper[4824]: I1211 02:02:41.675102 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:41 crc kubenswrapper[4824]: I1211 02:02:41.675190 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:41 crc kubenswrapper[4824]: I1211 02:02:41.675218 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:41 crc kubenswrapper[4824]: I1211 02:02:41.675288 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:41Z","lastTransitionTime":"2025-12-11T02:02:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:41 crc kubenswrapper[4824]: I1211 02:02:41.777953 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:41 crc kubenswrapper[4824]: I1211 02:02:41.778017 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:41 crc kubenswrapper[4824]: I1211 02:02:41.778036 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:41 crc kubenswrapper[4824]: I1211 02:02:41.778070 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:41 crc kubenswrapper[4824]: I1211 02:02:41.778089 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:41Z","lastTransitionTime":"2025-12-11T02:02:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:41 crc kubenswrapper[4824]: I1211 02:02:41.880326 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:41 crc kubenswrapper[4824]: I1211 02:02:41.880419 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:41 crc kubenswrapper[4824]: I1211 02:02:41.880462 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:41 crc kubenswrapper[4824]: I1211 02:02:41.880496 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:41 crc kubenswrapper[4824]: I1211 02:02:41.880516 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:41Z","lastTransitionTime":"2025-12-11T02:02:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:41 crc kubenswrapper[4824]: I1211 02:02:41.983496 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:41 crc kubenswrapper[4824]: I1211 02:02:41.983560 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:41 crc kubenswrapper[4824]: I1211 02:02:41.983576 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:41 crc kubenswrapper[4824]: I1211 02:02:41.983601 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:41 crc kubenswrapper[4824]: I1211 02:02:41.983620 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:41Z","lastTransitionTime":"2025-12-11T02:02:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:42 crc kubenswrapper[4824]: I1211 02:02:42.086616 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:42 crc kubenswrapper[4824]: I1211 02:02:42.086676 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:42 crc kubenswrapper[4824]: I1211 02:02:42.086692 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:42 crc kubenswrapper[4824]: I1211 02:02:42.086714 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:42 crc kubenswrapper[4824]: I1211 02:02:42.086728 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:42Z","lastTransitionTime":"2025-12-11T02:02:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:42 crc kubenswrapper[4824]: I1211 02:02:42.189760 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:42 crc kubenswrapper[4824]: I1211 02:02:42.189853 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:42 crc kubenswrapper[4824]: I1211 02:02:42.189869 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:42 crc kubenswrapper[4824]: I1211 02:02:42.189903 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:42 crc kubenswrapper[4824]: I1211 02:02:42.189916 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:42Z","lastTransitionTime":"2025-12-11T02:02:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:42 crc kubenswrapper[4824]: I1211 02:02:42.296992 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:42 crc kubenswrapper[4824]: I1211 02:02:42.297082 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:42 crc kubenswrapper[4824]: I1211 02:02:42.297101 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:42 crc kubenswrapper[4824]: I1211 02:02:42.297137 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:42 crc kubenswrapper[4824]: I1211 02:02:42.297150 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:42Z","lastTransitionTime":"2025-12-11T02:02:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:42 crc kubenswrapper[4824]: I1211 02:02:42.400777 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:42 crc kubenswrapper[4824]: I1211 02:02:42.400819 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:42 crc kubenswrapper[4824]: I1211 02:02:42.400828 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:42 crc kubenswrapper[4824]: I1211 02:02:42.400842 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:42 crc kubenswrapper[4824]: I1211 02:02:42.400850 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:42Z","lastTransitionTime":"2025-12-11T02:02:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:42 crc kubenswrapper[4824]: I1211 02:02:42.503323 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:42 crc kubenswrapper[4824]: I1211 02:02:42.503417 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:42 crc kubenswrapper[4824]: I1211 02:02:42.503439 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:42 crc kubenswrapper[4824]: I1211 02:02:42.503463 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:42 crc kubenswrapper[4824]: I1211 02:02:42.503482 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:42Z","lastTransitionTime":"2025-12-11T02:02:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:42 crc kubenswrapper[4824]: I1211 02:02:42.607270 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:42 crc kubenswrapper[4824]: I1211 02:02:42.607343 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:42 crc kubenswrapper[4824]: I1211 02:02:42.607394 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:42 crc kubenswrapper[4824]: I1211 02:02:42.607426 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:42 crc kubenswrapper[4824]: I1211 02:02:42.607450 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:42Z","lastTransitionTime":"2025-12-11T02:02:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:42 crc kubenswrapper[4824]: I1211 02:02:42.710606 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:42 crc kubenswrapper[4824]: I1211 02:02:42.710671 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:42 crc kubenswrapper[4824]: I1211 02:02:42.710688 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:42 crc kubenswrapper[4824]: I1211 02:02:42.710714 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:42 crc kubenswrapper[4824]: I1211 02:02:42.710732 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:42Z","lastTransitionTime":"2025-12-11T02:02:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:42 crc kubenswrapper[4824]: I1211 02:02:42.746080 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 02:02:42 crc kubenswrapper[4824]: I1211 02:02:42.746211 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 02:02:42 crc kubenswrapper[4824]: I1211 02:02:42.746244 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 02:02:42 crc kubenswrapper[4824]: I1211 02:02:42.746275 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 02:02:42 crc kubenswrapper[4824]: I1211 02:02:42.746296 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T02:02:42Z","lastTransitionTime":"2025-12-11T02:02:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 02:02:42 crc kubenswrapper[4824]: I1211 02:02:42.817331 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-qwl8k"] Dec 11 02:02:42 crc kubenswrapper[4824]: I1211 02:02:42.817929 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qwl8k" Dec 11 02:02:42 crc kubenswrapper[4824]: I1211 02:02:42.820629 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 11 02:02:42 crc kubenswrapper[4824]: I1211 02:02:42.820704 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 11 02:02:42 crc kubenswrapper[4824]: I1211 02:02:42.820853 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 11 02:02:42 crc kubenswrapper[4824]: I1211 02:02:42.820937 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 11 02:02:42 crc kubenswrapper[4824]: I1211 02:02:42.959257 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/b3a57e47-d4d1-45be-95d5-c9d3e568b4c2-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-qwl8k\" (UID: \"b3a57e47-d4d1-45be-95d5-c9d3e568b4c2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qwl8k" Dec 11 02:02:42 crc kubenswrapper[4824]: I1211 02:02:42.959316 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b3a57e47-d4d1-45be-95d5-c9d3e568b4c2-service-ca\") pod \"cluster-version-operator-5c965bbfc6-qwl8k\" (UID: \"b3a57e47-d4d1-45be-95d5-c9d3e568b4c2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qwl8k" Dec 11 02:02:42 crc kubenswrapper[4824]: I1211 02:02:42.959372 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b3a57e47-d4d1-45be-95d5-c9d3e568b4c2-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-qwl8k\" (UID: \"b3a57e47-d4d1-45be-95d5-c9d3e568b4c2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qwl8k" Dec 11 02:02:42 crc kubenswrapper[4824]: I1211 02:02:42.959625 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/b3a57e47-d4d1-45be-95d5-c9d3e568b4c2-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-qwl8k\" (UID: \"b3a57e47-d4d1-45be-95d5-c9d3e568b4c2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qwl8k" Dec 11 02:02:42 crc kubenswrapper[4824]: I1211 02:02:42.959743 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b3a57e47-d4d1-45be-95d5-c9d3e568b4c2-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-qwl8k\" (UID: \"b3a57e47-d4d1-45be-95d5-c9d3e568b4c2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qwl8k" Dec 11 02:02:43 crc kubenswrapper[4824]: I1211 02:02:43.060981 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b3a57e47-d4d1-45be-95d5-c9d3e568b4c2-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-qwl8k\" (UID: \"b3a57e47-d4d1-45be-95d5-c9d3e568b4c2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qwl8k" Dec 11 02:02:43 crc kubenswrapper[4824]: I1211 02:02:43.061079 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/b3a57e47-d4d1-45be-95d5-c9d3e568b4c2-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-qwl8k\" (UID: \"b3a57e47-d4d1-45be-95d5-c9d3e568b4c2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qwl8k" Dec 11 02:02:43 crc kubenswrapper[4824]: I1211 02:02:43.061166 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/b3a57e47-d4d1-45be-95d5-c9d3e568b4c2-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-qwl8k\" (UID: \"b3a57e47-d4d1-45be-95d5-c9d3e568b4c2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qwl8k" Dec 11 02:02:43 crc kubenswrapper[4824]: I1211 02:02:43.061110 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b3a57e47-d4d1-45be-95d5-c9d3e568b4c2-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-qwl8k\" (UID: \"b3a57e47-d4d1-45be-95d5-c9d3e568b4c2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qwl8k" Dec 11 02:02:43 crc kubenswrapper[4824]: I1211 02:02:43.061239 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/b3a57e47-d4d1-45be-95d5-c9d3e568b4c2-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-qwl8k\" (UID: \"b3a57e47-d4d1-45be-95d5-c9d3e568b4c2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qwl8k" Dec 11 02:02:43 crc kubenswrapper[4824]: I1211 02:02:43.061303 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b3a57e47-d4d1-45be-95d5-c9d3e568b4c2-service-ca\") pod \"cluster-version-operator-5c965bbfc6-qwl8k\" (UID: \"b3a57e47-d4d1-45be-95d5-c9d3e568b4c2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qwl8k" Dec 11 02:02:43 crc kubenswrapper[4824]: I1211 02:02:43.061281 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/b3a57e47-d4d1-45be-95d5-c9d3e568b4c2-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-qwl8k\" (UID: \"b3a57e47-d4d1-45be-95d5-c9d3e568b4c2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qwl8k" Dec 11 02:02:43 crc kubenswrapper[4824]: I1211 02:02:43.062192 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b3a57e47-d4d1-45be-95d5-c9d3e568b4c2-service-ca\") pod \"cluster-version-operator-5c965bbfc6-qwl8k\" (UID: \"b3a57e47-d4d1-45be-95d5-c9d3e568b4c2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qwl8k" Dec 11 02:02:43 crc kubenswrapper[4824]: I1211 02:02:43.070508 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b3a57e47-d4d1-45be-95d5-c9d3e568b4c2-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-qwl8k\" (UID: \"b3a57e47-d4d1-45be-95d5-c9d3e568b4c2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qwl8k" Dec 11 02:02:43 crc kubenswrapper[4824]: I1211 02:02:43.089003 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b3a57e47-d4d1-45be-95d5-c9d3e568b4c2-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-qwl8k\" (UID: \"b3a57e47-d4d1-45be-95d5-c9d3e568b4c2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qwl8k" Dec 11 02:02:43 crc kubenswrapper[4824]: I1211 02:02:43.134022 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qwl8k" Dec 11 02:02:43 crc kubenswrapper[4824]: I1211 02:02:43.320622 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qwl8k" event={"ID":"b3a57e47-d4d1-45be-95d5-c9d3e568b4c2","Type":"ContainerStarted","Data":"9abe45c8b5616b4c8847163b84b49a69591b1292f34798f6917baa91c1bcbaae"} Dec 11 02:02:43 crc kubenswrapper[4824]: I1211 02:02:43.321089 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qwl8k" event={"ID":"b3a57e47-d4d1-45be-95d5-c9d3e568b4c2","Type":"ContainerStarted","Data":"431d76deade365e997235860320912876aadff704ab75978fc556083ed0cbb67"} Dec 11 02:02:43 crc kubenswrapper[4824]: I1211 02:02:43.631961 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 02:02:43 crc kubenswrapper[4824]: I1211 02:02:43.632028 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 02:02:43 crc kubenswrapper[4824]: I1211 02:02:43.632038 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 02:02:43 crc kubenswrapper[4824]: I1211 02:02:43.631961 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-grnpt" Dec 11 02:02:43 crc kubenswrapper[4824]: E1211 02:02:43.632238 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 02:02:43 crc kubenswrapper[4824]: E1211 02:02:43.632342 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-grnpt" podUID="09a52948-0404-45ff-85b1-51479268bf71" Dec 11 02:02:43 crc kubenswrapper[4824]: E1211 02:02:43.632476 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 02:02:43 crc kubenswrapper[4824]: E1211 02:02:43.632595 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 02:02:45 crc kubenswrapper[4824]: I1211 02:02:45.631607 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-grnpt" Dec 11 02:02:45 crc kubenswrapper[4824]: I1211 02:02:45.631694 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 02:02:45 crc kubenswrapper[4824]: E1211 02:02:45.631739 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-grnpt" podUID="09a52948-0404-45ff-85b1-51479268bf71" Dec 11 02:02:45 crc kubenswrapper[4824]: I1211 02:02:45.631694 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 02:02:45 crc kubenswrapper[4824]: I1211 02:02:45.631859 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 02:02:45 crc kubenswrapper[4824]: E1211 02:02:45.632019 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 02:02:45 crc kubenswrapper[4824]: E1211 02:02:45.632197 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 02:02:45 crc kubenswrapper[4824]: E1211 02:02:45.632280 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 02:02:47 crc kubenswrapper[4824]: I1211 02:02:47.633345 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 02:02:47 crc kubenswrapper[4824]: E1211 02:02:47.633526 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 02:02:47 crc kubenswrapper[4824]: I1211 02:02:47.633901 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 02:02:47 crc kubenswrapper[4824]: E1211 02:02:47.634005 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 02:02:47 crc kubenswrapper[4824]: I1211 02:02:47.634300 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-grnpt" Dec 11 02:02:47 crc kubenswrapper[4824]: E1211 02:02:47.634403 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-grnpt" podUID="09a52948-0404-45ff-85b1-51479268bf71" Dec 11 02:02:47 crc kubenswrapper[4824]: I1211 02:02:47.634599 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 02:02:47 crc kubenswrapper[4824]: E1211 02:02:47.634685 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 02:02:49 crc kubenswrapper[4824]: I1211 02:02:49.632067 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 02:02:49 crc kubenswrapper[4824]: I1211 02:02:49.632153 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 02:02:49 crc kubenswrapper[4824]: I1211 02:02:49.632175 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-grnpt" Dec 11 02:02:49 crc kubenswrapper[4824]: I1211 02:02:49.632088 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 02:02:49 crc kubenswrapper[4824]: E1211 02:02:49.632307 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 02:02:49 crc kubenswrapper[4824]: E1211 02:02:49.632436 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 02:02:49 crc kubenswrapper[4824]: E1211 02:02:49.632584 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-grnpt" podUID="09a52948-0404-45ff-85b1-51479268bf71" Dec 11 02:02:49 crc kubenswrapper[4824]: E1211 02:02:49.632739 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 02:02:50 crc kubenswrapper[4824]: I1211 02:02:50.355003 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6bkc2_8022f696-d0e7-437b-848a-3cd25bd1f364/kube-multus/1.log" Dec 11 02:02:50 crc kubenswrapper[4824]: I1211 02:02:50.355946 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6bkc2_8022f696-d0e7-437b-848a-3cd25bd1f364/kube-multus/0.log" Dec 11 02:02:50 crc kubenswrapper[4824]: I1211 02:02:50.356073 4824 generic.go:334] "Generic (PLEG): container finished" podID="8022f696-d0e7-437b-848a-3cd25bd1f364" containerID="f40719f59881e077e7edbab3adb9efd101359bd461e5434c365e119e8cdbddb2" exitCode=1 Dec 11 02:02:50 crc kubenswrapper[4824]: I1211 02:02:50.356158 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-6bkc2" event={"ID":"8022f696-d0e7-437b-848a-3cd25bd1f364","Type":"ContainerDied","Data":"f40719f59881e077e7edbab3adb9efd101359bd461e5434c365e119e8cdbddb2"} Dec 11 02:02:50 crc kubenswrapper[4824]: I1211 02:02:50.356217 4824 scope.go:117] "RemoveContainer" containerID="f3e28f6778ec0dba6ea3e6b69e5a2c1c73c0d5db1abeead241fb58efdbfaefe4" Dec 11 02:02:50 crc kubenswrapper[4824]: I1211 02:02:50.356861 4824 scope.go:117] "RemoveContainer" containerID="f40719f59881e077e7edbab3adb9efd101359bd461e5434c365e119e8cdbddb2" Dec 11 02:02:50 crc kubenswrapper[4824]: E1211 02:02:50.357212 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-6bkc2_openshift-multus(8022f696-d0e7-437b-848a-3cd25bd1f364)\"" pod="openshift-multus/multus-6bkc2" podUID="8022f696-d0e7-437b-848a-3cd25bd1f364" Dec 11 02:02:50 crc kubenswrapper[4824]: I1211 02:02:50.383973 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qwl8k" podStartSLOduration=93.383952014 podStartE2EDuration="1m33.383952014s" podCreationTimestamp="2025-12-11 02:01:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:02:43.344267737 +0000 UTC m=+105.033305156" watchObservedRunningTime="2025-12-11 02:02:50.383952014 +0000 UTC m=+112.072989433" Dec 11 02:02:51 crc kubenswrapper[4824]: I1211 02:02:51.362310 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6bkc2_8022f696-d0e7-437b-848a-3cd25bd1f364/kube-multus/1.log" Dec 11 02:02:51 crc kubenswrapper[4824]: I1211 02:02:51.632181 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 02:02:51 crc kubenswrapper[4824]: E1211 02:02:51.632340 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 02:02:51 crc kubenswrapper[4824]: I1211 02:02:51.632415 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-grnpt" Dec 11 02:02:51 crc kubenswrapper[4824]: E1211 02:02:51.632564 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-grnpt" podUID="09a52948-0404-45ff-85b1-51479268bf71" Dec 11 02:02:51 crc kubenswrapper[4824]: I1211 02:02:51.632615 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 02:02:51 crc kubenswrapper[4824]: E1211 02:02:51.632966 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 02:02:51 crc kubenswrapper[4824]: I1211 02:02:51.633104 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 02:02:51 crc kubenswrapper[4824]: E1211 02:02:51.633306 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 02:02:51 crc kubenswrapper[4824]: I1211 02:02:51.634288 4824 scope.go:117] "RemoveContainer" containerID="154d82fd6acee749bfd37220547e52647dc7f9d86a73c28ce7d8577c9cadef6e" Dec 11 02:02:52 crc kubenswrapper[4824]: I1211 02:02:52.368185 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f6pgc_5b193bd2-b536-4056-92f1-94c9836ab2eb/ovnkube-controller/3.log" Dec 11 02:02:52 crc kubenswrapper[4824]: I1211 02:02:52.377818 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" event={"ID":"5b193bd2-b536-4056-92f1-94c9836ab2eb","Type":"ContainerStarted","Data":"8bb86ae130147feda9a5af751733489c3e0cc293e47e555879ae1fe0392c640e"} Dec 11 02:02:52 crc kubenswrapper[4824]: I1211 02:02:52.378591 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" Dec 11 02:02:52 crc kubenswrapper[4824]: I1211 02:02:52.417167 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" podStartSLOduration=95.417152565 podStartE2EDuration="1m35.417152565s" podCreationTimestamp="2025-12-11 02:01:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:02:52.416362823 +0000 UTC m=+114.105400222" watchObservedRunningTime="2025-12-11 02:02:52.417152565 +0000 UTC m=+114.106189964" Dec 11 02:02:52 crc kubenswrapper[4824]: I1211 02:02:52.589508 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-grnpt"] Dec 11 02:02:52 crc kubenswrapper[4824]: I1211 02:02:52.589690 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-grnpt" Dec 11 02:02:52 crc kubenswrapper[4824]: E1211 02:02:52.589817 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-grnpt" podUID="09a52948-0404-45ff-85b1-51479268bf71" Dec 11 02:02:53 crc kubenswrapper[4824]: I1211 02:02:53.631914 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 02:02:53 crc kubenswrapper[4824]: I1211 02:02:53.631945 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 02:02:53 crc kubenswrapper[4824]: I1211 02:02:53.631962 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 02:02:53 crc kubenswrapper[4824]: E1211 02:02:53.632056 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 02:02:53 crc kubenswrapper[4824]: E1211 02:02:53.632318 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 02:02:53 crc kubenswrapper[4824]: E1211 02:02:53.632492 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 02:02:54 crc kubenswrapper[4824]: I1211 02:02:54.632232 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-grnpt" Dec 11 02:02:54 crc kubenswrapper[4824]: E1211 02:02:54.632823 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-grnpt" podUID="09a52948-0404-45ff-85b1-51479268bf71" Dec 11 02:02:55 crc kubenswrapper[4824]: I1211 02:02:55.631670 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 02:02:55 crc kubenswrapper[4824]: I1211 02:02:55.631706 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 02:02:55 crc kubenswrapper[4824]: E1211 02:02:55.631776 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 02:02:55 crc kubenswrapper[4824]: E1211 02:02:55.631845 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 02:02:55 crc kubenswrapper[4824]: I1211 02:02:55.631869 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 02:02:55 crc kubenswrapper[4824]: E1211 02:02:55.632048 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 02:02:56 crc kubenswrapper[4824]: I1211 02:02:56.632358 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-grnpt" Dec 11 02:02:56 crc kubenswrapper[4824]: E1211 02:02:56.632564 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-grnpt" podUID="09a52948-0404-45ff-85b1-51479268bf71" Dec 11 02:02:57 crc kubenswrapper[4824]: I1211 02:02:57.632342 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 02:02:57 crc kubenswrapper[4824]: I1211 02:02:57.632368 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 02:02:57 crc kubenswrapper[4824]: E1211 02:02:57.632518 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 02:02:57 crc kubenswrapper[4824]: I1211 02:02:57.632562 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 02:02:57 crc kubenswrapper[4824]: E1211 02:02:57.632804 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 02:02:57 crc kubenswrapper[4824]: E1211 02:02:57.632865 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 02:02:58 crc kubenswrapper[4824]: E1211 02:02:58.608024 4824 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Dec 11 02:02:58 crc kubenswrapper[4824]: I1211 02:02:58.632414 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-grnpt" Dec 11 02:02:58 crc kubenswrapper[4824]: E1211 02:02:58.634575 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-grnpt" podUID="09a52948-0404-45ff-85b1-51479268bf71" Dec 11 02:02:58 crc kubenswrapper[4824]: E1211 02:02:58.763843 4824 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 11 02:02:59 crc kubenswrapper[4824]: I1211 02:02:59.632179 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 02:02:59 crc kubenswrapper[4824]: E1211 02:02:59.632363 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 02:02:59 crc kubenswrapper[4824]: I1211 02:02:59.632623 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 02:02:59 crc kubenswrapper[4824]: I1211 02:02:59.632664 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 02:02:59 crc kubenswrapper[4824]: E1211 02:02:59.632750 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 02:02:59 crc kubenswrapper[4824]: E1211 02:02:59.632981 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 02:03:00 crc kubenswrapper[4824]: I1211 02:03:00.632489 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-grnpt" Dec 11 02:03:00 crc kubenswrapper[4824]: E1211 02:03:00.632699 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-grnpt" podUID="09a52948-0404-45ff-85b1-51479268bf71" Dec 11 02:03:01 crc kubenswrapper[4824]: I1211 02:03:01.631895 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 02:03:01 crc kubenswrapper[4824]: I1211 02:03:01.632211 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 02:03:01 crc kubenswrapper[4824]: E1211 02:03:01.632358 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 02:03:01 crc kubenswrapper[4824]: E1211 02:03:01.632567 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 02:03:01 crc kubenswrapper[4824]: I1211 02:03:01.632643 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 02:03:01 crc kubenswrapper[4824]: E1211 02:03:01.632741 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 02:03:02 crc kubenswrapper[4824]: I1211 02:03:02.632465 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-grnpt" Dec 11 02:03:02 crc kubenswrapper[4824]: E1211 02:03:02.632621 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-grnpt" podUID="09a52948-0404-45ff-85b1-51479268bf71" Dec 11 02:03:03 crc kubenswrapper[4824]: I1211 02:03:03.632463 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 02:03:03 crc kubenswrapper[4824]: I1211 02:03:03.632531 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 02:03:03 crc kubenswrapper[4824]: I1211 02:03:03.632691 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 02:03:03 crc kubenswrapper[4824]: E1211 02:03:03.632880 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 02:03:03 crc kubenswrapper[4824]: E1211 02:03:03.632966 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 02:03:03 crc kubenswrapper[4824]: E1211 02:03:03.633020 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 02:03:03 crc kubenswrapper[4824]: E1211 02:03:03.765880 4824 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 11 02:03:04 crc kubenswrapper[4824]: I1211 02:03:04.632178 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-grnpt" Dec 11 02:03:04 crc kubenswrapper[4824]: E1211 02:03:04.632388 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-grnpt" podUID="09a52948-0404-45ff-85b1-51479268bf71" Dec 11 02:03:04 crc kubenswrapper[4824]: I1211 02:03:04.633035 4824 scope.go:117] "RemoveContainer" containerID="f40719f59881e077e7edbab3adb9efd101359bd461e5434c365e119e8cdbddb2" Dec 11 02:03:05 crc kubenswrapper[4824]: I1211 02:03:05.427788 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6bkc2_8022f696-d0e7-437b-848a-3cd25bd1f364/kube-multus/1.log" Dec 11 02:03:05 crc kubenswrapper[4824]: I1211 02:03:05.428221 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-6bkc2" event={"ID":"8022f696-d0e7-437b-848a-3cd25bd1f364","Type":"ContainerStarted","Data":"468cb1b4a3153f6a5f696a8cf76b784ea23f5a51db91c4c28a6994cbe54bd55c"} Dec 11 02:03:05 crc kubenswrapper[4824]: I1211 02:03:05.632282 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 02:03:05 crc kubenswrapper[4824]: I1211 02:03:05.632347 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 02:03:05 crc kubenswrapper[4824]: I1211 02:03:05.632363 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 02:03:05 crc kubenswrapper[4824]: E1211 02:03:05.632484 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 02:03:05 crc kubenswrapper[4824]: E1211 02:03:05.632613 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 02:03:05 crc kubenswrapper[4824]: E1211 02:03:05.632743 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 02:03:06 crc kubenswrapper[4824]: I1211 02:03:06.631647 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-grnpt" Dec 11 02:03:06 crc kubenswrapper[4824]: E1211 02:03:06.631842 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-grnpt" podUID="09a52948-0404-45ff-85b1-51479268bf71" Dec 11 02:03:07 crc kubenswrapper[4824]: I1211 02:03:07.632438 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 02:03:07 crc kubenswrapper[4824]: I1211 02:03:07.632530 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 02:03:07 crc kubenswrapper[4824]: I1211 02:03:07.632438 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 02:03:07 crc kubenswrapper[4824]: E1211 02:03:07.632595 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 02:03:07 crc kubenswrapper[4824]: E1211 02:03:07.632704 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 02:03:07 crc kubenswrapper[4824]: E1211 02:03:07.632830 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 02:03:08 crc kubenswrapper[4824]: I1211 02:03:08.632390 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-grnpt" Dec 11 02:03:08 crc kubenswrapper[4824]: E1211 02:03:08.634242 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-grnpt" podUID="09a52948-0404-45ff-85b1-51479268bf71" Dec 11 02:03:09 crc kubenswrapper[4824]: I1211 02:03:09.632420 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 02:03:09 crc kubenswrapper[4824]: I1211 02:03:09.632468 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 02:03:09 crc kubenswrapper[4824]: I1211 02:03:09.632570 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 02:03:09 crc kubenswrapper[4824]: I1211 02:03:09.638032 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 11 02:03:09 crc kubenswrapper[4824]: I1211 02:03:09.640163 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 11 02:03:09 crc kubenswrapper[4824]: I1211 02:03:09.640421 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 11 02:03:09 crc kubenswrapper[4824]: I1211 02:03:09.641956 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 11 02:03:10 crc kubenswrapper[4824]: I1211 02:03:10.632465 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-grnpt" Dec 11 02:03:10 crc kubenswrapper[4824]: I1211 02:03:10.635626 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 11 02:03:10 crc kubenswrapper[4824]: I1211 02:03:10.635807 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.142940 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.192358 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-g692d"] Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.193074 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g692d" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.194500 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-qr5xz"] Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.196484 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-qr5xz" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.198420 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-msljv"] Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.199189 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-msljv" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.201056 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-vtdlb"] Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.207045 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.207703 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.208947 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.209521 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.209668 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vtdlb" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.209704 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.209781 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.209811 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.210714 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.209938 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.210168 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.210179 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.210181 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.214697 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-l9d4f"] Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.215339 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2vcc6"] Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.215740 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-tfwnz"] Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.216224 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-tfwnz" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.216665 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-l9d4f" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.217069 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.217178 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2vcc6" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.268441 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.269404 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.271159 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.271238 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.271321 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.271790 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-lfxkq"] Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.272280 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-krqhc"] Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.272641 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-w6b4l"] Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.273030 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-tz2j9"] Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.273185 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-lfxkq" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.273867 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-krqhc" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.273899 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.274155 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.274793 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.275160 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.275338 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.275365 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.275623 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.275722 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.275830 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.275957 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.275974 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.275997 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.276302 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-w6b4l" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.276874 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-k6kls"] Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.277202 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-2ng4n"] Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.277381 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.277459 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-kkw8v"] Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.277813 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-g692d"] Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.277841 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-qr5xz"] Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.277897 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-kkw8v" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.278085 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lg9s6"] Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.278120 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-tz2j9" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.278280 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-k6kls" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.278338 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-2ng4n" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.278758 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lg9s6" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.279260 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.279418 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.281358 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.281567 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.281679 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.281790 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.281902 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.282024 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.282486 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.282669 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.282702 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.282833 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.282952 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.283435 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.283559 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.283684 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.283858 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.284005 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.284167 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.284334 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.284814 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-l88b4"] Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.285576 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-l88b4" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.288618 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.288832 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.289259 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.289410 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.289701 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.290034 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.290398 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.290790 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.290888 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-d89dq"] Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.291459 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-d89dq" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.291686 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-r89c2"] Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.291971 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-r89c2" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.294448 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gndwh"] Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.295130 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gndwh" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.312613 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.313466 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.324894 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.325326 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.325564 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.325897 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.326237 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.326559 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.326825 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.327094 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.327332 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.327581 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.327852 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.327137 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.328617 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.328789 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.329097 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.329607 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.330465 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.325389 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.330945 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.331545 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.322385 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-254w2"] Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.332630 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-254w2" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.332910 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.333184 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.333446 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.327418 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.330863 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.334397 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.334465 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.335251 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.335267 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.336525 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-hxqw8"] Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.337001 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ks7hm"] Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.340003 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.340214 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.340450 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.342843 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.343991 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.344254 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.344580 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hxqw8" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.345732 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2zq2n"] Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.346156 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2zq2n" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.346393 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ks7hm" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.349479 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bdf5eb2-fe05-4e24-9e00-43c72afc3583-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-l9d4f\" (UID: \"1bdf5eb2-fe05-4e24-9e00-43c72afc3583\") " pod="openshift-authentication/oauth-openshift-558db77b4-l9d4f" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.349512 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3404cb6f-6c3a-4e0a-99a4-27530beefa83-auth-proxy-config\") pod \"machine-approver-56656f9798-vtdlb\" (UID: \"3404cb6f-6c3a-4e0a-99a4-27530beefa83\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vtdlb" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.349532 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-874kt\" (UniqueName: \"kubernetes.io/projected/3404cb6f-6c3a-4e0a-99a4-27530beefa83-kube-api-access-874kt\") pod \"machine-approver-56656f9798-vtdlb\" (UID: \"3404cb6f-6c3a-4e0a-99a4-27530beefa83\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vtdlb" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.349550 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-htnkn\" (UniqueName: \"kubernetes.io/projected/29e1953f-668a-4393-82ab-a45fa4b79ec7-kube-api-access-htnkn\") pod \"openshift-apiserver-operator-796bbdcf4f-2vcc6\" (UID: \"29e1953f-668a-4393-82ab-a45fa4b79ec7\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2vcc6" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.349569 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3e0e52dd-503e-47a4-b68e-3f768ccecfc6-serving-cert\") pod \"apiserver-7bbb656c7d-g692d\" (UID: \"3e0e52dd-503e-47a4-b68e-3f768ccecfc6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g692d" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.349960 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5407364a-e2b6-43b4-9eb0-deb12a022289-client-ca\") pod \"route-controller-manager-6576b87f9c-msljv\" (UID: \"5407364a-e2b6-43b4-9eb0-deb12a022289\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-msljv" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.350074 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n28fr\" (UniqueName: \"kubernetes.io/projected/1bdf5eb2-fe05-4e24-9e00-43c72afc3583-kube-api-access-n28fr\") pod \"oauth-openshift-558db77b4-l9d4f\" (UID: \"1bdf5eb2-fe05-4e24-9e00-43c72afc3583\") " pod="openshift-authentication/oauth-openshift-558db77b4-l9d4f" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.350178 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/3404cb6f-6c3a-4e0a-99a4-27530beefa83-machine-approver-tls\") pod \"machine-approver-56656f9798-vtdlb\" (UID: \"3404cb6f-6c3a-4e0a-99a4-27530beefa83\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vtdlb" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.350214 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-grgzk\" (UniqueName: \"kubernetes.io/projected/f257970f-c1ce-44fb-b207-a78126794cfe-kube-api-access-grgzk\") pod \"authentication-operator-69f744f599-tfwnz\" (UID: \"f257970f-c1ce-44fb-b207-a78126794cfe\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-tfwnz" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.350294 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e099899-aaf7-4bd8-b37e-8616b5060e33-config\") pod \"controller-manager-879f6c89f-qr5xz\" (UID: \"9e099899-aaf7-4bd8-b37e-8616b5060e33\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qr5xz" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.350371 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/1bdf5eb2-fe05-4e24-9e00-43c72afc3583-audit-policies\") pod \"oauth-openshift-558db77b4-l9d4f\" (UID: \"1bdf5eb2-fe05-4e24-9e00-43c72afc3583\") " pod="openshift-authentication/oauth-openshift-558db77b4-l9d4f" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.350458 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3404cb6f-6c3a-4e0a-99a4-27530beefa83-config\") pod \"machine-approver-56656f9798-vtdlb\" (UID: \"3404cb6f-6c3a-4e0a-99a4-27530beefa83\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vtdlb" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.350580 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f257970f-c1ce-44fb-b207-a78126794cfe-serving-cert\") pod \"authentication-operator-69f744f599-tfwnz\" (UID: \"f257970f-c1ce-44fb-b207-a78126794cfe\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-tfwnz" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.350659 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6587n\" (UniqueName: \"kubernetes.io/projected/3e0e52dd-503e-47a4-b68e-3f768ccecfc6-kube-api-access-6587n\") pod \"apiserver-7bbb656c7d-g692d\" (UID: \"3e0e52dd-503e-47a4-b68e-3f768ccecfc6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g692d" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.350751 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f257970f-c1ce-44fb-b207-a78126794cfe-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-tfwnz\" (UID: \"f257970f-c1ce-44fb-b207-a78126794cfe\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-tfwnz" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.350886 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5407364a-e2b6-43b4-9eb0-deb12a022289-config\") pod \"route-controller-manager-6576b87f9c-msljv\" (UID: \"5407364a-e2b6-43b4-9eb0-deb12a022289\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-msljv" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.350946 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/1bdf5eb2-fe05-4e24-9e00-43c72afc3583-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-l9d4f\" (UID: \"1bdf5eb2-fe05-4e24-9e00-43c72afc3583\") " pod="openshift-authentication/oauth-openshift-558db77b4-l9d4f" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.350966 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f257970f-c1ce-44fb-b207-a78126794cfe-service-ca-bundle\") pod \"authentication-operator-69f744f599-tfwnz\" (UID: \"f257970f-c1ce-44fb-b207-a78126794cfe\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-tfwnz" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.351010 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/3e0e52dd-503e-47a4-b68e-3f768ccecfc6-encryption-config\") pod \"apiserver-7bbb656c7d-g692d\" (UID: \"3e0e52dd-503e-47a4-b68e-3f768ccecfc6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g692d" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.351033 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/1bdf5eb2-fe05-4e24-9e00-43c72afc3583-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-l9d4f\" (UID: \"1bdf5eb2-fe05-4e24-9e00-43c72afc3583\") " pod="openshift-authentication/oauth-openshift-558db77b4-l9d4f" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.351056 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/1bdf5eb2-fe05-4e24-9e00-43c72afc3583-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-l9d4f\" (UID: \"1bdf5eb2-fe05-4e24-9e00-43c72afc3583\") " pod="openshift-authentication/oauth-openshift-558db77b4-l9d4f" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.351137 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1bdf5eb2-fe05-4e24-9e00-43c72afc3583-audit-dir\") pod \"oauth-openshift-558db77b4-l9d4f\" (UID: \"1bdf5eb2-fe05-4e24-9e00-43c72afc3583\") " pod="openshift-authentication/oauth-openshift-558db77b4-l9d4f" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.351162 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/1bdf5eb2-fe05-4e24-9e00-43c72afc3583-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-l9d4f\" (UID: \"1bdf5eb2-fe05-4e24-9e00-43c72afc3583\") " pod="openshift-authentication/oauth-openshift-558db77b4-l9d4f" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.351182 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3e0e52dd-503e-47a4-b68e-3f768ccecfc6-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-g692d\" (UID: \"3e0e52dd-503e-47a4-b68e-3f768ccecfc6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g692d" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.351224 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/1bdf5eb2-fe05-4e24-9e00-43c72afc3583-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-l9d4f\" (UID: \"1bdf5eb2-fe05-4e24-9e00-43c72afc3583\") " pod="openshift-authentication/oauth-openshift-558db77b4-l9d4f" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.351247 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/3e0e52dd-503e-47a4-b68e-3f768ccecfc6-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-g692d\" (UID: \"3e0e52dd-503e-47a4-b68e-3f768ccecfc6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g692d" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.351307 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/3e0e52dd-503e-47a4-b68e-3f768ccecfc6-etcd-client\") pod \"apiserver-7bbb656c7d-g692d\" (UID: \"3e0e52dd-503e-47a4-b68e-3f768ccecfc6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g692d" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.351333 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4x4bd\" (UniqueName: \"kubernetes.io/projected/5407364a-e2b6-43b4-9eb0-deb12a022289-kube-api-access-4x4bd\") pod \"route-controller-manager-6576b87f9c-msljv\" (UID: \"5407364a-e2b6-43b4-9eb0-deb12a022289\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-msljv" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.351352 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/29e1953f-668a-4393-82ab-a45fa4b79ec7-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-2vcc6\" (UID: \"29e1953f-668a-4393-82ab-a45fa4b79ec7\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2vcc6" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.351401 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/29e1953f-668a-4393-82ab-a45fa4b79ec7-config\") pod \"openshift-apiserver-operator-796bbdcf4f-2vcc6\" (UID: \"29e1953f-668a-4393-82ab-a45fa4b79ec7\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2vcc6" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.351470 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9e099899-aaf7-4bd8-b37e-8616b5060e33-serving-cert\") pod \"controller-manager-879f6c89f-qr5xz\" (UID: \"9e099899-aaf7-4bd8-b37e-8616b5060e33\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qr5xz" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.351502 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/1bdf5eb2-fe05-4e24-9e00-43c72afc3583-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-l9d4f\" (UID: \"1bdf5eb2-fe05-4e24-9e00-43c72afc3583\") " pod="openshift-authentication/oauth-openshift-558db77b4-l9d4f" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.351563 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5407364a-e2b6-43b4-9eb0-deb12a022289-serving-cert\") pod \"route-controller-manager-6576b87f9c-msljv\" (UID: \"5407364a-e2b6-43b4-9eb0-deb12a022289\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-msljv" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.351589 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9e099899-aaf7-4bd8-b37e-8616b5060e33-client-ca\") pod \"controller-manager-879f6c89f-qr5xz\" (UID: \"9e099899-aaf7-4bd8-b37e-8616b5060e33\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qr5xz" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.351636 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f257970f-c1ce-44fb-b207-a78126794cfe-config\") pod \"authentication-operator-69f744f599-tfwnz\" (UID: \"f257970f-c1ce-44fb-b207-a78126794cfe\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-tfwnz" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.351662 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6r6md\" (UniqueName: \"kubernetes.io/projected/9e099899-aaf7-4bd8-b37e-8616b5060e33-kube-api-access-6r6md\") pod \"controller-manager-879f6c89f-qr5xz\" (UID: \"9e099899-aaf7-4bd8-b37e-8616b5060e33\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qr5xz" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.351676 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3e0e52dd-503e-47a4-b68e-3f768ccecfc6-audit-policies\") pod \"apiserver-7bbb656c7d-g692d\" (UID: \"3e0e52dd-503e-47a4-b68e-3f768ccecfc6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g692d" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.351728 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/1bdf5eb2-fe05-4e24-9e00-43c72afc3583-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-l9d4f\" (UID: \"1bdf5eb2-fe05-4e24-9e00-43c72afc3583\") " pod="openshift-authentication/oauth-openshift-558db77b4-l9d4f" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.351747 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/1bdf5eb2-fe05-4e24-9e00-43c72afc3583-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-l9d4f\" (UID: \"1bdf5eb2-fe05-4e24-9e00-43c72afc3583\") " pod="openshift-authentication/oauth-openshift-558db77b4-l9d4f" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.351765 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/1bdf5eb2-fe05-4e24-9e00-43c72afc3583-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-l9d4f\" (UID: \"1bdf5eb2-fe05-4e24-9e00-43c72afc3583\") " pod="openshift-authentication/oauth-openshift-558db77b4-l9d4f" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.351801 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/1bdf5eb2-fe05-4e24-9e00-43c72afc3583-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-l9d4f\" (UID: \"1bdf5eb2-fe05-4e24-9e00-43c72afc3583\") " pod="openshift-authentication/oauth-openshift-558db77b4-l9d4f" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.351819 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9e099899-aaf7-4bd8-b37e-8616b5060e33-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-qr5xz\" (UID: \"9e099899-aaf7-4bd8-b37e-8616b5060e33\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qr5xz" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.351836 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3e0e52dd-503e-47a4-b68e-3f768ccecfc6-audit-dir\") pod \"apiserver-7bbb656c7d-g692d\" (UID: \"3e0e52dd-503e-47a4-b68e-3f768ccecfc6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g692d" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.365991 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.366988 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.367185 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.367582 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.369231 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.369903 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.370896 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.373000 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.373091 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.373876 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kdzkc"] Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.374369 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kdzkc" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.375070 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.375091 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-jlqk5"] Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.375664 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-l9d4f"] Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.376173 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-jlqk5" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.379055 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-z42g5"] Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.379408 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-z42g5" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.382174 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-pg8rn"] Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.382518 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-pg8rn" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.383004 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-66mv8"] Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.383347 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-66mv8" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.384545 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-tfwnz"] Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.385549 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-lfxkq"] Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.386935 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2vcc6"] Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.387852 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-2ng4n"] Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.390962 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-4r8x8"] Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.393322 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p8x2v"] Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.395365 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-4r8x8" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.403272 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.403676 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-qp2jl"] Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.403915 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p8x2v" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.404353 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423640-42vd6"] Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.404772 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qqg25"] Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.405194 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qqg25" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.405258 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-qp2jl" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.405472 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423640-42vd6" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.405513 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-w6b4l"] Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.406593 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-59bgf"] Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.407231 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-59bgf" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.409890 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-msljv"] Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.409923 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-lpgf5"] Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.410879 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-vkdbw"] Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.411373 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-gp8dt"] Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.411719 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-gp8dt" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.411872 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-lpgf5" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.411991 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vkdbw" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.412162 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-8fksd"] Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.412790 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8fksd" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.414862 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.417799 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-zvcj8"] Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.418279 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-zvcj8" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.421529 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-tz2j9"] Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.423088 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-l88b4"] Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.423159 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.424383 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lg9s6"] Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.426354 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-k6kls"] Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.428412 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gndwh"] Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.430078 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ks7hm"] Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.431263 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-kkw8v"] Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.432661 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-krqhc"] Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.434728 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-r89c2"] Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.435783 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-hxqw8"] Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.436674 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.440177 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-vkdbw"] Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.440804 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-z42g5"] Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.442038 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-59bgf"] Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.445974 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-d89dq"] Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.447266 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-254w2"] Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.448848 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-4r8x8"] Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.449613 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423640-42vd6"] Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.450527 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-gp8dt"] Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.452444 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2zq2n"] Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.452647 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-hr4qs"] Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.452746 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/1bdf5eb2-fe05-4e24-9e00-43c72afc3583-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-l9d4f\" (UID: \"1bdf5eb2-fe05-4e24-9e00-43c72afc3583\") " pod="openshift-authentication/oauth-openshift-558db77b4-l9d4f" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.452767 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f257970f-c1ce-44fb-b207-a78126794cfe-service-ca-bundle\") pod \"authentication-operator-69f744f599-tfwnz\" (UID: \"f257970f-c1ce-44fb-b207-a78126794cfe\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-tfwnz" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.452787 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/3e0e52dd-503e-47a4-b68e-3f768ccecfc6-encryption-config\") pod \"apiserver-7bbb656c7d-g692d\" (UID: \"3e0e52dd-503e-47a4-b68e-3f768ccecfc6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g692d" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.452806 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4kfvr\" (UniqueName: \"kubernetes.io/projected/8e2095d8-25c6-407f-aadb-f4603ae44d15-kube-api-access-4kfvr\") pod \"cluster-samples-operator-665b6dd947-krqhc\" (UID: \"8e2095d8-25c6-407f-aadb-f4603ae44d15\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-krqhc" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.452822 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb4dde0f-222d-4609-b1c9-fed653668e3a-config\") pod \"console-operator-58897d9998-lfxkq\" (UID: \"bb4dde0f-222d-4609-b1c9-fed653668e3a\") " pod="openshift-console-operator/console-operator-58897d9998-lfxkq" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.452837 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/1bdf5eb2-fe05-4e24-9e00-43c72afc3583-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-l9d4f\" (UID: \"1bdf5eb2-fe05-4e24-9e00-43c72afc3583\") " pod="openshift-authentication/oauth-openshift-558db77b4-l9d4f" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.452854 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fb52538d-050a-4b69-91b2-aa48f9833fe7-trusted-ca\") pod \"ingress-operator-5b745b69d9-hxqw8\" (UID: \"fb52538d-050a-4b69-91b2-aa48f9833fe7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hxqw8" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.452868 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cgv6v\" (UniqueName: \"kubernetes.io/projected/fb52538d-050a-4b69-91b2-aa48f9833fe7-kube-api-access-cgv6v\") pod \"ingress-operator-5b745b69d9-hxqw8\" (UID: \"fb52538d-050a-4b69-91b2-aa48f9833fe7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hxqw8" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.452883 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e87b6d85-2d13-4b72-bb6b-8299b6c72d0a-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-gndwh\" (UID: \"e87b6d85-2d13-4b72-bb6b-8299b6c72d0a\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gndwh" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.452899 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/75da7b9a-7085-492b-954a-6bb920fbb186-serving-cert\") pod \"apiserver-76f77b778f-l88b4\" (UID: \"75da7b9a-7085-492b-954a-6bb920fbb186\") " pod="openshift-apiserver/apiserver-76f77b778f-l88b4" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.452914 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/1bdf5eb2-fe05-4e24-9e00-43c72afc3583-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-l9d4f\" (UID: \"1bdf5eb2-fe05-4e24-9e00-43c72afc3583\") " pod="openshift-authentication/oauth-openshift-558db77b4-l9d4f" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.452929 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1bdf5eb2-fe05-4e24-9e00-43c72afc3583-audit-dir\") pod \"oauth-openshift-558db77b4-l9d4f\" (UID: \"1bdf5eb2-fe05-4e24-9e00-43c72afc3583\") " pod="openshift-authentication/oauth-openshift-558db77b4-l9d4f" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.452944 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/1bdf5eb2-fe05-4e24-9e00-43c72afc3583-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-l9d4f\" (UID: \"1bdf5eb2-fe05-4e24-9e00-43c72afc3583\") " pod="openshift-authentication/oauth-openshift-558db77b4-l9d4f" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.452960 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3e0e52dd-503e-47a4-b68e-3f768ccecfc6-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-g692d\" (UID: \"3e0e52dd-503e-47a4-b68e-3f768ccecfc6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g692d" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.452980 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/1bdf5eb2-fe05-4e24-9e00-43c72afc3583-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-l9d4f\" (UID: \"1bdf5eb2-fe05-4e24-9e00-43c72afc3583\") " pod="openshift-authentication/oauth-openshift-558db77b4-l9d4f" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.452995 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/3e0e52dd-503e-47a4-b68e-3f768ccecfc6-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-g692d\" (UID: \"3e0e52dd-503e-47a4-b68e-3f768ccecfc6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g692d" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.453011 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/247d17b7-7064-4ef1-9aed-c205fec84bc2-config\") pod \"kube-apiserver-operator-766d6c64bb-2zq2n\" (UID: \"247d17b7-7064-4ef1-9aed-c205fec84bc2\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2zq2n" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.453036 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/3e0e52dd-503e-47a4-b68e-3f768ccecfc6-etcd-client\") pod \"apiserver-7bbb656c7d-g692d\" (UID: \"3e0e52dd-503e-47a4-b68e-3f768ccecfc6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g692d" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.453053 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zzw5l\" (UniqueName: \"kubernetes.io/projected/7704cded-5ba1-4474-8a31-c0ba947b7679-kube-api-access-zzw5l\") pod \"machine-api-operator-5694c8668f-d89dq\" (UID: \"7704cded-5ba1-4474-8a31-c0ba947b7679\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-d89dq" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.453071 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4x4bd\" (UniqueName: \"kubernetes.io/projected/5407364a-e2b6-43b4-9eb0-deb12a022289-kube-api-access-4x4bd\") pod \"route-controller-manager-6576b87f9c-msljv\" (UID: \"5407364a-e2b6-43b4-9eb0-deb12a022289\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-msljv" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.453101 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/29e1953f-668a-4393-82ab-a45fa4b79ec7-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-2vcc6\" (UID: \"29e1953f-668a-4393-82ab-a45fa4b79ec7\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2vcc6" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.453132 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ad510a18-cd72-4cf0-9b0c-dd3540d491fa-trusted-ca-bundle\") pod \"console-f9d7485db-k6kls\" (UID: \"ad510a18-cd72-4cf0-9b0c-dd3540d491fa\") " pod="openshift-console/console-f9d7485db-k6kls" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.453148 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/29e1953f-668a-4393-82ab-a45fa4b79ec7-config\") pod \"openshift-apiserver-operator-796bbdcf4f-2vcc6\" (UID: \"29e1953f-668a-4393-82ab-a45fa4b79ec7\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2vcc6" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.453166 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fa007484-938d-460e-9c5e-c3f5b31a25d2-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ks7hm\" (UID: \"fa007484-938d-460e-9c5e-c3f5b31a25d2\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ks7hm" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.453181 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/75da7b9a-7085-492b-954a-6bb920fbb186-image-import-ca\") pod \"apiserver-76f77b778f-l88b4\" (UID: \"75da7b9a-7085-492b-954a-6bb920fbb186\") " pod="openshift-apiserver/apiserver-76f77b778f-l88b4" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.453200 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9e099899-aaf7-4bd8-b37e-8616b5060e33-serving-cert\") pod \"controller-manager-879f6c89f-qr5xz\" (UID: \"9e099899-aaf7-4bd8-b37e-8616b5060e33\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qr5xz" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.453217 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/1bdf5eb2-fe05-4e24-9e00-43c72afc3583-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-l9d4f\" (UID: \"1bdf5eb2-fe05-4e24-9e00-43c72afc3583\") " pod="openshift-authentication/oauth-openshift-558db77b4-l9d4f" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.453239 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5407364a-e2b6-43b4-9eb0-deb12a022289-serving-cert\") pod \"route-controller-manager-6576b87f9c-msljv\" (UID: \"5407364a-e2b6-43b4-9eb0-deb12a022289\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-msljv" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.453255 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9e099899-aaf7-4bd8-b37e-8616b5060e33-client-ca\") pod \"controller-manager-879f6c89f-qr5xz\" (UID: \"9e099899-aaf7-4bd8-b37e-8616b5060e33\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qr5xz" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.453272 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d5fb9349-891c-447a-bb5a-5e18a0ce9f92-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-lg9s6\" (UID: \"d5fb9349-891c-447a-bb5a-5e18a0ce9f92\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lg9s6" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.453312 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7js74\" (UniqueName: \"kubernetes.io/projected/d5fb9349-891c-447a-bb5a-5e18a0ce9f92-kube-api-access-7js74\") pod \"cluster-image-registry-operator-dc59b4c8b-lg9s6\" (UID: \"d5fb9349-891c-447a-bb5a-5e18a0ce9f92\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lg9s6" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.453329 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f257970f-c1ce-44fb-b207-a78126794cfe-config\") pod \"authentication-operator-69f744f599-tfwnz\" (UID: \"f257970f-c1ce-44fb-b207-a78126794cfe\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-tfwnz" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.453345 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6r6md\" (UniqueName: \"kubernetes.io/projected/9e099899-aaf7-4bd8-b37e-8616b5060e33-kube-api-access-6r6md\") pod \"controller-manager-879f6c89f-qr5xz\" (UID: \"9e099899-aaf7-4bd8-b37e-8616b5060e33\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qr5xz" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.453361 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3e0e52dd-503e-47a4-b68e-3f768ccecfc6-audit-policies\") pod \"apiserver-7bbb656c7d-g692d\" (UID: \"3e0e52dd-503e-47a4-b68e-3f768ccecfc6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g692d" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.453393 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/247d17b7-7064-4ef1-9aed-c205fec84bc2-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-2zq2n\" (UID: \"247d17b7-7064-4ef1-9aed-c205fec84bc2\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2zq2n" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.453411 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wgmgc\" (UniqueName: \"kubernetes.io/projected/e87b6d85-2d13-4b72-bb6b-8299b6c72d0a-kube-api-access-wgmgc\") pod \"kube-storage-version-migrator-operator-b67b599dd-gndwh\" (UID: \"e87b6d85-2d13-4b72-bb6b-8299b6c72d0a\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gndwh" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.453425 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bb4dde0f-222d-4609-b1c9-fed653668e3a-trusted-ca\") pod \"console-operator-58897d9998-lfxkq\" (UID: \"bb4dde0f-222d-4609-b1c9-fed653668e3a\") " pod="openshift-console-operator/console-operator-58897d9998-lfxkq" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.453440 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ad510a18-cd72-4cf0-9b0c-dd3540d491fa-service-ca\") pod \"console-f9d7485db-k6kls\" (UID: \"ad510a18-cd72-4cf0-9b0c-dd3540d491fa\") " pod="openshift-console/console-f9d7485db-k6kls" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.453476 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/8c854339-76bb-439d-b267-b2617cdd3f41-etcd-client\") pod \"etcd-operator-b45778765-r89c2\" (UID: \"8c854339-76bb-439d-b267-b2617cdd3f41\") " pod="openshift-etcd-operator/etcd-operator-b45778765-r89c2" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.453491 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/d5fb9349-891c-447a-bb5a-5e18a0ce9f92-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-lg9s6\" (UID: \"d5fb9349-891c-447a-bb5a-5e18a0ce9f92\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lg9s6" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.453509 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/1bdf5eb2-fe05-4e24-9e00-43c72afc3583-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-l9d4f\" (UID: \"1bdf5eb2-fe05-4e24-9e00-43c72afc3583\") " pod="openshift-authentication/oauth-openshift-558db77b4-l9d4f" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.453562 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4d4xx\" (UniqueName: \"kubernetes.io/projected/8c854339-76bb-439d-b267-b2617cdd3f41-kube-api-access-4d4xx\") pod \"etcd-operator-b45778765-r89c2\" (UID: \"8c854339-76bb-439d-b267-b2617cdd3f41\") " pod="openshift-etcd-operator/etcd-operator-b45778765-r89c2" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.453576 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-hr4qs" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.453577 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/247d17b7-7064-4ef1-9aed-c205fec84bc2-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-2zq2n\" (UID: \"247d17b7-7064-4ef1-9aed-c205fec84bc2\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2zq2n" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.453600 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d5fb9349-891c-447a-bb5a-5e18a0ce9f92-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-lg9s6\" (UID: \"d5fb9349-891c-447a-bb5a-5e18a0ce9f92\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lg9s6" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.453633 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/1bdf5eb2-fe05-4e24-9e00-43c72afc3583-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-l9d4f\" (UID: \"1bdf5eb2-fe05-4e24-9e00-43c72afc3583\") " pod="openshift-authentication/oauth-openshift-558db77b4-l9d4f" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.453648 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9e099899-aaf7-4bd8-b37e-8616b5060e33-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-qr5xz\" (UID: \"9e099899-aaf7-4bd8-b37e-8616b5060e33\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qr5xz" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.453663 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3e0e52dd-503e-47a4-b68e-3f768ccecfc6-audit-dir\") pod \"apiserver-7bbb656c7d-g692d\" (UID: \"3e0e52dd-503e-47a4-b68e-3f768ccecfc6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g692d" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.453702 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/8650bbf0-78d2-410e-a62d-d4adf673cb55-metrics-tls\") pod \"dns-operator-744455d44c-kkw8v\" (UID: \"8650bbf0-78d2-410e-a62d-d4adf673cb55\") " pod="openshift-dns-operator/dns-operator-744455d44c-kkw8v" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.453720 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/75da7b9a-7085-492b-954a-6bb920fbb186-audit-dir\") pod \"apiserver-76f77b778f-l88b4\" (UID: \"75da7b9a-7085-492b-954a-6bb920fbb186\") " pod="openshift-apiserver/apiserver-76f77b778f-l88b4" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.453736 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/1bdf5eb2-fe05-4e24-9e00-43c72afc3583-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-l9d4f\" (UID: \"1bdf5eb2-fe05-4e24-9e00-43c72afc3583\") " pod="openshift-authentication/oauth-openshift-558db77b4-l9d4f" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.453754 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/1bdf5eb2-fe05-4e24-9e00-43c72afc3583-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-l9d4f\" (UID: \"1bdf5eb2-fe05-4e24-9e00-43c72afc3583\") " pod="openshift-authentication/oauth-openshift-558db77b4-l9d4f" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.453789 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fa007484-938d-460e-9c5e-c3f5b31a25d2-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ks7hm\" (UID: \"fa007484-938d-460e-9c5e-c3f5b31a25d2\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ks7hm" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.453805 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3404cb6f-6c3a-4e0a-99a4-27530beefa83-auth-proxy-config\") pod \"machine-approver-56656f9798-vtdlb\" (UID: \"3404cb6f-6c3a-4e0a-99a4-27530beefa83\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vtdlb" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.453833 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-874kt\" (UniqueName: \"kubernetes.io/projected/3404cb6f-6c3a-4e0a-99a4-27530beefa83-kube-api-access-874kt\") pod \"machine-approver-56656f9798-vtdlb\" (UID: \"3404cb6f-6c3a-4e0a-99a4-27530beefa83\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vtdlb" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.453865 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/8c854339-76bb-439d-b267-b2617cdd3f41-etcd-service-ca\") pod \"etcd-operator-b45778765-r89c2\" (UID: \"8c854339-76bb-439d-b267-b2617cdd3f41\") " pod="openshift-etcd-operator/etcd-operator-b45778765-r89c2" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.453882 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/75da7b9a-7085-492b-954a-6bb920fbb186-config\") pod \"apiserver-76f77b778f-l88b4\" (UID: \"75da7b9a-7085-492b-954a-6bb920fbb186\") " pod="openshift-apiserver/apiserver-76f77b778f-l88b4" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.453896 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/75da7b9a-7085-492b-954a-6bb920fbb186-etcd-client\") pod \"apiserver-76f77b778f-l88b4\" (UID: \"75da7b9a-7085-492b-954a-6bb920fbb186\") " pod="openshift-apiserver/apiserver-76f77b778f-l88b4" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.453912 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bdf5eb2-fe05-4e24-9e00-43c72afc3583-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-l9d4f\" (UID: \"1bdf5eb2-fe05-4e24-9e00-43c72afc3583\") " pod="openshift-authentication/oauth-openshift-558db77b4-l9d4f" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.453944 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-htnkn\" (UniqueName: \"kubernetes.io/projected/29e1953f-668a-4393-82ab-a45fa4b79ec7-kube-api-access-htnkn\") pod \"openshift-apiserver-operator-796bbdcf4f-2vcc6\" (UID: \"29e1953f-668a-4393-82ab-a45fa4b79ec7\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2vcc6" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.453959 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3e0e52dd-503e-47a4-b68e-3f768ccecfc6-serving-cert\") pod \"apiserver-7bbb656c7d-g692d\" (UID: \"3e0e52dd-503e-47a4-b68e-3f768ccecfc6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g692d" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.453975 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa007484-938d-460e-9c5e-c3f5b31a25d2-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ks7hm\" (UID: \"fa007484-938d-460e-9c5e-c3f5b31a25d2\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ks7hm" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.453990 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/75da7b9a-7085-492b-954a-6bb920fbb186-etcd-serving-ca\") pod \"apiserver-76f77b778f-l88b4\" (UID: \"75da7b9a-7085-492b-954a-6bb920fbb186\") " pod="openshift-apiserver/apiserver-76f77b778f-l88b4" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.454025 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/75da7b9a-7085-492b-954a-6bb920fbb186-trusted-ca-bundle\") pod \"apiserver-76f77b778f-l88b4\" (UID: \"75da7b9a-7085-492b-954a-6bb920fbb186\") " pod="openshift-apiserver/apiserver-76f77b778f-l88b4" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.454040 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/8e2095d8-25c6-407f-aadb-f4603ae44d15-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-krqhc\" (UID: \"8e2095d8-25c6-407f-aadb-f4603ae44d15\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-krqhc" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.454057 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ad510a18-cd72-4cf0-9b0c-dd3540d491fa-oauth-serving-cert\") pod \"console-f9d7485db-k6kls\" (UID: \"ad510a18-cd72-4cf0-9b0c-dd3540d491fa\") " pod="openshift-console/console-f9d7485db-k6kls" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.454073 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rdj7r\" (UniqueName: \"kubernetes.io/projected/f185e5b2-54c0-44ef-bdda-118efa89cb5f-kube-api-access-rdj7r\") pod \"downloads-7954f5f757-2ng4n\" (UID: \"f185e5b2-54c0-44ef-bdda-118efa89cb5f\") " pod="openshift-console/downloads-7954f5f757-2ng4n" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.454657 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f257970f-c1ce-44fb-b207-a78126794cfe-service-ca-bundle\") pod \"authentication-operator-69f744f599-tfwnz\" (UID: \"f257970f-c1ce-44fb-b207-a78126794cfe\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-tfwnz" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.455547 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-6b6br"] Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.455769 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/1bdf5eb2-fe05-4e24-9e00-43c72afc3583-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-l9d4f\" (UID: \"1bdf5eb2-fe05-4e24-9e00-43c72afc3583\") " pod="openshift-authentication/oauth-openshift-558db77b4-l9d4f" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.455901 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3e0e52dd-503e-47a4-b68e-3f768ccecfc6-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-g692d\" (UID: \"3e0e52dd-503e-47a4-b68e-3f768ccecfc6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g692d" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.454103 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n28fr\" (UniqueName: \"kubernetes.io/projected/1bdf5eb2-fe05-4e24-9e00-43c72afc3583-kube-api-access-n28fr\") pod \"oauth-openshift-558db77b4-l9d4f\" (UID: \"1bdf5eb2-fe05-4e24-9e00-43c72afc3583\") " pod="openshift-authentication/oauth-openshift-558db77b4-l9d4f" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.456070 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/3404cb6f-6c3a-4e0a-99a4-27530beefa83-machine-approver-tls\") pod \"machine-approver-56656f9798-vtdlb\" (UID: \"3404cb6f-6c3a-4e0a-99a4-27530beefa83\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vtdlb" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.456089 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-grgzk\" (UniqueName: \"kubernetes.io/projected/f257970f-c1ce-44fb-b207-a78126794cfe-kube-api-access-grgzk\") pod \"authentication-operator-69f744f599-tfwnz\" (UID: \"f257970f-c1ce-44fb-b207-a78126794cfe\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-tfwnz" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.456169 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e099899-aaf7-4bd8-b37e-8616b5060e33-config\") pod \"controller-manager-879f6c89f-qr5xz\" (UID: \"9e099899-aaf7-4bd8-b37e-8616b5060e33\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qr5xz" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.456188 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/fb52538d-050a-4b69-91b2-aa48f9833fe7-bound-sa-token\") pod \"ingress-operator-5b745b69d9-hxqw8\" (UID: \"fb52538d-050a-4b69-91b2-aa48f9833fe7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hxqw8" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.456248 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5407364a-e2b6-43b4-9eb0-deb12a022289-client-ca\") pod \"route-controller-manager-6576b87f9c-msljv\" (UID: \"5407364a-e2b6-43b4-9eb0-deb12a022289\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-msljv" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.456265 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/75da7b9a-7085-492b-954a-6bb920fbb186-audit\") pod \"apiserver-76f77b778f-l88b4\" (UID: \"75da7b9a-7085-492b-954a-6bb920fbb186\") " pod="openshift-apiserver/apiserver-76f77b778f-l88b4" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.456282 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-shkmj\" (UniqueName: \"kubernetes.io/projected/8650bbf0-78d2-410e-a62d-d4adf673cb55-kube-api-access-shkmj\") pod \"dns-operator-744455d44c-kkw8v\" (UID: \"8650bbf0-78d2-410e-a62d-d4adf673cb55\") " pod="openshift-dns-operator/dns-operator-744455d44c-kkw8v" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.456323 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/7704cded-5ba1-4474-8a31-c0ba947b7679-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-d89dq\" (UID: \"7704cded-5ba1-4474-8a31-c0ba947b7679\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-d89dq" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.456344 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.456385 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/1bdf5eb2-fe05-4e24-9e00-43c72afc3583-audit-policies\") pod \"oauth-openshift-558db77b4-l9d4f\" (UID: \"1bdf5eb2-fe05-4e24-9e00-43c72afc3583\") " pod="openshift-authentication/oauth-openshift-558db77b4-l9d4f" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.456403 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/fb52538d-050a-4b69-91b2-aa48f9833fe7-metrics-tls\") pod \"ingress-operator-5b745b69d9-hxqw8\" (UID: \"fb52538d-050a-4b69-91b2-aa48f9833fe7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hxqw8" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.456425 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8c854339-76bb-439d-b267-b2617cdd3f41-serving-cert\") pod \"etcd-operator-b45778765-r89c2\" (UID: \"8c854339-76bb-439d-b267-b2617cdd3f41\") " pod="openshift-etcd-operator/etcd-operator-b45778765-r89c2" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.456441 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7704cded-5ba1-4474-8a31-c0ba947b7679-config\") pod \"machine-api-operator-5694c8668f-d89dq\" (UID: \"7704cded-5ba1-4474-8a31-c0ba947b7679\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-d89dq" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.456485 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dvkb4\" (UniqueName: \"kubernetes.io/projected/bb4dde0f-222d-4609-b1c9-fed653668e3a-kube-api-access-dvkb4\") pod \"console-operator-58897d9998-lfxkq\" (UID: \"bb4dde0f-222d-4609-b1c9-fed653668e3a\") " pod="openshift-console-operator/console-operator-58897d9998-lfxkq" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.456503 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3404cb6f-6c3a-4e0a-99a4-27530beefa83-config\") pod \"machine-approver-56656f9798-vtdlb\" (UID: \"3404cb6f-6c3a-4e0a-99a4-27530beefa83\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vtdlb" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.456530 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/1bdf5eb2-fe05-4e24-9e00-43c72afc3583-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-l9d4f\" (UID: \"1bdf5eb2-fe05-4e24-9e00-43c72afc3583\") " pod="openshift-authentication/oauth-openshift-558db77b4-l9d4f" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.456149 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-6b6br" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.457022 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3404cb6f-6c3a-4e0a-99a4-27530beefa83-config\") pod \"machine-approver-56656f9798-vtdlb\" (UID: \"3404cb6f-6c3a-4e0a-99a4-27530beefa83\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vtdlb" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.457041 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3e0e52dd-503e-47a4-b68e-3f768ccecfc6-audit-dir\") pod \"apiserver-7bbb656c7d-g692d\" (UID: \"3e0e52dd-503e-47a4-b68e-3f768ccecfc6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g692d" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.457254 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9e099899-aaf7-4bd8-b37e-8616b5060e33-client-ca\") pod \"controller-manager-879f6c89f-qr5xz\" (UID: \"9e099899-aaf7-4bd8-b37e-8616b5060e33\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qr5xz" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.457288 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1bdf5eb2-fe05-4e24-9e00-43c72afc3583-audit-dir\") pod \"oauth-openshift-558db77b4-l9d4f\" (UID: \"1bdf5eb2-fe05-4e24-9e00-43c72afc3583\") " pod="openshift-authentication/oauth-openshift-558db77b4-l9d4f" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.457406 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/3e0e52dd-503e-47a4-b68e-3f768ccecfc6-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-g692d\" (UID: \"3e0e52dd-503e-47a4-b68e-3f768ccecfc6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g692d" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.457419 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f257970f-c1ce-44fb-b207-a78126794cfe-serving-cert\") pod \"authentication-operator-69f744f599-tfwnz\" (UID: \"f257970f-c1ce-44fb-b207-a78126794cfe\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-tfwnz" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.457451 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6587n\" (UniqueName: \"kubernetes.io/projected/3e0e52dd-503e-47a4-b68e-3f768ccecfc6-kube-api-access-6587n\") pod \"apiserver-7bbb656c7d-g692d\" (UID: \"3e0e52dd-503e-47a4-b68e-3f768ccecfc6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g692d" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.457494 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8c854339-76bb-439d-b267-b2617cdd3f41-config\") pod \"etcd-operator-b45778765-r89c2\" (UID: \"8c854339-76bb-439d-b267-b2617cdd3f41\") " pod="openshift-etcd-operator/etcd-operator-b45778765-r89c2" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.457542 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f257970f-c1ce-44fb-b207-a78126794cfe-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-tfwnz\" (UID: \"f257970f-c1ce-44fb-b207-a78126794cfe\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-tfwnz" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.457572 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ad510a18-cd72-4cf0-9b0c-dd3540d491fa-console-serving-cert\") pod \"console-f9d7485db-k6kls\" (UID: \"ad510a18-cd72-4cf0-9b0c-dd3540d491fa\") " pod="openshift-console/console-f9d7485db-k6kls" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.457595 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ad510a18-cd72-4cf0-9b0c-dd3540d491fa-console-oauth-config\") pod \"console-f9d7485db-k6kls\" (UID: \"ad510a18-cd72-4cf0-9b0c-dd3540d491fa\") " pod="openshift-console/console-f9d7485db-k6kls" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.457625 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ad510a18-cd72-4cf0-9b0c-dd3540d491fa-console-config\") pod \"console-f9d7485db-k6kls\" (UID: \"ad510a18-cd72-4cf0-9b0c-dd3540d491fa\") " pod="openshift-console/console-f9d7485db-k6kls" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.457653 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/7704cded-5ba1-4474-8a31-c0ba947b7679-images\") pod \"machine-api-operator-5694c8668f-d89dq\" (UID: \"7704cded-5ba1-4474-8a31-c0ba947b7679\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-d89dq" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.457677 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d825v\" (UniqueName: \"kubernetes.io/projected/75da7b9a-7085-492b-954a-6bb920fbb186-kube-api-access-d825v\") pod \"apiserver-76f77b778f-l88b4\" (UID: \"75da7b9a-7085-492b-954a-6bb920fbb186\") " pod="openshift-apiserver/apiserver-76f77b778f-l88b4" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.457701 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rm2hc\" (UniqueName: \"kubernetes.io/projected/ad510a18-cd72-4cf0-9b0c-dd3540d491fa-kube-api-access-rm2hc\") pod \"console-f9d7485db-k6kls\" (UID: \"ad510a18-cd72-4cf0-9b0c-dd3540d491fa\") " pod="openshift-console/console-f9d7485db-k6kls" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.457736 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bb4dde0f-222d-4609-b1c9-fed653668e3a-serving-cert\") pod \"console-operator-58897d9998-lfxkq\" (UID: \"bb4dde0f-222d-4609-b1c9-fed653668e3a\") " pod="openshift-console-operator/console-operator-58897d9998-lfxkq" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.458081 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/29e1953f-668a-4393-82ab-a45fa4b79ec7-config\") pod \"openshift-apiserver-operator-796bbdcf4f-2vcc6\" (UID: \"29e1953f-668a-4393-82ab-a45fa4b79ec7\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2vcc6" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.458088 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f257970f-c1ce-44fb-b207-a78126794cfe-config\") pod \"authentication-operator-69f744f599-tfwnz\" (UID: \"f257970f-c1ce-44fb-b207-a78126794cfe\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-tfwnz" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.458694 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3e0e52dd-503e-47a4-b68e-3f768ccecfc6-audit-policies\") pod \"apiserver-7bbb656c7d-g692d\" (UID: \"3e0e52dd-503e-47a4-b68e-3f768ccecfc6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g692d" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.458732 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e099899-aaf7-4bd8-b37e-8616b5060e33-config\") pod \"controller-manager-879f6c89f-qr5xz\" (UID: \"9e099899-aaf7-4bd8-b37e-8616b5060e33\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qr5xz" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.459000 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/1bdf5eb2-fe05-4e24-9e00-43c72afc3583-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-l9d4f\" (UID: \"1bdf5eb2-fe05-4e24-9e00-43c72afc3583\") " pod="openshift-authentication/oauth-openshift-558db77b4-l9d4f" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.459532 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/8c854339-76bb-439d-b267-b2617cdd3f41-etcd-ca\") pod \"etcd-operator-b45778765-r89c2\" (UID: \"8c854339-76bb-439d-b267-b2617cdd3f41\") " pod="openshift-etcd-operator/etcd-operator-b45778765-r89c2" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.459583 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5407364a-e2b6-43b4-9eb0-deb12a022289-config\") pod \"route-controller-manager-6576b87f9c-msljv\" (UID: \"5407364a-e2b6-43b4-9eb0-deb12a022289\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-msljv" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.459609 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e87b6d85-2d13-4b72-bb6b-8299b6c72d0a-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-gndwh\" (UID: \"e87b6d85-2d13-4b72-bb6b-8299b6c72d0a\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gndwh" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.459640 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/75da7b9a-7085-492b-954a-6bb920fbb186-node-pullsecrets\") pod \"apiserver-76f77b778f-l88b4\" (UID: \"75da7b9a-7085-492b-954a-6bb920fbb186\") " pod="openshift-apiserver/apiserver-76f77b778f-l88b4" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.459663 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/75da7b9a-7085-492b-954a-6bb920fbb186-encryption-config\") pod \"apiserver-76f77b778f-l88b4\" (UID: \"75da7b9a-7085-492b-954a-6bb920fbb186\") " pod="openshift-apiserver/apiserver-76f77b778f-l88b4" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.460508 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5407364a-e2b6-43b4-9eb0-deb12a022289-serving-cert\") pod \"route-controller-manager-6576b87f9c-msljv\" (UID: \"5407364a-e2b6-43b4-9eb0-deb12a022289\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-msljv" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.461990 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3404cb6f-6c3a-4e0a-99a4-27530beefa83-auth-proxy-config\") pod \"machine-approver-56656f9798-vtdlb\" (UID: \"3404cb6f-6c3a-4e0a-99a4-27530beefa83\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vtdlb" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.462948 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/1bdf5eb2-fe05-4e24-9e00-43c72afc3583-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-l9d4f\" (UID: \"1bdf5eb2-fe05-4e24-9e00-43c72afc3583\") " pod="openshift-authentication/oauth-openshift-558db77b4-l9d4f" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.463440 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/1bdf5eb2-fe05-4e24-9e00-43c72afc3583-audit-policies\") pod \"oauth-openshift-558db77b4-l9d4f\" (UID: \"1bdf5eb2-fe05-4e24-9e00-43c72afc3583\") " pod="openshift-authentication/oauth-openshift-558db77b4-l9d4f" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.460978 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5407364a-e2b6-43b4-9eb0-deb12a022289-client-ca\") pod \"route-controller-manager-6576b87f9c-msljv\" (UID: \"5407364a-e2b6-43b4-9eb0-deb12a022289\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-msljv" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.464181 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/1bdf5eb2-fe05-4e24-9e00-43c72afc3583-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-l9d4f\" (UID: \"1bdf5eb2-fe05-4e24-9e00-43c72afc3583\") " pod="openshift-authentication/oauth-openshift-558db77b4-l9d4f" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.464676 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/1bdf5eb2-fe05-4e24-9e00-43c72afc3583-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-l9d4f\" (UID: \"1bdf5eb2-fe05-4e24-9e00-43c72afc3583\") " pod="openshift-authentication/oauth-openshift-558db77b4-l9d4f" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.464931 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3e0e52dd-503e-47a4-b68e-3f768ccecfc6-serving-cert\") pod \"apiserver-7bbb656c7d-g692d\" (UID: \"3e0e52dd-503e-47a4-b68e-3f768ccecfc6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g692d" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.465208 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/3404cb6f-6c3a-4e0a-99a4-27530beefa83-machine-approver-tls\") pod \"machine-approver-56656f9798-vtdlb\" (UID: \"3404cb6f-6c3a-4e0a-99a4-27530beefa83\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vtdlb" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.466919 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f257970f-c1ce-44fb-b207-a78126794cfe-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-tfwnz\" (UID: \"f257970f-c1ce-44fb-b207-a78126794cfe\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-tfwnz" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.467078 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bdf5eb2-fe05-4e24-9e00-43c72afc3583-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-l9d4f\" (UID: \"1bdf5eb2-fe05-4e24-9e00-43c72afc3583\") " pod="openshift-authentication/oauth-openshift-558db77b4-l9d4f" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.467172 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5407364a-e2b6-43b4-9eb0-deb12a022289-config\") pod \"route-controller-manager-6576b87f9c-msljv\" (UID: \"5407364a-e2b6-43b4-9eb0-deb12a022289\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-msljv" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.467524 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f257970f-c1ce-44fb-b207-a78126794cfe-serving-cert\") pod \"authentication-operator-69f744f599-tfwnz\" (UID: \"f257970f-c1ce-44fb-b207-a78126794cfe\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-tfwnz" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.467559 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qqg25"] Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.468359 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/1bdf5eb2-fe05-4e24-9e00-43c72afc3583-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-l9d4f\" (UID: \"1bdf5eb2-fe05-4e24-9e00-43c72afc3583\") " pod="openshift-authentication/oauth-openshift-558db77b4-l9d4f" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.468513 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/3e0e52dd-503e-47a4-b68e-3f768ccecfc6-encryption-config\") pod \"apiserver-7bbb656c7d-g692d\" (UID: \"3e0e52dd-503e-47a4-b68e-3f768ccecfc6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g692d" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.468520 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9e099899-aaf7-4bd8-b37e-8616b5060e33-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-qr5xz\" (UID: \"9e099899-aaf7-4bd8-b37e-8616b5060e33\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qr5xz" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.468636 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/1bdf5eb2-fe05-4e24-9e00-43c72afc3583-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-l9d4f\" (UID: \"1bdf5eb2-fe05-4e24-9e00-43c72afc3583\") " pod="openshift-authentication/oauth-openshift-558db77b4-l9d4f" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.469026 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/1bdf5eb2-fe05-4e24-9e00-43c72afc3583-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-l9d4f\" (UID: \"1bdf5eb2-fe05-4e24-9e00-43c72afc3583\") " pod="openshift-authentication/oauth-openshift-558db77b4-l9d4f" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.469322 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9e099899-aaf7-4bd8-b37e-8616b5060e33-serving-cert\") pod \"controller-manager-879f6c89f-qr5xz\" (UID: \"9e099899-aaf7-4bd8-b37e-8616b5060e33\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qr5xz" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.469351 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/1bdf5eb2-fe05-4e24-9e00-43c72afc3583-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-l9d4f\" (UID: \"1bdf5eb2-fe05-4e24-9e00-43c72afc3583\") " pod="openshift-authentication/oauth-openshift-558db77b4-l9d4f" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.469407 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/3e0e52dd-503e-47a4-b68e-3f768ccecfc6-etcd-client\") pod \"apiserver-7bbb656c7d-g692d\" (UID: \"3e0e52dd-503e-47a4-b68e-3f768ccecfc6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g692d" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.473406 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-zvcj8"] Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.475535 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.476779 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kdzkc"] Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.480621 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-8fksd"] Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.484753 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-66mv8"] Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.485062 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-lpgf5"] Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.485973 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/29e1953f-668a-4393-82ab-a45fa4b79ec7-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-2vcc6\" (UID: \"29e1953f-668a-4393-82ab-a45fa4b79ec7\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2vcc6" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.486164 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-jlqk5"] Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.488643 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p8x2v"] Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.490938 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-qp2jl"] Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.492426 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-hr4qs"] Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.493500 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-qtll9"] Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.494778 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-qtll9" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.495265 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-qtll9"] Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.495375 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.516948 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.535381 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.555831 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.560343 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d5fb9349-891c-447a-bb5a-5e18a0ce9f92-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-lg9s6\" (UID: \"d5fb9349-891c-447a-bb5a-5e18a0ce9f92\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lg9s6" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.560391 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7js74\" (UniqueName: \"kubernetes.io/projected/d5fb9349-891c-447a-bb5a-5e18a0ce9f92-kube-api-access-7js74\") pod \"cluster-image-registry-operator-dc59b4c8b-lg9s6\" (UID: \"d5fb9349-891c-447a-bb5a-5e18a0ce9f92\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lg9s6" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.560417 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/247d17b7-7064-4ef1-9aed-c205fec84bc2-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-2zq2n\" (UID: \"247d17b7-7064-4ef1-9aed-c205fec84bc2\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2zq2n" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.560458 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bb4dde0f-222d-4609-b1c9-fed653668e3a-trusted-ca\") pod \"console-operator-58897d9998-lfxkq\" (UID: \"bb4dde0f-222d-4609-b1c9-fed653668e3a\") " pod="openshift-console-operator/console-operator-58897d9998-lfxkq" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.560475 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wgmgc\" (UniqueName: \"kubernetes.io/projected/e87b6d85-2d13-4b72-bb6b-8299b6c72d0a-kube-api-access-wgmgc\") pod \"kube-storage-version-migrator-operator-b67b599dd-gndwh\" (UID: \"e87b6d85-2d13-4b72-bb6b-8299b6c72d0a\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gndwh" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.560491 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/d5fb9349-891c-447a-bb5a-5e18a0ce9f92-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-lg9s6\" (UID: \"d5fb9349-891c-447a-bb5a-5e18a0ce9f92\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lg9s6" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.560507 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ad510a18-cd72-4cf0-9b0c-dd3540d491fa-service-ca\") pod \"console-f9d7485db-k6kls\" (UID: \"ad510a18-cd72-4cf0-9b0c-dd3540d491fa\") " pod="openshift-console/console-f9d7485db-k6kls" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.560539 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/8c854339-76bb-439d-b267-b2617cdd3f41-etcd-client\") pod \"etcd-operator-b45778765-r89c2\" (UID: \"8c854339-76bb-439d-b267-b2617cdd3f41\") " pod="openshift-etcd-operator/etcd-operator-b45778765-r89c2" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.561500 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d5fb9349-891c-447a-bb5a-5e18a0ce9f92-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-lg9s6\" (UID: \"d5fb9349-891c-447a-bb5a-5e18a0ce9f92\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lg9s6" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.561519 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4d4xx\" (UniqueName: \"kubernetes.io/projected/8c854339-76bb-439d-b267-b2617cdd3f41-kube-api-access-4d4xx\") pod \"etcd-operator-b45778765-r89c2\" (UID: \"8c854339-76bb-439d-b267-b2617cdd3f41\") " pod="openshift-etcd-operator/etcd-operator-b45778765-r89c2" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.562521 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/247d17b7-7064-4ef1-9aed-c205fec84bc2-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-2zq2n\" (UID: \"247d17b7-7064-4ef1-9aed-c205fec84bc2\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2zq2n" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.562545 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/8650bbf0-78d2-410e-a62d-d4adf673cb55-metrics-tls\") pod \"dns-operator-744455d44c-kkw8v\" (UID: \"8650bbf0-78d2-410e-a62d-d4adf673cb55\") " pod="openshift-dns-operator/dns-operator-744455d44c-kkw8v" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.562383 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d5fb9349-891c-447a-bb5a-5e18a0ce9f92-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-lg9s6\" (UID: \"d5fb9349-891c-447a-bb5a-5e18a0ce9f92\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lg9s6" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.561423 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ad510a18-cd72-4cf0-9b0c-dd3540d491fa-service-ca\") pod \"console-f9d7485db-k6kls\" (UID: \"ad510a18-cd72-4cf0-9b0c-dd3540d491fa\") " pod="openshift-console/console-f9d7485db-k6kls" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.562677 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/75da7b9a-7085-492b-954a-6bb920fbb186-audit-dir\") pod \"apiserver-76f77b778f-l88b4\" (UID: \"75da7b9a-7085-492b-954a-6bb920fbb186\") " pod="openshift-apiserver/apiserver-76f77b778f-l88b4" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.562697 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fa007484-938d-460e-9c5e-c3f5b31a25d2-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ks7hm\" (UID: \"fa007484-938d-460e-9c5e-c3f5b31a25d2\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ks7hm" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.562734 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/8c854339-76bb-439d-b267-b2617cdd3f41-etcd-service-ca\") pod \"etcd-operator-b45778765-r89c2\" (UID: \"8c854339-76bb-439d-b267-b2617cdd3f41\") " pod="openshift-etcd-operator/etcd-operator-b45778765-r89c2" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.562751 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/75da7b9a-7085-492b-954a-6bb920fbb186-config\") pod \"apiserver-76f77b778f-l88b4\" (UID: \"75da7b9a-7085-492b-954a-6bb920fbb186\") " pod="openshift-apiserver/apiserver-76f77b778f-l88b4" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.562844 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/75da7b9a-7085-492b-954a-6bb920fbb186-audit-dir\") pod \"apiserver-76f77b778f-l88b4\" (UID: \"75da7b9a-7085-492b-954a-6bb920fbb186\") " pod="openshift-apiserver/apiserver-76f77b778f-l88b4" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.563018 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bb4dde0f-222d-4609-b1c9-fed653668e3a-trusted-ca\") pod \"console-operator-58897d9998-lfxkq\" (UID: \"bb4dde0f-222d-4609-b1c9-fed653668e3a\") " pod="openshift-console-operator/console-operator-58897d9998-lfxkq" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.563371 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/8c854339-76bb-439d-b267-b2617cdd3f41-etcd-service-ca\") pod \"etcd-operator-b45778765-r89c2\" (UID: \"8c854339-76bb-439d-b267-b2617cdd3f41\") " pod="openshift-etcd-operator/etcd-operator-b45778765-r89c2" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.563506 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/75da7b9a-7085-492b-954a-6bb920fbb186-config\") pod \"apiserver-76f77b778f-l88b4\" (UID: \"75da7b9a-7085-492b-954a-6bb920fbb186\") " pod="openshift-apiserver/apiserver-76f77b778f-l88b4" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.562766 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/75da7b9a-7085-492b-954a-6bb920fbb186-etcd-client\") pod \"apiserver-76f77b778f-l88b4\" (UID: \"75da7b9a-7085-492b-954a-6bb920fbb186\") " pod="openshift-apiserver/apiserver-76f77b778f-l88b4" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.564037 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/75da7b9a-7085-492b-954a-6bb920fbb186-trusted-ca-bundle\") pod \"apiserver-76f77b778f-l88b4\" (UID: \"75da7b9a-7085-492b-954a-6bb920fbb186\") " pod="openshift-apiserver/apiserver-76f77b778f-l88b4" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.564058 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/8e2095d8-25c6-407f-aadb-f4603ae44d15-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-krqhc\" (UID: \"8e2095d8-25c6-407f-aadb-f4603ae44d15\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-krqhc" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.564089 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ad510a18-cd72-4cf0-9b0c-dd3540d491fa-oauth-serving-cert\") pod \"console-f9d7485db-k6kls\" (UID: \"ad510a18-cd72-4cf0-9b0c-dd3540d491fa\") " pod="openshift-console/console-f9d7485db-k6kls" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.564282 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/8c854339-76bb-439d-b267-b2617cdd3f41-etcd-client\") pod \"etcd-operator-b45778765-r89c2\" (UID: \"8c854339-76bb-439d-b267-b2617cdd3f41\") " pod="openshift-etcd-operator/etcd-operator-b45778765-r89c2" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.564475 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdj7r\" (UniqueName: \"kubernetes.io/projected/f185e5b2-54c0-44ef-bdda-118efa89cb5f-kube-api-access-rdj7r\") pod \"downloads-7954f5f757-2ng4n\" (UID: \"f185e5b2-54c0-44ef-bdda-118efa89cb5f\") " pod="openshift-console/downloads-7954f5f757-2ng4n" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.564503 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa007484-938d-460e-9c5e-c3f5b31a25d2-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ks7hm\" (UID: \"fa007484-938d-460e-9c5e-c3f5b31a25d2\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ks7hm" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.564694 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/75da7b9a-7085-492b-954a-6bb920fbb186-etcd-serving-ca\") pod \"apiserver-76f77b778f-l88b4\" (UID: \"75da7b9a-7085-492b-954a-6bb920fbb186\") " pod="openshift-apiserver/apiserver-76f77b778f-l88b4" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.564745 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/fb52538d-050a-4b69-91b2-aa48f9833fe7-bound-sa-token\") pod \"ingress-operator-5b745b69d9-hxqw8\" (UID: \"fb52538d-050a-4b69-91b2-aa48f9833fe7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hxqw8" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.564760 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/75da7b9a-7085-492b-954a-6bb920fbb186-trusted-ca-bundle\") pod \"apiserver-76f77b778f-l88b4\" (UID: \"75da7b9a-7085-492b-954a-6bb920fbb186\") " pod="openshift-apiserver/apiserver-76f77b778f-l88b4" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.564779 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/75da7b9a-7085-492b-954a-6bb920fbb186-audit\") pod \"apiserver-76f77b778f-l88b4\" (UID: \"75da7b9a-7085-492b-954a-6bb920fbb186\") " pod="openshift-apiserver/apiserver-76f77b778f-l88b4" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.564919 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/fb52538d-050a-4b69-91b2-aa48f9833fe7-metrics-tls\") pod \"ingress-operator-5b745b69d9-hxqw8\" (UID: \"fb52538d-050a-4b69-91b2-aa48f9833fe7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hxqw8" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.565031 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-shkmj\" (UniqueName: \"kubernetes.io/projected/8650bbf0-78d2-410e-a62d-d4adf673cb55-kube-api-access-shkmj\") pod \"dns-operator-744455d44c-kkw8v\" (UID: \"8650bbf0-78d2-410e-a62d-d4adf673cb55\") " pod="openshift-dns-operator/dns-operator-744455d44c-kkw8v" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.565053 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/7704cded-5ba1-4474-8a31-c0ba947b7679-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-d89dq\" (UID: \"7704cded-5ba1-4474-8a31-c0ba947b7679\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-d89dq" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.565071 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dvkb4\" (UniqueName: \"kubernetes.io/projected/bb4dde0f-222d-4609-b1c9-fed653668e3a-kube-api-access-dvkb4\") pod \"console-operator-58897d9998-lfxkq\" (UID: \"bb4dde0f-222d-4609-b1c9-fed653668e3a\") " pod="openshift-console-operator/console-operator-58897d9998-lfxkq" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.565104 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8c854339-76bb-439d-b267-b2617cdd3f41-serving-cert\") pod \"etcd-operator-b45778765-r89c2\" (UID: \"8c854339-76bb-439d-b267-b2617cdd3f41\") " pod="openshift-etcd-operator/etcd-operator-b45778765-r89c2" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.565135 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7704cded-5ba1-4474-8a31-c0ba947b7679-config\") pod \"machine-api-operator-5694c8668f-d89dq\" (UID: \"7704cded-5ba1-4474-8a31-c0ba947b7679\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-d89dq" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.565156 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8c854339-76bb-439d-b267-b2617cdd3f41-config\") pod \"etcd-operator-b45778765-r89c2\" (UID: \"8c854339-76bb-439d-b267-b2617cdd3f41\") " pod="openshift-etcd-operator/etcd-operator-b45778765-r89c2" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.565198 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ad510a18-cd72-4cf0-9b0c-dd3540d491fa-console-serving-cert\") pod \"console-f9d7485db-k6kls\" (UID: \"ad510a18-cd72-4cf0-9b0c-dd3540d491fa\") " pod="openshift-console/console-f9d7485db-k6kls" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.565213 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ad510a18-cd72-4cf0-9b0c-dd3540d491fa-console-oauth-config\") pod \"console-f9d7485db-k6kls\" (UID: \"ad510a18-cd72-4cf0-9b0c-dd3540d491fa\") " pod="openshift-console/console-f9d7485db-k6kls" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.565222 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/75da7b9a-7085-492b-954a-6bb920fbb186-etcd-serving-ca\") pod \"apiserver-76f77b778f-l88b4\" (UID: \"75da7b9a-7085-492b-954a-6bb920fbb186\") " pod="openshift-apiserver/apiserver-76f77b778f-l88b4" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.565231 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ad510a18-cd72-4cf0-9b0c-dd3540d491fa-console-config\") pod \"console-f9d7485db-k6kls\" (UID: \"ad510a18-cd72-4cf0-9b0c-dd3540d491fa\") " pod="openshift-console/console-f9d7485db-k6kls" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.565274 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/7704cded-5ba1-4474-8a31-c0ba947b7679-images\") pod \"machine-api-operator-5694c8668f-d89dq\" (UID: \"7704cded-5ba1-4474-8a31-c0ba947b7679\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-d89dq" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.565292 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d825v\" (UniqueName: \"kubernetes.io/projected/75da7b9a-7085-492b-954a-6bb920fbb186-kube-api-access-d825v\") pod \"apiserver-76f77b778f-l88b4\" (UID: \"75da7b9a-7085-492b-954a-6bb920fbb186\") " pod="openshift-apiserver/apiserver-76f77b778f-l88b4" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.565309 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rm2hc\" (UniqueName: \"kubernetes.io/projected/ad510a18-cd72-4cf0-9b0c-dd3540d491fa-kube-api-access-rm2hc\") pod \"console-f9d7485db-k6kls\" (UID: \"ad510a18-cd72-4cf0-9b0c-dd3540d491fa\") " pod="openshift-console/console-f9d7485db-k6kls" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.565354 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bb4dde0f-222d-4609-b1c9-fed653668e3a-serving-cert\") pod \"console-operator-58897d9998-lfxkq\" (UID: \"bb4dde0f-222d-4609-b1c9-fed653668e3a\") " pod="openshift-console-operator/console-operator-58897d9998-lfxkq" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.565374 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/8c854339-76bb-439d-b267-b2617cdd3f41-etcd-ca\") pod \"etcd-operator-b45778765-r89c2\" (UID: \"8c854339-76bb-439d-b267-b2617cdd3f41\") " pod="openshift-etcd-operator/etcd-operator-b45778765-r89c2" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.565382 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/75da7b9a-7085-492b-954a-6bb920fbb186-audit\") pod \"apiserver-76f77b778f-l88b4\" (UID: \"75da7b9a-7085-492b-954a-6bb920fbb186\") " pod="openshift-apiserver/apiserver-76f77b778f-l88b4" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.565394 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e87b6d85-2d13-4b72-bb6b-8299b6c72d0a-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-gndwh\" (UID: \"e87b6d85-2d13-4b72-bb6b-8299b6c72d0a\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gndwh" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.565427 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/75da7b9a-7085-492b-954a-6bb920fbb186-encryption-config\") pod \"apiserver-76f77b778f-l88b4\" (UID: \"75da7b9a-7085-492b-954a-6bb920fbb186\") " pod="openshift-apiserver/apiserver-76f77b778f-l88b4" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.565444 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/75da7b9a-7085-492b-954a-6bb920fbb186-node-pullsecrets\") pod \"apiserver-76f77b778f-l88b4\" (UID: \"75da7b9a-7085-492b-954a-6bb920fbb186\") " pod="openshift-apiserver/apiserver-76f77b778f-l88b4" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.565462 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4kfvr\" (UniqueName: \"kubernetes.io/projected/8e2095d8-25c6-407f-aadb-f4603ae44d15-kube-api-access-4kfvr\") pod \"cluster-samples-operator-665b6dd947-krqhc\" (UID: \"8e2095d8-25c6-407f-aadb-f4603ae44d15\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-krqhc" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.565498 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb4dde0f-222d-4609-b1c9-fed653668e3a-config\") pod \"console-operator-58897d9998-lfxkq\" (UID: \"bb4dde0f-222d-4609-b1c9-fed653668e3a\") " pod="openshift-console-operator/console-operator-58897d9998-lfxkq" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.565516 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cgv6v\" (UniqueName: \"kubernetes.io/projected/fb52538d-050a-4b69-91b2-aa48f9833fe7-kube-api-access-cgv6v\") pod \"ingress-operator-5b745b69d9-hxqw8\" (UID: \"fb52538d-050a-4b69-91b2-aa48f9833fe7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hxqw8" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.565533 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fb52538d-050a-4b69-91b2-aa48f9833fe7-trusted-ca\") pod \"ingress-operator-5b745b69d9-hxqw8\" (UID: \"fb52538d-050a-4b69-91b2-aa48f9833fe7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hxqw8" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.565552 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e87b6d85-2d13-4b72-bb6b-8299b6c72d0a-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-gndwh\" (UID: \"e87b6d85-2d13-4b72-bb6b-8299b6c72d0a\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gndwh" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.565586 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/75da7b9a-7085-492b-954a-6bb920fbb186-serving-cert\") pod \"apiserver-76f77b778f-l88b4\" (UID: \"75da7b9a-7085-492b-954a-6bb920fbb186\") " pod="openshift-apiserver/apiserver-76f77b778f-l88b4" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.565611 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/247d17b7-7064-4ef1-9aed-c205fec84bc2-config\") pod \"kube-apiserver-operator-766d6c64bb-2zq2n\" (UID: \"247d17b7-7064-4ef1-9aed-c205fec84bc2\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2zq2n" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.565652 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zzw5l\" (UniqueName: \"kubernetes.io/projected/7704cded-5ba1-4474-8a31-c0ba947b7679-kube-api-access-zzw5l\") pod \"machine-api-operator-5694c8668f-d89dq\" (UID: \"7704cded-5ba1-4474-8a31-c0ba947b7679\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-d89dq" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.565670 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ad510a18-cd72-4cf0-9b0c-dd3540d491fa-trusted-ca-bundle\") pod \"console-f9d7485db-k6kls\" (UID: \"ad510a18-cd72-4cf0-9b0c-dd3540d491fa\") " pod="openshift-console/console-f9d7485db-k6kls" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.565692 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fa007484-938d-460e-9c5e-c3f5b31a25d2-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ks7hm\" (UID: \"fa007484-938d-460e-9c5e-c3f5b31a25d2\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ks7hm" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.565706 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/75da7b9a-7085-492b-954a-6bb920fbb186-image-import-ca\") pod \"apiserver-76f77b778f-l88b4\" (UID: \"75da7b9a-7085-492b-954a-6bb920fbb186\") " pod="openshift-apiserver/apiserver-76f77b778f-l88b4" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.564955 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ad510a18-cd72-4cf0-9b0c-dd3540d491fa-oauth-serving-cert\") pod \"console-f9d7485db-k6kls\" (UID: \"ad510a18-cd72-4cf0-9b0c-dd3540d491fa\") " pod="openshift-console/console-f9d7485db-k6kls" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.565876 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ad510a18-cd72-4cf0-9b0c-dd3540d491fa-console-config\") pod \"console-f9d7485db-k6kls\" (UID: \"ad510a18-cd72-4cf0-9b0c-dd3540d491fa\") " pod="openshift-console/console-f9d7485db-k6kls" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.566713 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/75da7b9a-7085-492b-954a-6bb920fbb186-node-pullsecrets\") pod \"apiserver-76f77b778f-l88b4\" (UID: \"75da7b9a-7085-492b-954a-6bb920fbb186\") " pod="openshift-apiserver/apiserver-76f77b778f-l88b4" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.567038 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/8c854339-76bb-439d-b267-b2617cdd3f41-etcd-ca\") pod \"etcd-operator-b45778765-r89c2\" (UID: \"8c854339-76bb-439d-b267-b2617cdd3f41\") " pod="openshift-etcd-operator/etcd-operator-b45778765-r89c2" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.567055 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ad510a18-cd72-4cf0-9b0c-dd3540d491fa-trusted-ca-bundle\") pod \"console-f9d7485db-k6kls\" (UID: \"ad510a18-cd72-4cf0-9b0c-dd3540d491fa\") " pod="openshift-console/console-f9d7485db-k6kls" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.567197 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8c854339-76bb-439d-b267-b2617cdd3f41-config\") pod \"etcd-operator-b45778765-r89c2\" (UID: \"8c854339-76bb-439d-b267-b2617cdd3f41\") " pod="openshift-etcd-operator/etcd-operator-b45778765-r89c2" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.567586 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/7704cded-5ba1-4474-8a31-c0ba947b7679-images\") pod \"machine-api-operator-5694c8668f-d89dq\" (UID: \"7704cded-5ba1-4474-8a31-c0ba947b7679\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-d89dq" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.567807 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb4dde0f-222d-4609-b1c9-fed653668e3a-config\") pod \"console-operator-58897d9998-lfxkq\" (UID: \"bb4dde0f-222d-4609-b1c9-fed653668e3a\") " pod="openshift-console-operator/console-operator-58897d9998-lfxkq" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.567901 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/75da7b9a-7085-492b-954a-6bb920fbb186-etcd-client\") pod \"apiserver-76f77b778f-l88b4\" (UID: \"75da7b9a-7085-492b-954a-6bb920fbb186\") " pod="openshift-apiserver/apiserver-76f77b778f-l88b4" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.567915 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/8e2095d8-25c6-407f-aadb-f4603ae44d15-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-krqhc\" (UID: \"8e2095d8-25c6-407f-aadb-f4603ae44d15\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-krqhc" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.568376 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8c854339-76bb-439d-b267-b2617cdd3f41-serving-cert\") pod \"etcd-operator-b45778765-r89c2\" (UID: \"8c854339-76bb-439d-b267-b2617cdd3f41\") " pod="openshift-etcd-operator/etcd-operator-b45778765-r89c2" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.568791 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7704cded-5ba1-4474-8a31-c0ba947b7679-config\") pod \"machine-api-operator-5694c8668f-d89dq\" (UID: \"7704cded-5ba1-4474-8a31-c0ba947b7679\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-d89dq" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.568976 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/d5fb9349-891c-447a-bb5a-5e18a0ce9f92-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-lg9s6\" (UID: \"d5fb9349-891c-447a-bb5a-5e18a0ce9f92\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lg9s6" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.569325 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/8650bbf0-78d2-410e-a62d-d4adf673cb55-metrics-tls\") pod \"dns-operator-744455d44c-kkw8v\" (UID: \"8650bbf0-78d2-410e-a62d-d4adf673cb55\") " pod="openshift-dns-operator/dns-operator-744455d44c-kkw8v" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.570192 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ad510a18-cd72-4cf0-9b0c-dd3540d491fa-console-oauth-config\") pod \"console-f9d7485db-k6kls\" (UID: \"ad510a18-cd72-4cf0-9b0c-dd3540d491fa\") " pod="openshift-console/console-f9d7485db-k6kls" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.570820 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/75da7b9a-7085-492b-954a-6bb920fbb186-encryption-config\") pod \"apiserver-76f77b778f-l88b4\" (UID: \"75da7b9a-7085-492b-954a-6bb920fbb186\") " pod="openshift-apiserver/apiserver-76f77b778f-l88b4" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.571168 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/75da7b9a-7085-492b-954a-6bb920fbb186-image-import-ca\") pod \"apiserver-76f77b778f-l88b4\" (UID: \"75da7b9a-7085-492b-954a-6bb920fbb186\") " pod="openshift-apiserver/apiserver-76f77b778f-l88b4" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.571277 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ad510a18-cd72-4cf0-9b0c-dd3540d491fa-console-serving-cert\") pod \"console-f9d7485db-k6kls\" (UID: \"ad510a18-cd72-4cf0-9b0c-dd3540d491fa\") " pod="openshift-console/console-f9d7485db-k6kls" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.572265 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/75da7b9a-7085-492b-954a-6bb920fbb186-serving-cert\") pod \"apiserver-76f77b778f-l88b4\" (UID: \"75da7b9a-7085-492b-954a-6bb920fbb186\") " pod="openshift-apiserver/apiserver-76f77b778f-l88b4" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.572829 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bb4dde0f-222d-4609-b1c9-fed653668e3a-serving-cert\") pod \"console-operator-58897d9998-lfxkq\" (UID: \"bb4dde0f-222d-4609-b1c9-fed653668e3a\") " pod="openshift-console-operator/console-operator-58897d9998-lfxkq" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.574214 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/7704cded-5ba1-4474-8a31-c0ba947b7679-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-d89dq\" (UID: \"7704cded-5ba1-4474-8a31-c0ba947b7679\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-d89dq" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.576155 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.595394 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.603382 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e87b6d85-2d13-4b72-bb6b-8299b6c72d0a-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-gndwh\" (UID: \"e87b6d85-2d13-4b72-bb6b-8299b6c72d0a\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gndwh" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.615655 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.617881 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e87b6d85-2d13-4b72-bb6b-8299b6c72d0a-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-gndwh\" (UID: \"e87b6d85-2d13-4b72-bb6b-8299b6c72d0a\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gndwh" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.635246 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.655719 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.675960 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.696454 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.735452 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.755218 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.775849 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.796196 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.818311 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.829998 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/fb52538d-050a-4b69-91b2-aa48f9833fe7-metrics-tls\") pod \"ingress-operator-5b745b69d9-hxqw8\" (UID: \"fb52538d-050a-4b69-91b2-aa48f9833fe7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hxqw8" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.845684 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.849969 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fb52538d-050a-4b69-91b2-aa48f9833fe7-trusted-ca\") pod \"ingress-operator-5b745b69d9-hxqw8\" (UID: \"fb52538d-050a-4b69-91b2-aa48f9833fe7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hxqw8" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.856024 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.876658 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.896282 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.917020 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.936004 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.945770 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/247d17b7-7064-4ef1-9aed-c205fec84bc2-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-2zq2n\" (UID: \"247d17b7-7064-4ef1-9aed-c205fec84bc2\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2zq2n" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.957190 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.958446 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/247d17b7-7064-4ef1-9aed-c205fec84bc2-config\") pod \"kube-apiserver-operator-766d6c64bb-2zq2n\" (UID: \"247d17b7-7064-4ef1-9aed-c205fec84bc2\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2zq2n" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.976654 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 11 02:03:13 crc kubenswrapper[4824]: I1211 02:03:13.997217 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 11 02:03:14 crc kubenswrapper[4824]: I1211 02:03:14.012249 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fa007484-938d-460e-9c5e-c3f5b31a25d2-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ks7hm\" (UID: \"fa007484-938d-460e-9c5e-c3f5b31a25d2\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ks7hm" Dec 11 02:03:14 crc kubenswrapper[4824]: I1211 02:03:14.015892 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 11 02:03:14 crc kubenswrapper[4824]: I1211 02:03:14.026725 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa007484-938d-460e-9c5e-c3f5b31a25d2-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ks7hm\" (UID: \"fa007484-938d-460e-9c5e-c3f5b31a25d2\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ks7hm" Dec 11 02:03:14 crc kubenswrapper[4824]: I1211 02:03:14.035718 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 11 02:03:14 crc kubenswrapper[4824]: I1211 02:03:14.063886 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 11 02:03:14 crc kubenswrapper[4824]: I1211 02:03:14.075672 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 11 02:03:14 crc kubenswrapper[4824]: I1211 02:03:14.097502 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 11 02:03:14 crc kubenswrapper[4824]: I1211 02:03:14.136463 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 11 02:03:14 crc kubenswrapper[4824]: I1211 02:03:14.155825 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 11 02:03:14 crc kubenswrapper[4824]: I1211 02:03:14.176610 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 11 02:03:14 crc kubenswrapper[4824]: I1211 02:03:14.196255 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 11 02:03:14 crc kubenswrapper[4824]: I1211 02:03:14.216167 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 11 02:03:14 crc kubenswrapper[4824]: I1211 02:03:14.236534 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 11 02:03:14 crc kubenswrapper[4824]: I1211 02:03:14.266586 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 11 02:03:14 crc kubenswrapper[4824]: I1211 02:03:14.275766 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 11 02:03:14 crc kubenswrapper[4824]: I1211 02:03:14.296779 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 11 02:03:14 crc kubenswrapper[4824]: I1211 02:03:14.316919 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 11 02:03:14 crc kubenswrapper[4824]: I1211 02:03:14.336522 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 11 02:03:14 crc kubenswrapper[4824]: I1211 02:03:14.356191 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 11 02:03:14 crc kubenswrapper[4824]: I1211 02:03:14.376552 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 11 02:03:14 crc kubenswrapper[4824]: I1211 02:03:14.394302 4824 request.go:700] Waited for 1.01141297s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-ingress/configmaps?fieldSelector=metadata.name%3Dopenshift-service-ca.crt&limit=500&resourceVersion=0 Dec 11 02:03:14 crc kubenswrapper[4824]: I1211 02:03:14.396351 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 11 02:03:14 crc kubenswrapper[4824]: I1211 02:03:14.416659 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 11 02:03:14 crc kubenswrapper[4824]: I1211 02:03:14.436623 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 11 02:03:14 crc kubenswrapper[4824]: I1211 02:03:14.455928 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 11 02:03:14 crc kubenswrapper[4824]: I1211 02:03:14.476430 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 11 02:03:14 crc kubenswrapper[4824]: I1211 02:03:14.496494 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 11 02:03:14 crc kubenswrapper[4824]: I1211 02:03:14.516669 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 11 02:03:14 crc kubenswrapper[4824]: I1211 02:03:14.536103 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 11 02:03:14 crc kubenswrapper[4824]: I1211 02:03:14.556592 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 11 02:03:14 crc kubenswrapper[4824]: I1211 02:03:14.575963 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 11 02:03:14 crc kubenswrapper[4824]: I1211 02:03:14.597068 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 11 02:03:14 crc kubenswrapper[4824]: I1211 02:03:14.615728 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 11 02:03:14 crc kubenswrapper[4824]: I1211 02:03:14.636357 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 11 02:03:14 crc kubenswrapper[4824]: I1211 02:03:14.656622 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 11 02:03:14 crc kubenswrapper[4824]: I1211 02:03:14.675934 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 11 02:03:14 crc kubenswrapper[4824]: I1211 02:03:14.696561 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 11 02:03:14 crc kubenswrapper[4824]: I1211 02:03:14.716898 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 11 02:03:14 crc kubenswrapper[4824]: I1211 02:03:14.735904 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 11 02:03:14 crc kubenswrapper[4824]: I1211 02:03:14.757464 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 11 02:03:14 crc kubenswrapper[4824]: I1211 02:03:14.775617 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 11 02:03:14 crc kubenswrapper[4824]: I1211 02:03:14.797300 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 11 02:03:14 crc kubenswrapper[4824]: I1211 02:03:14.815619 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 11 02:03:14 crc kubenswrapper[4824]: I1211 02:03:14.836619 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 11 02:03:14 crc kubenswrapper[4824]: I1211 02:03:14.856660 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 11 02:03:14 crc kubenswrapper[4824]: I1211 02:03:14.876791 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 11 02:03:14 crc kubenswrapper[4824]: I1211 02:03:14.896205 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 11 02:03:14 crc kubenswrapper[4824]: I1211 02:03:14.916749 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 11 02:03:14 crc kubenswrapper[4824]: I1211 02:03:14.936079 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 11 02:03:14 crc kubenswrapper[4824]: I1211 02:03:14.956847 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 11 02:03:14 crc kubenswrapper[4824]: I1211 02:03:14.976309 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 11 02:03:14 crc kubenswrapper[4824]: I1211 02:03:14.996609 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.016052 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.036704 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.056440 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.076246 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.096752 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.116381 4824 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.136649 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.156577 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.176784 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.196534 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.238331 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6r6md\" (UniqueName: \"kubernetes.io/projected/9e099899-aaf7-4bd8-b37e-8616b5060e33-kube-api-access-6r6md\") pod \"controller-manager-879f6c89f-qr5xz\" (UID: \"9e099899-aaf7-4bd8-b37e-8616b5060e33\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qr5xz" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.267589 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4x4bd\" (UniqueName: \"kubernetes.io/projected/5407364a-e2b6-43b4-9eb0-deb12a022289-kube-api-access-4x4bd\") pod \"route-controller-manager-6576b87f9c-msljv\" (UID: \"5407364a-e2b6-43b4-9eb0-deb12a022289\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-msljv" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.276471 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n28fr\" (UniqueName: \"kubernetes.io/projected/1bdf5eb2-fe05-4e24-9e00-43c72afc3583-kube-api-access-n28fr\") pod \"oauth-openshift-558db77b4-l9d4f\" (UID: \"1bdf5eb2-fe05-4e24-9e00-43c72afc3583\") " pod="openshift-authentication/oauth-openshift-558db77b4-l9d4f" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.306394 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6587n\" (UniqueName: \"kubernetes.io/projected/3e0e52dd-503e-47a4-b68e-3f768ccecfc6-kube-api-access-6587n\") pod \"apiserver-7bbb656c7d-g692d\" (UID: \"3e0e52dd-503e-47a4-b68e-3f768ccecfc6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g692d" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.319792 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-grgzk\" (UniqueName: \"kubernetes.io/projected/f257970f-c1ce-44fb-b207-a78126794cfe-kube-api-access-grgzk\") pod \"authentication-operator-69f744f599-tfwnz\" (UID: \"f257970f-c1ce-44fb-b207-a78126794cfe\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-tfwnz" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.338296 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-874kt\" (UniqueName: \"kubernetes.io/projected/3404cb6f-6c3a-4e0a-99a4-27530beefa83-kube-api-access-874kt\") pod \"machine-approver-56656f9798-vtdlb\" (UID: \"3404cb6f-6c3a-4e0a-99a4-27530beefa83\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vtdlb" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.355866 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.362841 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-htnkn\" (UniqueName: \"kubernetes.io/projected/29e1953f-668a-4393-82ab-a45fa4b79ec7-kube-api-access-htnkn\") pod \"openshift-apiserver-operator-796bbdcf4f-2vcc6\" (UID: \"29e1953f-668a-4393-82ab-a45fa4b79ec7\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2vcc6" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.372754 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g692d" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.375711 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.388090 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-qr5xz" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.396297 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.414084 4824 request.go:700] Waited for 1.853512787s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/serviceaccounts/cluster-image-registry-operator/token Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.436151 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-msljv" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.447340 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d5fb9349-891c-447a-bb5a-5e18a0ce9f92-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-lg9s6\" (UID: \"d5fb9349-891c-447a-bb5a-5e18a0ce9f92\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lg9s6" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.449906 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vtdlb" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.462100 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-tfwnz" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.467706 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7js74\" (UniqueName: \"kubernetes.io/projected/d5fb9349-891c-447a-bb5a-5e18a0ce9f92-kube-api-access-7js74\") pod \"cluster-image-registry-operator-dc59b4c8b-lg9s6\" (UID: \"d5fb9349-891c-447a-bb5a-5e18a0ce9f92\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lg9s6" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.476989 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wgmgc\" (UniqueName: \"kubernetes.io/projected/e87b6d85-2d13-4b72-bb6b-8299b6c72d0a-kube-api-access-wgmgc\") pod \"kube-storage-version-migrator-operator-b67b599dd-gndwh\" (UID: \"e87b6d85-2d13-4b72-bb6b-8299b6c72d0a\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gndwh" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.483620 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-l9d4f" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.488587 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2vcc6" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.506328 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4d4xx\" (UniqueName: \"kubernetes.io/projected/8c854339-76bb-439d-b267-b2617cdd3f41-kube-api-access-4d4xx\") pod \"etcd-operator-b45778765-r89c2\" (UID: \"8c854339-76bb-439d-b267-b2617cdd3f41\") " pod="openshift-etcd-operator/etcd-operator-b45778765-r89c2" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.527893 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/247d17b7-7064-4ef1-9aed-c205fec84bc2-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-2zq2n\" (UID: \"247d17b7-7064-4ef1-9aed-c205fec84bc2\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2zq2n" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.545083 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fa007484-938d-460e-9c5e-c3f5b31a25d2-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ks7hm\" (UID: \"fa007484-938d-460e-9c5e-c3f5b31a25d2\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ks7hm" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.573230 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdj7r\" (UniqueName: \"kubernetes.io/projected/f185e5b2-54c0-44ef-bdda-118efa89cb5f-kube-api-access-rdj7r\") pod \"downloads-7954f5f757-2ng4n\" (UID: \"f185e5b2-54c0-44ef-bdda-118efa89cb5f\") " pod="openshift-console/downloads-7954f5f757-2ng4n" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.579508 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/fb52538d-050a-4b69-91b2-aa48f9833fe7-bound-sa-token\") pod \"ingress-operator-5b745b69d9-hxqw8\" (UID: \"fb52538d-050a-4b69-91b2-aa48f9833fe7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hxqw8" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.591774 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-shkmj\" (UniqueName: \"kubernetes.io/projected/8650bbf0-78d2-410e-a62d-d4adf673cb55-kube-api-access-shkmj\") pod \"dns-operator-744455d44c-kkw8v\" (UID: \"8650bbf0-78d2-410e-a62d-d4adf673cb55\") " pod="openshift-dns-operator/dns-operator-744455d44c-kkw8v" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.601420 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-2ng4n" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.608461 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lg9s6" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.615979 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dvkb4\" (UniqueName: \"kubernetes.io/projected/bb4dde0f-222d-4609-b1c9-fed653668e3a-kube-api-access-dvkb4\") pod \"console-operator-58897d9998-lfxkq\" (UID: \"bb4dde0f-222d-4609-b1c9-fed653668e3a\") " pod="openshift-console-operator/console-operator-58897d9998-lfxkq" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.633461 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cgv6v\" (UniqueName: \"kubernetes.io/projected/fb52538d-050a-4b69-91b2-aa48f9833fe7-kube-api-access-cgv6v\") pod \"ingress-operator-5b745b69d9-hxqw8\" (UID: \"fb52538d-050a-4b69-91b2-aa48f9833fe7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hxqw8" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.633648 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-r89c2" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.643512 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gndwh" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.650288 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d825v\" (UniqueName: \"kubernetes.io/projected/75da7b9a-7085-492b-954a-6bb920fbb186-kube-api-access-d825v\") pod \"apiserver-76f77b778f-l88b4\" (UID: \"75da7b9a-7085-492b-954a-6bb920fbb186\") " pod="openshift-apiserver/apiserver-76f77b778f-l88b4" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.655398 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hxqw8" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.666368 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2zq2n" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.671342 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4kfvr\" (UniqueName: \"kubernetes.io/projected/8e2095d8-25c6-407f-aadb-f4603ae44d15-kube-api-access-4kfvr\") pod \"cluster-samples-operator-665b6dd947-krqhc\" (UID: \"8e2095d8-25c6-407f-aadb-f4603ae44d15\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-krqhc" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.692058 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rm2hc\" (UniqueName: \"kubernetes.io/projected/ad510a18-cd72-4cf0-9b0c-dd3540d491fa-kube-api-access-rm2hc\") pod \"console-f9d7485db-k6kls\" (UID: \"ad510a18-cd72-4cf0-9b0c-dd3540d491fa\") " pod="openshift-console/console-f9d7485db-k6kls" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.698547 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ks7hm" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.712843 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zzw5l\" (UniqueName: \"kubernetes.io/projected/7704cded-5ba1-4474-8a31-c0ba947b7679-kube-api-access-zzw5l\") pod \"machine-api-operator-5694c8668f-d89dq\" (UID: \"7704cded-5ba1-4474-8a31-c0ba947b7679\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-d89dq" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.792736 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-msljv"] Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.800226 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/18d74f73-0f04-4f3c-88c4-a431d490db9d-trusted-ca\") pod \"image-registry-697d97f7c8-254w2\" (UID: \"18d74f73-0f04-4f3c-88c4-a431d490db9d\") " pod="openshift-image-registry/image-registry-697d97f7c8-254w2" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.800266 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4d6h\" (UniqueName: \"kubernetes.io/projected/18d74f73-0f04-4f3c-88c4-a431d490db9d-kube-api-access-z4d6h\") pod \"image-registry-697d97f7c8-254w2\" (UID: \"18d74f73-0f04-4f3c-88c4-a431d490db9d\") " pod="openshift-image-registry/image-registry-697d97f7c8-254w2" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.800345 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qwjc7\" (UniqueName: \"kubernetes.io/projected/dbdb9d15-7786-4fef-a38e-e3be94f1bcb2-kube-api-access-qwjc7\") pod \"openshift-controller-manager-operator-756b6f6bc6-tz2j9\" (UID: \"dbdb9d15-7786-4fef-a38e-e3be94f1bcb2\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-tz2j9" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.800369 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/f5149a23-92a4-42e6-8b20-8411d1d0536f-available-featuregates\") pod \"openshift-config-operator-7777fb866f-w6b4l\" (UID: \"f5149a23-92a4-42e6-8b20-8411d1d0536f\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-w6b4l" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.800425 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-254w2\" (UID: \"18d74f73-0f04-4f3c-88c4-a431d490db9d\") " pod="openshift-image-registry/image-registry-697d97f7c8-254w2" Dec 11 02:03:15 crc kubenswrapper[4824]: E1211 02:03:15.800720 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 02:03:16.300708882 +0000 UTC m=+137.989746261 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-254w2" (UID: "18d74f73-0f04-4f3c-88c4-a431d490db9d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.800902 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/18d74f73-0f04-4f3c-88c4-a431d490db9d-ca-trust-extracted\") pod \"image-registry-697d97f7c8-254w2\" (UID: \"18d74f73-0f04-4f3c-88c4-a431d490db9d\") " pod="openshift-image-registry/image-registry-697d97f7c8-254w2" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.801385 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zm5nf\" (UniqueName: \"kubernetes.io/projected/f5149a23-92a4-42e6-8b20-8411d1d0536f-kube-api-access-zm5nf\") pod \"openshift-config-operator-7777fb866f-w6b4l\" (UID: \"f5149a23-92a4-42e6-8b20-8411d1d0536f\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-w6b4l" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.801434 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/18d74f73-0f04-4f3c-88c4-a431d490db9d-installation-pull-secrets\") pod \"image-registry-697d97f7c8-254w2\" (UID: \"18d74f73-0f04-4f3c-88c4-a431d490db9d\") " pod="openshift-image-registry/image-registry-697d97f7c8-254w2" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.801462 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dbdb9d15-7786-4fef-a38e-e3be94f1bcb2-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-tz2j9\" (UID: \"dbdb9d15-7786-4fef-a38e-e3be94f1bcb2\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-tz2j9" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.801485 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/18d74f73-0f04-4f3c-88c4-a431d490db9d-registry-certificates\") pod \"image-registry-697d97f7c8-254w2\" (UID: \"18d74f73-0f04-4f3c-88c4-a431d490db9d\") " pod="openshift-image-registry/image-registry-697d97f7c8-254w2" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.801500 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f5149a23-92a4-42e6-8b20-8411d1d0536f-serving-cert\") pod \"openshift-config-operator-7777fb866f-w6b4l\" (UID: \"f5149a23-92a4-42e6-8b20-8411d1d0536f\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-w6b4l" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.801541 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dbdb9d15-7786-4fef-a38e-e3be94f1bcb2-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-tz2j9\" (UID: \"dbdb9d15-7786-4fef-a38e-e3be94f1bcb2\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-tz2j9" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.801573 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/18d74f73-0f04-4f3c-88c4-a431d490db9d-registry-tls\") pod \"image-registry-697d97f7c8-254w2\" (UID: \"18d74f73-0f04-4f3c-88c4-a431d490db9d\") " pod="openshift-image-registry/image-registry-697d97f7c8-254w2" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.801623 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/18d74f73-0f04-4f3c-88c4-a431d490db9d-bound-sa-token\") pod \"image-registry-697d97f7c8-254w2\" (UID: \"18d74f73-0f04-4f3c-88c4-a431d490db9d\") " pod="openshift-image-registry/image-registry-697d97f7c8-254w2" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.810577 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-krqhc" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.814569 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-lfxkq" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.840343 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-kkw8v" Dec 11 02:03:15 crc kubenswrapper[4824]: W1211 02:03:15.840728 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5407364a_e2b6_43b4_9eb0_deb12a022289.slice/crio-1311461ca6b530fd2d5299f8f6c5a8f9f353e64c24afe69528bcac3772261b60 WatchSource:0}: Error finding container 1311461ca6b530fd2d5299f8f6c5a8f9f353e64c24afe69528bcac3772261b60: Status 404 returned error can't find the container with id 1311461ca6b530fd2d5299f8f6c5a8f9f353e64c24afe69528bcac3772261b60 Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.863498 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-k6kls" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.891147 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-qr5xz"] Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.891199 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-g692d"] Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.902728 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.902943 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dbdb9d15-7786-4fef-a38e-e3be94f1bcb2-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-tz2j9\" (UID: \"dbdb9d15-7786-4fef-a38e-e3be94f1bcb2\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-tz2j9" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.902984 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/335b007e-ea5c-4e20-9402-5599da94ef22-plugins-dir\") pod \"csi-hostpathplugin-hr4qs\" (UID: \"335b007e-ea5c-4e20-9402-5599da94ef22\") " pod="hostpath-provisioner/csi-hostpathplugin-hr4qs" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.903032 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4j7mj\" (UniqueName: \"kubernetes.io/projected/56378af0-da53-4ffc-a66c-a381083e1ba6-kube-api-access-4j7mj\") pod \"catalog-operator-68c6474976-qqg25\" (UID: \"56378af0-da53-4ffc-a66c-a381083e1ba6\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qqg25" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.903056 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kpq4n\" (UniqueName: \"kubernetes.io/projected/09112ee7-18d2-4a6d-83ce-eddfce56b083-kube-api-access-kpq4n\") pod \"service-ca-operator-777779d784-59bgf\" (UID: \"09112ee7-18d2-4a6d-83ce-eddfce56b083\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-59bgf" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.903126 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dccdfcce-6b35-40e6-8bdc-7eed59518f24-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-kdzkc\" (UID: \"dccdfcce-6b35-40e6-8bdc-7eed59518f24\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kdzkc" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.903151 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/56378af0-da53-4ffc-a66c-a381083e1ba6-srv-cert\") pod \"catalog-operator-68c6474976-qqg25\" (UID: \"56378af0-da53-4ffc-a66c-a381083e1ba6\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qqg25" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.903171 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/335b007e-ea5c-4e20-9402-5599da94ef22-mountpoint-dir\") pod \"csi-hostpathplugin-hr4qs\" (UID: \"335b007e-ea5c-4e20-9402-5599da94ef22\") " pod="hostpath-provisioner/csi-hostpathplugin-hr4qs" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.903220 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/1c402f2f-f79a-4a1a-bf40-07da4223059e-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-qp2jl\" (UID: \"1c402f2f-f79a-4a1a-bf40-07da4223059e\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-qp2jl" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.903245 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/52938126-55df-4106-9bed-ad7cc0f1130d-node-bootstrap-token\") pod \"machine-config-server-6b6br\" (UID: \"52938126-55df-4106-9bed-ad7cc0f1130d\") " pod="openshift-machine-config-operator/machine-config-server-6b6br" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.903269 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/505f6458-5d5a-4725-a529-1075770afa95-proxy-tls\") pod \"machine-config-controller-84d6567774-vkdbw\" (UID: \"505f6458-5d5a-4725-a529-1075770afa95\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vkdbw" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.903335 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4d6h\" (UniqueName: \"kubernetes.io/projected/18d74f73-0f04-4f3c-88c4-a431d490db9d-kube-api-access-z4d6h\") pod \"image-registry-697d97f7c8-254w2\" (UID: \"18d74f73-0f04-4f3c-88c4-a431d490db9d\") " pod="openshift-image-registry/image-registry-697d97f7c8-254w2" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.903358 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/04cb3d8f-ccf2-410c-b51c-445add07699c-proxy-tls\") pod \"machine-config-operator-74547568cd-jlqk5\" (UID: \"04cb3d8f-ccf2-410c-b51c-445add07699c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-jlqk5" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.903408 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/335b007e-ea5c-4e20-9402-5599da94ef22-registration-dir\") pod \"csi-hostpathplugin-hr4qs\" (UID: \"335b007e-ea5c-4e20-9402-5599da94ef22\") " pod="hostpath-provisioner/csi-hostpathplugin-hr4qs" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.903433 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/877d4a7f-46b0-4f9c-a702-b6b83620c3cb-metrics-certs\") pod \"router-default-5444994796-pg8rn\" (UID: \"877d4a7f-46b0-4f9c-a702-b6b83620c3cb\") " pod="openshift-ingress/router-default-5444994796-pg8rn" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.903457 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dccdfcce-6b35-40e6-8bdc-7eed59518f24-config\") pod \"kube-controller-manager-operator-78b949d7b-kdzkc\" (UID: \"dccdfcce-6b35-40e6-8bdc-7eed59518f24\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kdzkc" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.903478 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dccdfcce-6b35-40e6-8bdc-7eed59518f24-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-kdzkc\" (UID: \"dccdfcce-6b35-40e6-8bdc-7eed59518f24\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kdzkc" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.903564 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ffb4acd4-543f-4509-b993-249e3656048d-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-z42g5\" (UID: \"ffb4acd4-543f-4509-b993-249e3656048d\") " pod="openshift-marketplace/marketplace-operator-79b997595-z42g5" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.903661 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/335b007e-ea5c-4e20-9402-5599da94ef22-socket-dir\") pod \"csi-hostpathplugin-hr4qs\" (UID: \"335b007e-ea5c-4e20-9402-5599da94ef22\") " pod="hostpath-provisioner/csi-hostpathplugin-hr4qs" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.903713 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zm5nf\" (UniqueName: \"kubernetes.io/projected/f5149a23-92a4-42e6-8b20-8411d1d0536f-kube-api-access-zm5nf\") pod \"openshift-config-operator-7777fb866f-w6b4l\" (UID: \"f5149a23-92a4-42e6-8b20-8411d1d0536f\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-w6b4l" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.903737 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/56378af0-da53-4ffc-a66c-a381083e1ba6-profile-collector-cert\") pod \"catalog-operator-68c6474976-qqg25\" (UID: \"56378af0-da53-4ffc-a66c-a381083e1ba6\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qqg25" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.903901 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/09f964e8-942f-4577-983a-bdc08487c0e8-signing-key\") pod \"service-ca-9c57cc56f-4r8x8\" (UID: \"09f964e8-942f-4577-983a-bdc08487c0e8\") " pod="openshift-service-ca/service-ca-9c57cc56f-4r8x8" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.903933 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a4f0d12d-4a03-4ffd-9058-f311bf0041fe-config-volume\") pod \"dns-default-qtll9\" (UID: \"a4f0d12d-4a03-4ffd-9058-f311bf0041fe\") " pod="openshift-dns/dns-default-qtll9" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.903960 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dbdb9d15-7786-4fef-a38e-e3be94f1bcb2-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-tz2j9\" (UID: \"dbdb9d15-7786-4fef-a38e-e3be94f1bcb2\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-tz2j9" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.903981 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09112ee7-18d2-4a6d-83ce-eddfce56b083-config\") pod \"service-ca-operator-777779d784-59bgf\" (UID: \"09112ee7-18d2-4a6d-83ce-eddfce56b083\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-59bgf" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.904081 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cwz4n\" (UniqueName: \"kubernetes.io/projected/877d4a7f-46b0-4f9c-a702-b6b83620c3cb-kube-api-access-cwz4n\") pod \"router-default-5444994796-pg8rn\" (UID: \"877d4a7f-46b0-4f9c-a702-b6b83620c3cb\") " pod="openshift-ingress/router-default-5444994796-pg8rn" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.904137 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/59dd9dac-e58a-4869-b1ee-e0c5e8a78974-webhook-cert\") pod \"packageserver-d55dfcdfc-p8x2v\" (UID: \"59dd9dac-e58a-4869-b1ee-e0c5e8a78974\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p8x2v" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.904224 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/09f964e8-942f-4577-983a-bdc08487c0e8-signing-cabundle\") pod \"service-ca-9c57cc56f-4r8x8\" (UID: \"09f964e8-942f-4577-983a-bdc08487c0e8\") " pod="openshift-service-ca/service-ca-9c57cc56f-4r8x8" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.904263 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-njgbl\" (UniqueName: \"kubernetes.io/projected/63fe488c-de20-435f-83fd-47a4179341e6-kube-api-access-njgbl\") pod \"ingress-canary-zvcj8\" (UID: \"63fe488c-de20-435f-83fd-47a4179341e6\") " pod="openshift-ingress-canary/ingress-canary-zvcj8" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.904303 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/04cb3d8f-ccf2-410c-b51c-445add07699c-images\") pod \"machine-config-operator-74547568cd-jlqk5\" (UID: \"04cb3d8f-ccf2-410c-b51c-445add07699c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-jlqk5" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.904326 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/59dd9dac-e58a-4869-b1ee-e0c5e8a78974-apiservice-cert\") pod \"packageserver-d55dfcdfc-p8x2v\" (UID: \"59dd9dac-e58a-4869-b1ee-e0c5e8a78974\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p8x2v" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.904360 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ffb4acd4-543f-4509-b993-249e3656048d-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-z42g5\" (UID: \"ffb4acd4-543f-4509-b993-249e3656048d\") " pod="openshift-marketplace/marketplace-operator-79b997595-z42g5" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.904383 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/04cb3d8f-ccf2-410c-b51c-445add07699c-auth-proxy-config\") pod \"machine-config-operator-74547568cd-jlqk5\" (UID: \"04cb3d8f-ccf2-410c-b51c-445add07699c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-jlqk5" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.904419 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/18d74f73-0f04-4f3c-88c4-a431d490db9d-registry-tls\") pod \"image-registry-697d97f7c8-254w2\" (UID: \"18d74f73-0f04-4f3c-88c4-a431d490db9d\") " pod="openshift-image-registry/image-registry-697d97f7c8-254w2" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.904443 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4c22bf31-ad33-44cd-a495-05d83688f2bd-config-volume\") pod \"collect-profiles-29423640-42vd6\" (UID: \"4c22bf31-ad33-44cd-a495-05d83688f2bd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423640-42vd6" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.904494 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a4f0d12d-4a03-4ffd-9058-f311bf0041fe-metrics-tls\") pod \"dns-default-qtll9\" (UID: \"a4f0d12d-4a03-4ffd-9058-f311bf0041fe\") " pod="openshift-dns/dns-default-qtll9" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.904517 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/18d74f73-0f04-4f3c-88c4-a431d490db9d-bound-sa-token\") pod \"image-registry-697d97f7c8-254w2\" (UID: \"18d74f73-0f04-4f3c-88c4-a431d490db9d\") " pod="openshift-image-registry/image-registry-697d97f7c8-254w2" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.904540 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bbwzb\" (UniqueName: \"kubernetes.io/projected/a4f0d12d-4a03-4ffd-9058-f311bf0041fe-kube-api-access-bbwzb\") pod \"dns-default-qtll9\" (UID: \"a4f0d12d-4a03-4ffd-9058-f311bf0041fe\") " pod="openshift-dns/dns-default-qtll9" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.904563 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/877d4a7f-46b0-4f9c-a702-b6b83620c3cb-service-ca-bundle\") pod \"router-default-5444994796-pg8rn\" (UID: \"877d4a7f-46b0-4f9c-a702-b6b83620c3cb\") " pod="openshift-ingress/router-default-5444994796-pg8rn" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.904588 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/18d74f73-0f04-4f3c-88c4-a431d490db9d-trusted-ca\") pod \"image-registry-697d97f7c8-254w2\" (UID: \"18d74f73-0f04-4f3c-88c4-a431d490db9d\") " pod="openshift-image-registry/image-registry-697d97f7c8-254w2" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.904623 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/398ffffe-110f-4f78-a784-2ee501ee58a3-srv-cert\") pod \"olm-operator-6b444d44fb-66mv8\" (UID: \"398ffffe-110f-4f78-a784-2ee501ee58a3\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-66mv8" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.904646 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/63fe488c-de20-435f-83fd-47a4179341e6-cert\") pod \"ingress-canary-zvcj8\" (UID: \"63fe488c-de20-435f-83fd-47a4179341e6\") " pod="openshift-ingress-canary/ingress-canary-zvcj8" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.904666 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sgqgs\" (UniqueName: \"kubernetes.io/projected/95b90361-2f73-4649-bc1a-1202ba511baf-kube-api-access-sgqgs\") pod \"migrator-59844c95c7-8fksd\" (UID: \"95b90361-2f73-4649-bc1a-1202ba511baf\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8fksd" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.904737 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qwjc7\" (UniqueName: \"kubernetes.io/projected/dbdb9d15-7786-4fef-a38e-e3be94f1bcb2-kube-api-access-qwjc7\") pod \"openshift-controller-manager-operator-756b6f6bc6-tz2j9\" (UID: \"dbdb9d15-7786-4fef-a38e-e3be94f1bcb2\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-tz2j9" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.904762 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hlmx4\" (UniqueName: \"kubernetes.io/projected/59dd9dac-e58a-4869-b1ee-e0c5e8a78974-kube-api-access-hlmx4\") pod \"packageserver-d55dfcdfc-p8x2v\" (UID: \"59dd9dac-e58a-4869-b1ee-e0c5e8a78974\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p8x2v" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.904828 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/f5149a23-92a4-42e6-8b20-8411d1d0536f-available-featuregates\") pod \"openshift-config-operator-7777fb866f-w6b4l\" (UID: \"f5149a23-92a4-42e6-8b20-8411d1d0536f\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-w6b4l" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.904853 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/59dd9dac-e58a-4869-b1ee-e0c5e8a78974-tmpfs\") pod \"packageserver-d55dfcdfc-p8x2v\" (UID: \"59dd9dac-e58a-4869-b1ee-e0c5e8a78974\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p8x2v" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.904904 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-675ss\" (UniqueName: \"kubernetes.io/projected/52938126-55df-4106-9bed-ad7cc0f1130d-kube-api-access-675ss\") pod \"machine-config-server-6b6br\" (UID: \"52938126-55df-4106-9bed-ad7cc0f1130d\") " pod="openshift-machine-config-operator/machine-config-server-6b6br" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.904930 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/c70805eb-e095-4462-a202-e01860559471-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-gp8dt\" (UID: \"c70805eb-e095-4462-a202-e01860559471\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-gp8dt" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.904953 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/505f6458-5d5a-4725-a529-1075770afa95-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-vkdbw\" (UID: \"505f6458-5d5a-4725-a529-1075770afa95\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vkdbw" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.904992 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/18d74f73-0f04-4f3c-88c4-a431d490db9d-ca-trust-extracted\") pod \"image-registry-697d97f7c8-254w2\" (UID: \"18d74f73-0f04-4f3c-88c4-a431d490db9d\") " pod="openshift-image-registry/image-registry-697d97f7c8-254w2" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.905031 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/335b007e-ea5c-4e20-9402-5599da94ef22-csi-data-dir\") pod \"csi-hostpathplugin-hr4qs\" (UID: \"335b007e-ea5c-4e20-9402-5599da94ef22\") " pod="hostpath-provisioner/csi-hostpathplugin-hr4qs" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.905052 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dqzmb\" (UniqueName: \"kubernetes.io/projected/505f6458-5d5a-4725-a529-1075770afa95-kube-api-access-dqzmb\") pod \"machine-config-controller-84d6567774-vkdbw\" (UID: \"505f6458-5d5a-4725-a529-1075770afa95\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vkdbw" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.905073 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wt25m\" (UniqueName: \"kubernetes.io/projected/04cb3d8f-ccf2-410c-b51c-445add07699c-kube-api-access-wt25m\") pod \"machine-config-operator-74547568cd-jlqk5\" (UID: \"04cb3d8f-ccf2-410c-b51c-445add07699c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-jlqk5" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.905094 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/877d4a7f-46b0-4f9c-a702-b6b83620c3cb-default-certificate\") pod \"router-default-5444994796-pg8rn\" (UID: \"877d4a7f-46b0-4f9c-a702-b6b83620c3cb\") " pod="openshift-ingress/router-default-5444994796-pg8rn" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.905193 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4h2lj\" (UniqueName: \"kubernetes.io/projected/69814666-b43a-4cab-9864-864862277916-kube-api-access-4h2lj\") pod \"package-server-manager-789f6589d5-lpgf5\" (UID: \"69814666-b43a-4cab-9864-864862277916\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-lpgf5" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.905218 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-drmvw\" (UniqueName: \"kubernetes.io/projected/4c22bf31-ad33-44cd-a495-05d83688f2bd-kube-api-access-drmvw\") pod \"collect-profiles-29423640-42vd6\" (UID: \"4c22bf31-ad33-44cd-a495-05d83688f2bd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423640-42vd6" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.905267 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09112ee7-18d2-4a6d-83ce-eddfce56b083-serving-cert\") pod \"service-ca-operator-777779d784-59bgf\" (UID: \"09112ee7-18d2-4a6d-83ce-eddfce56b083\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-59bgf" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.905311 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/18d74f73-0f04-4f3c-88c4-a431d490db9d-installation-pull-secrets\") pod \"image-registry-697d97f7c8-254w2\" (UID: \"18d74f73-0f04-4f3c-88c4-a431d490db9d\") " pod="openshift-image-registry/image-registry-697d97f7c8-254w2" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.905334 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/398ffffe-110f-4f78-a784-2ee501ee58a3-profile-collector-cert\") pod \"olm-operator-6b444d44fb-66mv8\" (UID: \"398ffffe-110f-4f78-a784-2ee501ee58a3\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-66mv8" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.905374 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c6vtj\" (UniqueName: \"kubernetes.io/projected/09f964e8-942f-4577-983a-bdc08487c0e8-kube-api-access-c6vtj\") pod \"service-ca-9c57cc56f-4r8x8\" (UID: \"09f964e8-942f-4577-983a-bdc08487c0e8\") " pod="openshift-service-ca/service-ca-9c57cc56f-4r8x8" Dec 11 02:03:15 crc kubenswrapper[4824]: E1211 02:03:15.905762 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 02:03:16.405711186 +0000 UTC m=+138.094748565 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.906903 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dbdb9d15-7786-4fef-a38e-e3be94f1bcb2-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-tz2j9\" (UID: \"dbdb9d15-7786-4fef-a38e-e3be94f1bcb2\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-tz2j9" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.909105 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/f5149a23-92a4-42e6-8b20-8411d1d0536f-available-featuregates\") pod \"openshift-config-operator-7777fb866f-w6b4l\" (UID: \"f5149a23-92a4-42e6-8b20-8411d1d0536f\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-w6b4l" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.910425 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/18d74f73-0f04-4f3c-88c4-a431d490db9d-ca-trust-extracted\") pod \"image-registry-697d97f7c8-254w2\" (UID: \"18d74f73-0f04-4f3c-88c4-a431d490db9d\") " pod="openshift-image-registry/image-registry-697d97f7c8-254w2" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.910470 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hfh5h\" (UniqueName: \"kubernetes.io/projected/398ffffe-110f-4f78-a784-2ee501ee58a3-kube-api-access-hfh5h\") pod \"olm-operator-6b444d44fb-66mv8\" (UID: \"398ffffe-110f-4f78-a784-2ee501ee58a3\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-66mv8" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.910505 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/18d74f73-0f04-4f3c-88c4-a431d490db9d-registry-certificates\") pod \"image-registry-697d97f7c8-254w2\" (UID: \"18d74f73-0f04-4f3c-88c4-a431d490db9d\") " pod="openshift-image-registry/image-registry-697d97f7c8-254w2" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.910538 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/877d4a7f-46b0-4f9c-a702-b6b83620c3cb-stats-auth\") pod \"router-default-5444994796-pg8rn\" (UID: \"877d4a7f-46b0-4f9c-a702-b6b83620c3cb\") " pod="openshift-ingress/router-default-5444994796-pg8rn" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.910556 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4c22bf31-ad33-44cd-a495-05d83688f2bd-secret-volume\") pod \"collect-profiles-29423640-42vd6\" (UID: \"4c22bf31-ad33-44cd-a495-05d83688f2bd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423640-42vd6" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.910575 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f5149a23-92a4-42e6-8b20-8411d1d0536f-serving-cert\") pod \"openshift-config-operator-7777fb866f-w6b4l\" (UID: \"f5149a23-92a4-42e6-8b20-8411d1d0536f\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-w6b4l" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.910591 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/52938126-55df-4106-9bed-ad7cc0f1130d-certs\") pod \"machine-config-server-6b6br\" (UID: \"52938126-55df-4106-9bed-ad7cc0f1130d\") " pod="openshift-machine-config-operator/machine-config-server-6b6br" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.910612 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-whkfd\" (UniqueName: \"kubernetes.io/projected/335b007e-ea5c-4e20-9402-5599da94ef22-kube-api-access-whkfd\") pod \"csi-hostpathplugin-hr4qs\" (UID: \"335b007e-ea5c-4e20-9402-5599da94ef22\") " pod="hostpath-provisioner/csi-hostpathplugin-hr4qs" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.910631 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/69814666-b43a-4cab-9864-864862277916-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-lpgf5\" (UID: \"69814666-b43a-4cab-9864-864862277916\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-lpgf5" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.910649 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kq4zx\" (UniqueName: \"kubernetes.io/projected/ffb4acd4-543f-4509-b993-249e3656048d-kube-api-access-kq4zx\") pod \"marketplace-operator-79b997595-z42g5\" (UID: \"ffb4acd4-543f-4509-b993-249e3656048d\") " pod="openshift-marketplace/marketplace-operator-79b997595-z42g5" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.910672 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wtcp6\" (UniqueName: \"kubernetes.io/projected/1c402f2f-f79a-4a1a-bf40-07da4223059e-kube-api-access-wtcp6\") pod \"multus-admission-controller-857f4d67dd-qp2jl\" (UID: \"1c402f2f-f79a-4a1a-bf40-07da4223059e\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-qp2jl" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.910690 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kgdtw\" (UniqueName: \"kubernetes.io/projected/c70805eb-e095-4462-a202-e01860559471-kube-api-access-kgdtw\") pod \"control-plane-machine-set-operator-78cbb6b69f-gp8dt\" (UID: \"c70805eb-e095-4462-a202-e01860559471\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-gp8dt" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.912364 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/18d74f73-0f04-4f3c-88c4-a431d490db9d-registry-certificates\") pod \"image-registry-697d97f7c8-254w2\" (UID: \"18d74f73-0f04-4f3c-88c4-a431d490db9d\") " pod="openshift-image-registry/image-registry-697d97f7c8-254w2" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.913314 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/18d74f73-0f04-4f3c-88c4-a431d490db9d-trusted-ca\") pod \"image-registry-697d97f7c8-254w2\" (UID: \"18d74f73-0f04-4f3c-88c4-a431d490db9d\") " pod="openshift-image-registry/image-registry-697d97f7c8-254w2" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.917296 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dbdb9d15-7786-4fef-a38e-e3be94f1bcb2-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-tz2j9\" (UID: \"dbdb9d15-7786-4fef-a38e-e3be94f1bcb2\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-tz2j9" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.917423 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f5149a23-92a4-42e6-8b20-8411d1d0536f-serving-cert\") pod \"openshift-config-operator-7777fb866f-w6b4l\" (UID: \"f5149a23-92a4-42e6-8b20-8411d1d0536f\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-w6b4l" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.917957 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/18d74f73-0f04-4f3c-88c4-a431d490db9d-installation-pull-secrets\") pod \"image-registry-697d97f7c8-254w2\" (UID: \"18d74f73-0f04-4f3c-88c4-a431d490db9d\") " pod="openshift-image-registry/image-registry-697d97f7c8-254w2" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.920523 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/18d74f73-0f04-4f3c-88c4-a431d490db9d-registry-tls\") pod \"image-registry-697d97f7c8-254w2\" (UID: \"18d74f73-0f04-4f3c-88c4-a431d490db9d\") " pod="openshift-image-registry/image-registry-697d97f7c8-254w2" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.920771 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-l88b4" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.928412 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-d89dq" Dec 11 02:03:15 crc kubenswrapper[4824]: W1211 02:03:15.948844 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9e099899_aaf7_4bd8_b37e_8616b5060e33.slice/crio-c3b9fdc4b65c7a93573ec7c1b416fa6a98dd8adb41cd979ae09ab5dded0ce475 WatchSource:0}: Error finding container c3b9fdc4b65c7a93573ec7c1b416fa6a98dd8adb41cd979ae09ab5dded0ce475: Status 404 returned error can't find the container with id c3b9fdc4b65c7a93573ec7c1b416fa6a98dd8adb41cd979ae09ab5dded0ce475 Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.953045 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4d6h\" (UniqueName: \"kubernetes.io/projected/18d74f73-0f04-4f3c-88c4-a431d490db9d-kube-api-access-z4d6h\") pod \"image-registry-697d97f7c8-254w2\" (UID: \"18d74f73-0f04-4f3c-88c4-a431d490db9d\") " pod="openshift-image-registry/image-registry-697d97f7c8-254w2" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.970287 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zm5nf\" (UniqueName: \"kubernetes.io/projected/f5149a23-92a4-42e6-8b20-8411d1d0536f-kube-api-access-zm5nf\") pod \"openshift-config-operator-7777fb866f-w6b4l\" (UID: \"f5149a23-92a4-42e6-8b20-8411d1d0536f\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-w6b4l" Dec 11 02:03:15 crc kubenswrapper[4824]: I1211 02:03:15.989677 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/18d74f73-0f04-4f3c-88c4-a431d490db9d-bound-sa-token\") pod \"image-registry-697d97f7c8-254w2\" (UID: \"18d74f73-0f04-4f3c-88c4-a431d490db9d\") " pod="openshift-image-registry/image-registry-697d97f7c8-254w2" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.013913 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/09f964e8-942f-4577-983a-bdc08487c0e8-signing-key\") pod \"service-ca-9c57cc56f-4r8x8\" (UID: \"09f964e8-942f-4577-983a-bdc08487c0e8\") " pod="openshift-service-ca/service-ca-9c57cc56f-4r8x8" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.013961 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a4f0d12d-4a03-4ffd-9058-f311bf0041fe-config-volume\") pod \"dns-default-qtll9\" (UID: \"a4f0d12d-4a03-4ffd-9058-f311bf0041fe\") " pod="openshift-dns/dns-default-qtll9" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.013986 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09112ee7-18d2-4a6d-83ce-eddfce56b083-config\") pod \"service-ca-operator-777779d784-59bgf\" (UID: \"09112ee7-18d2-4a6d-83ce-eddfce56b083\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-59bgf" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.014014 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cwz4n\" (UniqueName: \"kubernetes.io/projected/877d4a7f-46b0-4f9c-a702-b6b83620c3cb-kube-api-access-cwz4n\") pod \"router-default-5444994796-pg8rn\" (UID: \"877d4a7f-46b0-4f9c-a702-b6b83620c3cb\") " pod="openshift-ingress/router-default-5444994796-pg8rn" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.014039 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/59dd9dac-e58a-4869-b1ee-e0c5e8a78974-webhook-cert\") pod \"packageserver-d55dfcdfc-p8x2v\" (UID: \"59dd9dac-e58a-4869-b1ee-e0c5e8a78974\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p8x2v" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.014060 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/09f964e8-942f-4577-983a-bdc08487c0e8-signing-cabundle\") pod \"service-ca-9c57cc56f-4r8x8\" (UID: \"09f964e8-942f-4577-983a-bdc08487c0e8\") " pod="openshift-service-ca/service-ca-9c57cc56f-4r8x8" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.014077 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-njgbl\" (UniqueName: \"kubernetes.io/projected/63fe488c-de20-435f-83fd-47a4179341e6-kube-api-access-njgbl\") pod \"ingress-canary-zvcj8\" (UID: \"63fe488c-de20-435f-83fd-47a4179341e6\") " pod="openshift-ingress-canary/ingress-canary-zvcj8" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.014103 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/04cb3d8f-ccf2-410c-b51c-445add07699c-images\") pod \"machine-config-operator-74547568cd-jlqk5\" (UID: \"04cb3d8f-ccf2-410c-b51c-445add07699c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-jlqk5" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.014136 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/59dd9dac-e58a-4869-b1ee-e0c5e8a78974-apiservice-cert\") pod \"packageserver-d55dfcdfc-p8x2v\" (UID: \"59dd9dac-e58a-4869-b1ee-e0c5e8a78974\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p8x2v" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.014150 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ffb4acd4-543f-4509-b993-249e3656048d-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-z42g5\" (UID: \"ffb4acd4-543f-4509-b993-249e3656048d\") " pod="openshift-marketplace/marketplace-operator-79b997595-z42g5" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.014177 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/04cb3d8f-ccf2-410c-b51c-445add07699c-auth-proxy-config\") pod \"machine-config-operator-74547568cd-jlqk5\" (UID: \"04cb3d8f-ccf2-410c-b51c-445add07699c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-jlqk5" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.014194 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4c22bf31-ad33-44cd-a495-05d83688f2bd-config-volume\") pod \"collect-profiles-29423640-42vd6\" (UID: \"4c22bf31-ad33-44cd-a495-05d83688f2bd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423640-42vd6" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.014212 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a4f0d12d-4a03-4ffd-9058-f311bf0041fe-metrics-tls\") pod \"dns-default-qtll9\" (UID: \"a4f0d12d-4a03-4ffd-9058-f311bf0041fe\") " pod="openshift-dns/dns-default-qtll9" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.014230 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bbwzb\" (UniqueName: \"kubernetes.io/projected/a4f0d12d-4a03-4ffd-9058-f311bf0041fe-kube-api-access-bbwzb\") pod \"dns-default-qtll9\" (UID: \"a4f0d12d-4a03-4ffd-9058-f311bf0041fe\") " pod="openshift-dns/dns-default-qtll9" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.014245 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/877d4a7f-46b0-4f9c-a702-b6b83620c3cb-service-ca-bundle\") pod \"router-default-5444994796-pg8rn\" (UID: \"877d4a7f-46b0-4f9c-a702-b6b83620c3cb\") " pod="openshift-ingress/router-default-5444994796-pg8rn" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.014263 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/398ffffe-110f-4f78-a784-2ee501ee58a3-srv-cert\") pod \"olm-operator-6b444d44fb-66mv8\" (UID: \"398ffffe-110f-4f78-a784-2ee501ee58a3\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-66mv8" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.014277 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/63fe488c-de20-435f-83fd-47a4179341e6-cert\") pod \"ingress-canary-zvcj8\" (UID: \"63fe488c-de20-435f-83fd-47a4179341e6\") " pod="openshift-ingress-canary/ingress-canary-zvcj8" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.014292 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sgqgs\" (UniqueName: \"kubernetes.io/projected/95b90361-2f73-4649-bc1a-1202ba511baf-kube-api-access-sgqgs\") pod \"migrator-59844c95c7-8fksd\" (UID: \"95b90361-2f73-4649-bc1a-1202ba511baf\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8fksd" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.014315 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hlmx4\" (UniqueName: \"kubernetes.io/projected/59dd9dac-e58a-4869-b1ee-e0c5e8a78974-kube-api-access-hlmx4\") pod \"packageserver-d55dfcdfc-p8x2v\" (UID: \"59dd9dac-e58a-4869-b1ee-e0c5e8a78974\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p8x2v" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.014332 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/59dd9dac-e58a-4869-b1ee-e0c5e8a78974-tmpfs\") pod \"packageserver-d55dfcdfc-p8x2v\" (UID: \"59dd9dac-e58a-4869-b1ee-e0c5e8a78974\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p8x2v" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.014348 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-675ss\" (UniqueName: \"kubernetes.io/projected/52938126-55df-4106-9bed-ad7cc0f1130d-kube-api-access-675ss\") pod \"machine-config-server-6b6br\" (UID: \"52938126-55df-4106-9bed-ad7cc0f1130d\") " pod="openshift-machine-config-operator/machine-config-server-6b6br" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.014365 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/c70805eb-e095-4462-a202-e01860559471-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-gp8dt\" (UID: \"c70805eb-e095-4462-a202-e01860559471\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-gp8dt" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.014383 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/505f6458-5d5a-4725-a529-1075770afa95-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-vkdbw\" (UID: \"505f6458-5d5a-4725-a529-1075770afa95\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vkdbw" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.014401 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dqzmb\" (UniqueName: \"kubernetes.io/projected/505f6458-5d5a-4725-a529-1075770afa95-kube-api-access-dqzmb\") pod \"machine-config-controller-84d6567774-vkdbw\" (UID: \"505f6458-5d5a-4725-a529-1075770afa95\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vkdbw" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.014418 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/335b007e-ea5c-4e20-9402-5599da94ef22-csi-data-dir\") pod \"csi-hostpathplugin-hr4qs\" (UID: \"335b007e-ea5c-4e20-9402-5599da94ef22\") " pod="hostpath-provisioner/csi-hostpathplugin-hr4qs" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.014435 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4h2lj\" (UniqueName: \"kubernetes.io/projected/69814666-b43a-4cab-9864-864862277916-kube-api-access-4h2lj\") pod \"package-server-manager-789f6589d5-lpgf5\" (UID: \"69814666-b43a-4cab-9864-864862277916\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-lpgf5" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.014451 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wt25m\" (UniqueName: \"kubernetes.io/projected/04cb3d8f-ccf2-410c-b51c-445add07699c-kube-api-access-wt25m\") pod \"machine-config-operator-74547568cd-jlqk5\" (UID: \"04cb3d8f-ccf2-410c-b51c-445add07699c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-jlqk5" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.014466 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/877d4a7f-46b0-4f9c-a702-b6b83620c3cb-default-certificate\") pod \"router-default-5444994796-pg8rn\" (UID: \"877d4a7f-46b0-4f9c-a702-b6b83620c3cb\") " pod="openshift-ingress/router-default-5444994796-pg8rn" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.014487 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-drmvw\" (UniqueName: \"kubernetes.io/projected/4c22bf31-ad33-44cd-a495-05d83688f2bd-kube-api-access-drmvw\") pod \"collect-profiles-29423640-42vd6\" (UID: \"4c22bf31-ad33-44cd-a495-05d83688f2bd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423640-42vd6" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.014502 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09112ee7-18d2-4a6d-83ce-eddfce56b083-serving-cert\") pod \"service-ca-operator-777779d784-59bgf\" (UID: \"09112ee7-18d2-4a6d-83ce-eddfce56b083\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-59bgf" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.014519 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/398ffffe-110f-4f78-a784-2ee501ee58a3-profile-collector-cert\") pod \"olm-operator-6b444d44fb-66mv8\" (UID: \"398ffffe-110f-4f78-a784-2ee501ee58a3\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-66mv8" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.014539 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c6vtj\" (UniqueName: \"kubernetes.io/projected/09f964e8-942f-4577-983a-bdc08487c0e8-kube-api-access-c6vtj\") pod \"service-ca-9c57cc56f-4r8x8\" (UID: \"09f964e8-942f-4577-983a-bdc08487c0e8\") " pod="openshift-service-ca/service-ca-9c57cc56f-4r8x8" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.014561 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hfh5h\" (UniqueName: \"kubernetes.io/projected/398ffffe-110f-4f78-a784-2ee501ee58a3-kube-api-access-hfh5h\") pod \"olm-operator-6b444d44fb-66mv8\" (UID: \"398ffffe-110f-4f78-a784-2ee501ee58a3\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-66mv8" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.014593 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/52938126-55df-4106-9bed-ad7cc0f1130d-certs\") pod \"machine-config-server-6b6br\" (UID: \"52938126-55df-4106-9bed-ad7cc0f1130d\") " pod="openshift-machine-config-operator/machine-config-server-6b6br" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.014610 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/877d4a7f-46b0-4f9c-a702-b6b83620c3cb-stats-auth\") pod \"router-default-5444994796-pg8rn\" (UID: \"877d4a7f-46b0-4f9c-a702-b6b83620c3cb\") " pod="openshift-ingress/router-default-5444994796-pg8rn" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.014627 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4c22bf31-ad33-44cd-a495-05d83688f2bd-secret-volume\") pod \"collect-profiles-29423640-42vd6\" (UID: \"4c22bf31-ad33-44cd-a495-05d83688f2bd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423640-42vd6" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.014646 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/69814666-b43a-4cab-9864-864862277916-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-lpgf5\" (UID: \"69814666-b43a-4cab-9864-864862277916\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-lpgf5" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.014669 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kq4zx\" (UniqueName: \"kubernetes.io/projected/ffb4acd4-543f-4509-b993-249e3656048d-kube-api-access-kq4zx\") pod \"marketplace-operator-79b997595-z42g5\" (UID: \"ffb4acd4-543f-4509-b993-249e3656048d\") " pod="openshift-marketplace/marketplace-operator-79b997595-z42g5" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.014687 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-whkfd\" (UniqueName: \"kubernetes.io/projected/335b007e-ea5c-4e20-9402-5599da94ef22-kube-api-access-whkfd\") pod \"csi-hostpathplugin-hr4qs\" (UID: \"335b007e-ea5c-4e20-9402-5599da94ef22\") " pod="hostpath-provisioner/csi-hostpathplugin-hr4qs" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.014704 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wtcp6\" (UniqueName: \"kubernetes.io/projected/1c402f2f-f79a-4a1a-bf40-07da4223059e-kube-api-access-wtcp6\") pod \"multus-admission-controller-857f4d67dd-qp2jl\" (UID: \"1c402f2f-f79a-4a1a-bf40-07da4223059e\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-qp2jl" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.014719 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kgdtw\" (UniqueName: \"kubernetes.io/projected/c70805eb-e095-4462-a202-e01860559471-kube-api-access-kgdtw\") pod \"control-plane-machine-set-operator-78cbb6b69f-gp8dt\" (UID: \"c70805eb-e095-4462-a202-e01860559471\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-gp8dt" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.014737 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/335b007e-ea5c-4e20-9402-5599da94ef22-plugins-dir\") pod \"csi-hostpathplugin-hr4qs\" (UID: \"335b007e-ea5c-4e20-9402-5599da94ef22\") " pod="hostpath-provisioner/csi-hostpathplugin-hr4qs" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.014756 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4j7mj\" (UniqueName: \"kubernetes.io/projected/56378af0-da53-4ffc-a66c-a381083e1ba6-kube-api-access-4j7mj\") pod \"catalog-operator-68c6474976-qqg25\" (UID: \"56378af0-da53-4ffc-a66c-a381083e1ba6\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qqg25" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.014770 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kpq4n\" (UniqueName: \"kubernetes.io/projected/09112ee7-18d2-4a6d-83ce-eddfce56b083-kube-api-access-kpq4n\") pod \"service-ca-operator-777779d784-59bgf\" (UID: \"09112ee7-18d2-4a6d-83ce-eddfce56b083\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-59bgf" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.014785 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dccdfcce-6b35-40e6-8bdc-7eed59518f24-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-kdzkc\" (UID: \"dccdfcce-6b35-40e6-8bdc-7eed59518f24\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kdzkc" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.014798 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/56378af0-da53-4ffc-a66c-a381083e1ba6-srv-cert\") pod \"catalog-operator-68c6474976-qqg25\" (UID: \"56378af0-da53-4ffc-a66c-a381083e1ba6\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qqg25" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.014812 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/335b007e-ea5c-4e20-9402-5599da94ef22-mountpoint-dir\") pod \"csi-hostpathplugin-hr4qs\" (UID: \"335b007e-ea5c-4e20-9402-5599da94ef22\") " pod="hostpath-provisioner/csi-hostpathplugin-hr4qs" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.014827 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/1c402f2f-f79a-4a1a-bf40-07da4223059e-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-qp2jl\" (UID: \"1c402f2f-f79a-4a1a-bf40-07da4223059e\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-qp2jl" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.014843 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/52938126-55df-4106-9bed-ad7cc0f1130d-node-bootstrap-token\") pod \"machine-config-server-6b6br\" (UID: \"52938126-55df-4106-9bed-ad7cc0f1130d\") " pod="openshift-machine-config-operator/machine-config-server-6b6br" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.014858 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/505f6458-5d5a-4725-a529-1075770afa95-proxy-tls\") pod \"machine-config-controller-84d6567774-vkdbw\" (UID: \"505f6458-5d5a-4725-a529-1075770afa95\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vkdbw" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.014882 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/04cb3d8f-ccf2-410c-b51c-445add07699c-proxy-tls\") pod \"machine-config-operator-74547568cd-jlqk5\" (UID: \"04cb3d8f-ccf2-410c-b51c-445add07699c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-jlqk5" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.014895 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/335b007e-ea5c-4e20-9402-5599da94ef22-registration-dir\") pod \"csi-hostpathplugin-hr4qs\" (UID: \"335b007e-ea5c-4e20-9402-5599da94ef22\") " pod="hostpath-provisioner/csi-hostpathplugin-hr4qs" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.014909 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dccdfcce-6b35-40e6-8bdc-7eed59518f24-config\") pod \"kube-controller-manager-operator-78b949d7b-kdzkc\" (UID: \"dccdfcce-6b35-40e6-8bdc-7eed59518f24\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kdzkc" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.014922 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dccdfcce-6b35-40e6-8bdc-7eed59518f24-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-kdzkc\" (UID: \"dccdfcce-6b35-40e6-8bdc-7eed59518f24\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kdzkc" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.014938 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/877d4a7f-46b0-4f9c-a702-b6b83620c3cb-metrics-certs\") pod \"router-default-5444994796-pg8rn\" (UID: \"877d4a7f-46b0-4f9c-a702-b6b83620c3cb\") " pod="openshift-ingress/router-default-5444994796-pg8rn" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.014957 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ffb4acd4-543f-4509-b993-249e3656048d-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-z42g5\" (UID: \"ffb4acd4-543f-4509-b993-249e3656048d\") " pod="openshift-marketplace/marketplace-operator-79b997595-z42g5" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.014978 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-254w2\" (UID: \"18d74f73-0f04-4f3c-88c4-a431d490db9d\") " pod="openshift-image-registry/image-registry-697d97f7c8-254w2" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.014993 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/335b007e-ea5c-4e20-9402-5599da94ef22-socket-dir\") pod \"csi-hostpathplugin-hr4qs\" (UID: \"335b007e-ea5c-4e20-9402-5599da94ef22\") " pod="hostpath-provisioner/csi-hostpathplugin-hr4qs" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.015009 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/56378af0-da53-4ffc-a66c-a381083e1ba6-profile-collector-cert\") pod \"catalog-operator-68c6474976-qqg25\" (UID: \"56378af0-da53-4ffc-a66c-a381083e1ba6\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qqg25" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.015897 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qwjc7\" (UniqueName: \"kubernetes.io/projected/dbdb9d15-7786-4fef-a38e-e3be94f1bcb2-kube-api-access-qwjc7\") pod \"openshift-controller-manager-operator-756b6f6bc6-tz2j9\" (UID: \"dbdb9d15-7786-4fef-a38e-e3be94f1bcb2\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-tz2j9" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.016566 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/877d4a7f-46b0-4f9c-a702-b6b83620c3cb-service-ca-bundle\") pod \"router-default-5444994796-pg8rn\" (UID: \"877d4a7f-46b0-4f9c-a702-b6b83620c3cb\") " pod="openshift-ingress/router-default-5444994796-pg8rn" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.017081 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a4f0d12d-4a03-4ffd-9058-f311bf0041fe-config-volume\") pod \"dns-default-qtll9\" (UID: \"a4f0d12d-4a03-4ffd-9058-f311bf0041fe\") " pod="openshift-dns/dns-default-qtll9" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.017987 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/335b007e-ea5c-4e20-9402-5599da94ef22-registration-dir\") pod \"csi-hostpathplugin-hr4qs\" (UID: \"335b007e-ea5c-4e20-9402-5599da94ef22\") " pod="hostpath-provisioner/csi-hostpathplugin-hr4qs" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.018083 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09112ee7-18d2-4a6d-83ce-eddfce56b083-config\") pod \"service-ca-operator-777779d784-59bgf\" (UID: \"09112ee7-18d2-4a6d-83ce-eddfce56b083\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-59bgf" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.018471 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/09f964e8-942f-4577-983a-bdc08487c0e8-signing-key\") pod \"service-ca-9c57cc56f-4r8x8\" (UID: \"09f964e8-942f-4577-983a-bdc08487c0e8\") " pod="openshift-service-ca/service-ca-9c57cc56f-4r8x8" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.018510 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dccdfcce-6b35-40e6-8bdc-7eed59518f24-config\") pod \"kube-controller-manager-operator-78b949d7b-kdzkc\" (UID: \"dccdfcce-6b35-40e6-8bdc-7eed59518f24\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kdzkc" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.019506 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/505f6458-5d5a-4725-a529-1075770afa95-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-vkdbw\" (UID: \"505f6458-5d5a-4725-a529-1075770afa95\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vkdbw" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.019722 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/335b007e-ea5c-4e20-9402-5599da94ef22-csi-data-dir\") pod \"csi-hostpathplugin-hr4qs\" (UID: \"335b007e-ea5c-4e20-9402-5599da94ef22\") " pod="hostpath-provisioner/csi-hostpathplugin-hr4qs" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.021367 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/59dd9dac-e58a-4869-b1ee-e0c5e8a78974-apiservice-cert\") pod \"packageserver-d55dfcdfc-p8x2v\" (UID: \"59dd9dac-e58a-4869-b1ee-e0c5e8a78974\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p8x2v" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.021471 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/877d4a7f-46b0-4f9c-a702-b6b83620c3cb-metrics-certs\") pod \"router-default-5444994796-pg8rn\" (UID: \"877d4a7f-46b0-4f9c-a702-b6b83620c3cb\") " pod="openshift-ingress/router-default-5444994796-pg8rn" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.021944 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/398ffffe-110f-4f78-a784-2ee501ee58a3-srv-cert\") pod \"olm-operator-6b444d44fb-66mv8\" (UID: \"398ffffe-110f-4f78-a784-2ee501ee58a3\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-66mv8" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.022100 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/335b007e-ea5c-4e20-9402-5599da94ef22-mountpoint-dir\") pod \"csi-hostpathplugin-hr4qs\" (UID: \"335b007e-ea5c-4e20-9402-5599da94ef22\") " pod="hostpath-provisioner/csi-hostpathplugin-hr4qs" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.022280 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/59dd9dac-e58a-4869-b1ee-e0c5e8a78974-tmpfs\") pod \"packageserver-d55dfcdfc-p8x2v\" (UID: \"59dd9dac-e58a-4869-b1ee-e0c5e8a78974\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p8x2v" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.022342 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/335b007e-ea5c-4e20-9402-5599da94ef22-socket-dir\") pod \"csi-hostpathplugin-hr4qs\" (UID: \"335b007e-ea5c-4e20-9402-5599da94ef22\") " pod="hostpath-provisioner/csi-hostpathplugin-hr4qs" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.022354 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/335b007e-ea5c-4e20-9402-5599da94ef22-plugins-dir\") pod \"csi-hostpathplugin-hr4qs\" (UID: \"335b007e-ea5c-4e20-9402-5599da94ef22\") " pod="hostpath-provisioner/csi-hostpathplugin-hr4qs" Dec 11 02:03:16 crc kubenswrapper[4824]: E1211 02:03:16.022514 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 02:03:16.522496942 +0000 UTC m=+138.211534321 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-254w2" (UID: "18d74f73-0f04-4f3c-88c4-a431d490db9d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.022803 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4c22bf31-ad33-44cd-a495-05d83688f2bd-config-volume\") pod \"collect-profiles-29423640-42vd6\" (UID: \"4c22bf31-ad33-44cd-a495-05d83688f2bd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423640-42vd6" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.023232 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/04cb3d8f-ccf2-410c-b51c-445add07699c-proxy-tls\") pod \"machine-config-operator-74547568cd-jlqk5\" (UID: \"04cb3d8f-ccf2-410c-b51c-445add07699c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-jlqk5" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.023268 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/877d4a7f-46b0-4f9c-a702-b6b83620c3cb-default-certificate\") pod \"router-default-5444994796-pg8rn\" (UID: \"877d4a7f-46b0-4f9c-a702-b6b83620c3cb\") " pod="openshift-ingress/router-default-5444994796-pg8rn" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.025180 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/59dd9dac-e58a-4869-b1ee-e0c5e8a78974-webhook-cert\") pod \"packageserver-d55dfcdfc-p8x2v\" (UID: \"59dd9dac-e58a-4869-b1ee-e0c5e8a78974\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p8x2v" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.025686 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ffb4acd4-543f-4509-b993-249e3656048d-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-z42g5\" (UID: \"ffb4acd4-543f-4509-b993-249e3656048d\") " pod="openshift-marketplace/marketplace-operator-79b997595-z42g5" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.026588 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/505f6458-5d5a-4725-a529-1075770afa95-proxy-tls\") pod \"machine-config-controller-84d6567774-vkdbw\" (UID: \"505f6458-5d5a-4725-a529-1075770afa95\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vkdbw" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.026861 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/52938126-55df-4106-9bed-ad7cc0f1130d-node-bootstrap-token\") pod \"machine-config-server-6b6br\" (UID: \"52938126-55df-4106-9bed-ad7cc0f1130d\") " pod="openshift-machine-config-operator/machine-config-server-6b6br" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.026925 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/56378af0-da53-4ffc-a66c-a381083e1ba6-profile-collector-cert\") pod \"catalog-operator-68c6474976-qqg25\" (UID: \"56378af0-da53-4ffc-a66c-a381083e1ba6\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qqg25" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.027048 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a4f0d12d-4a03-4ffd-9058-f311bf0041fe-metrics-tls\") pod \"dns-default-qtll9\" (UID: \"a4f0d12d-4a03-4ffd-9058-f311bf0041fe\") " pod="openshift-dns/dns-default-qtll9" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.027959 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ffb4acd4-543f-4509-b993-249e3656048d-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-z42g5\" (UID: \"ffb4acd4-543f-4509-b993-249e3656048d\") " pod="openshift-marketplace/marketplace-operator-79b997595-z42g5" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.028482 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/63fe488c-de20-435f-83fd-47a4179341e6-cert\") pod \"ingress-canary-zvcj8\" (UID: \"63fe488c-de20-435f-83fd-47a4179341e6\") " pod="openshift-ingress-canary/ingress-canary-zvcj8" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.029050 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/398ffffe-110f-4f78-a784-2ee501ee58a3-profile-collector-cert\") pod \"olm-operator-6b444d44fb-66mv8\" (UID: \"398ffffe-110f-4f78-a784-2ee501ee58a3\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-66mv8" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.029088 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/69814666-b43a-4cab-9864-864862277916-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-lpgf5\" (UID: \"69814666-b43a-4cab-9864-864862277916\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-lpgf5" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.029379 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/52938126-55df-4106-9bed-ad7cc0f1130d-certs\") pod \"machine-config-server-6b6br\" (UID: \"52938126-55df-4106-9bed-ad7cc0f1130d\") " pod="openshift-machine-config-operator/machine-config-server-6b6br" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.030324 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/56378af0-da53-4ffc-a66c-a381083e1ba6-srv-cert\") pod \"catalog-operator-68c6474976-qqg25\" (UID: \"56378af0-da53-4ffc-a66c-a381083e1ba6\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qqg25" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.030691 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4c22bf31-ad33-44cd-a495-05d83688f2bd-secret-volume\") pod \"collect-profiles-29423640-42vd6\" (UID: \"4c22bf31-ad33-44cd-a495-05d83688f2bd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423640-42vd6" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.030697 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/1c402f2f-f79a-4a1a-bf40-07da4223059e-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-qp2jl\" (UID: \"1c402f2f-f79a-4a1a-bf40-07da4223059e\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-qp2jl" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.031534 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/04cb3d8f-ccf2-410c-b51c-445add07699c-auth-proxy-config\") pod \"machine-config-operator-74547568cd-jlqk5\" (UID: \"04cb3d8f-ccf2-410c-b51c-445add07699c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-jlqk5" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.032586 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09112ee7-18d2-4a6d-83ce-eddfce56b083-serving-cert\") pod \"service-ca-operator-777779d784-59bgf\" (UID: \"09112ee7-18d2-4a6d-83ce-eddfce56b083\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-59bgf" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.033213 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/c70805eb-e095-4462-a202-e01860559471-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-gp8dt\" (UID: \"c70805eb-e095-4462-a202-e01860559471\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-gp8dt" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.035982 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/09f964e8-942f-4577-983a-bdc08487c0e8-signing-cabundle\") pod \"service-ca-9c57cc56f-4r8x8\" (UID: \"09f964e8-942f-4577-983a-bdc08487c0e8\") " pod="openshift-service-ca/service-ca-9c57cc56f-4r8x8" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.054601 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-drmvw\" (UniqueName: \"kubernetes.io/projected/4c22bf31-ad33-44cd-a495-05d83688f2bd-kube-api-access-drmvw\") pod \"collect-profiles-29423640-42vd6\" (UID: \"4c22bf31-ad33-44cd-a495-05d83688f2bd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423640-42vd6" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.070442 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sgqgs\" (UniqueName: \"kubernetes.io/projected/95b90361-2f73-4649-bc1a-1202ba511baf-kube-api-access-sgqgs\") pod \"migrator-59844c95c7-8fksd\" (UID: \"95b90361-2f73-4649-bc1a-1202ba511baf\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8fksd" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.078156 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423640-42vd6" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.086633 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kpq4n\" (UniqueName: \"kubernetes.io/projected/09112ee7-18d2-4a6d-83ce-eddfce56b083-kube-api-access-kpq4n\") pod \"service-ca-operator-777779d784-59bgf\" (UID: \"09112ee7-18d2-4a6d-83ce-eddfce56b083\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-59bgf" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.110881 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dccdfcce-6b35-40e6-8bdc-7eed59518f24-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-kdzkc\" (UID: \"dccdfcce-6b35-40e6-8bdc-7eed59518f24\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kdzkc" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.115647 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.116372 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8fksd" Dec 11 02:03:16 crc kubenswrapper[4824]: E1211 02:03:16.116514 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 02:03:16.616494775 +0000 UTC m=+138.305532154 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.130309 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-w6b4l" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.137318 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dqzmb\" (UniqueName: \"kubernetes.io/projected/505f6458-5d5a-4725-a529-1075770afa95-kube-api-access-dqzmb\") pod \"machine-config-controller-84d6567774-vkdbw\" (UID: \"505f6458-5d5a-4725-a529-1075770afa95\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vkdbw" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.144836 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-tfwnz"] Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.148285 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-tz2j9" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.166068 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4h2lj\" (UniqueName: \"kubernetes.io/projected/69814666-b43a-4cab-9864-864862277916-kube-api-access-4h2lj\") pod \"package-server-manager-789f6589d5-lpgf5\" (UID: \"69814666-b43a-4cab-9864-864862277916\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-lpgf5" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.172301 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2vcc6"] Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.174675 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wt25m\" (UniqueName: \"kubernetes.io/projected/04cb3d8f-ccf2-410c-b51c-445add07699c-kube-api-access-wt25m\") pod \"machine-config-operator-74547568cd-jlqk5\" (UID: \"04cb3d8f-ccf2-410c-b51c-445add07699c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-jlqk5" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.194705 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cwz4n\" (UniqueName: \"kubernetes.io/projected/877d4a7f-46b0-4f9c-a702-b6b83620c3cb-kube-api-access-cwz4n\") pod \"router-default-5444994796-pg8rn\" (UID: \"877d4a7f-46b0-4f9c-a702-b6b83620c3cb\") " pod="openshift-ingress/router-default-5444994796-pg8rn" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.218018 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-254w2\" (UID: \"18d74f73-0f04-4f3c-88c4-a431d490db9d\") " pod="openshift-image-registry/image-registry-697d97f7c8-254w2" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.218389 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-2ng4n"] Dec 11 02:03:16 crc kubenswrapper[4824]: E1211 02:03:16.218449 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 02:03:16.718415645 +0000 UTC m=+138.407453024 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-254w2" (UID: "18d74f73-0f04-4f3c-88c4-a431d490db9d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.221022 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gndwh"] Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.230338 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-675ss\" (UniqueName: \"kubernetes.io/projected/52938126-55df-4106-9bed-ad7cc0f1130d-kube-api-access-675ss\") pod \"machine-config-server-6b6br\" (UID: \"52938126-55df-4106-9bed-ad7cc0f1130d\") " pod="openshift-machine-config-operator/machine-config-server-6b6br" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.241690 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-l9d4f"] Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.251604 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4j7mj\" (UniqueName: \"kubernetes.io/projected/56378af0-da53-4ffc-a66c-a381083e1ba6-kube-api-access-4j7mj\") pod \"catalog-operator-68c6474976-qqg25\" (UID: \"56378af0-da53-4ffc-a66c-a381083e1ba6\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qqg25" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.260139 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-hxqw8"] Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.265084 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2zq2n"] Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.271069 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-r89c2"] Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.277368 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bbwzb\" (UniqueName: \"kubernetes.io/projected/a4f0d12d-4a03-4ffd-9058-f311bf0041fe-kube-api-access-bbwzb\") pod \"dns-default-qtll9\" (UID: \"a4f0d12d-4a03-4ffd-9058-f311bf0041fe\") " pod="openshift-dns/dns-default-qtll9" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.289414 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kq4zx\" (UniqueName: \"kubernetes.io/projected/ffb4acd4-543f-4509-b993-249e3656048d-kube-api-access-kq4zx\") pod \"marketplace-operator-79b997595-z42g5\" (UID: \"ffb4acd4-543f-4509-b993-249e3656048d\") " pod="openshift-marketplace/marketplace-operator-79b997595-z42g5" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.294559 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/04cb3d8f-ccf2-410c-b51c-445add07699c-images\") pod \"machine-config-operator-74547568cd-jlqk5\" (UID: \"04cb3d8f-ccf2-410c-b51c-445add07699c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-jlqk5" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.296594 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lg9s6"] Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.297405 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dccdfcce-6b35-40e6-8bdc-7eed59518f24-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-kdzkc\" (UID: \"dccdfcce-6b35-40e6-8bdc-7eed59518f24\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kdzkc" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.304932 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kdzkc" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.308228 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/877d4a7f-46b0-4f9c-a702-b6b83620c3cb-stats-auth\") pod \"router-default-5444994796-pg8rn\" (UID: \"877d4a7f-46b0-4f9c-a702-b6b83620c3cb\") " pod="openshift-ingress/router-default-5444994796-pg8rn" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.315292 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hlmx4\" (UniqueName: \"kubernetes.io/projected/59dd9dac-e58a-4869-b1ee-e0c5e8a78974-kube-api-access-hlmx4\") pod \"packageserver-d55dfcdfc-p8x2v\" (UID: \"59dd9dac-e58a-4869-b1ee-e0c5e8a78974\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p8x2v" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.318498 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-jlqk5" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.318611 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 02:03:16 crc kubenswrapper[4824]: E1211 02:03:16.318930 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 02:03:16.818913183 +0000 UTC m=+138.507950652 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.319074 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-254w2\" (UID: \"18d74f73-0f04-4f3c-88c4-a431d490db9d\") " pod="openshift-image-registry/image-registry-697d97f7c8-254w2" Dec 11 02:03:16 crc kubenswrapper[4824]: E1211 02:03:16.319468 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 02:03:16.819460079 +0000 UTC m=+138.508497458 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-254w2" (UID: "18d74f73-0f04-4f3c-88c4-a431d490db9d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.325262 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-z42g5" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.333456 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-pg8rn" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.333902 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-whkfd\" (UniqueName: \"kubernetes.io/projected/335b007e-ea5c-4e20-9402-5599da94ef22-kube-api-access-whkfd\") pod \"csi-hostpathplugin-hr4qs\" (UID: \"335b007e-ea5c-4e20-9402-5599da94ef22\") " pod="hostpath-provisioner/csi-hostpathplugin-hr4qs" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.342667 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-l88b4"] Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.348579 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kgdtw\" (UniqueName: \"kubernetes.io/projected/c70805eb-e095-4462-a202-e01860559471-kube-api-access-kgdtw\") pod \"control-plane-machine-set-operator-78cbb6b69f-gp8dt\" (UID: \"c70805eb-e095-4462-a202-e01860559471\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-gp8dt" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.355585 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p8x2v" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.363428 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qqg25" Dec 11 02:03:16 crc kubenswrapper[4824]: W1211 02:03:16.363687 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod75da7b9a_7085_492b_954a_6bb920fbb186.slice/crio-c950a9ea49f11cec9f8d112c07a59b1838f7a82c4c1d175768d3f45567e890d3 WatchSource:0}: Error finding container c950a9ea49f11cec9f8d112c07a59b1838f7a82c4c1d175768d3f45567e890d3: Status 404 returned error can't find the container with id c950a9ea49f11cec9f8d112c07a59b1838f7a82c4c1d175768d3f45567e890d3 Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.369698 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wtcp6\" (UniqueName: \"kubernetes.io/projected/1c402f2f-f79a-4a1a-bf40-07da4223059e-kube-api-access-wtcp6\") pod \"multus-admission-controller-857f4d67dd-qp2jl\" (UID: \"1c402f2f-f79a-4a1a-bf40-07da4223059e\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-qp2jl" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.369904 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-qp2jl" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.384476 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-59bgf" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.389460 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-njgbl\" (UniqueName: \"kubernetes.io/projected/63fe488c-de20-435f-83fd-47a4179341e6-kube-api-access-njgbl\") pod \"ingress-canary-zvcj8\" (UID: \"63fe488c-de20-435f-83fd-47a4179341e6\") " pod="openshift-ingress-canary/ingress-canary-zvcj8" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.389622 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-gp8dt" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.398085 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-lpgf5" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.408626 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vkdbw" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.419904 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 02:03:16 crc kubenswrapper[4824]: E1211 02:03:16.420141 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 02:03:16.920092741 +0000 UTC m=+138.609130130 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.420350 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-254w2\" (UID: \"18d74f73-0f04-4f3c-88c4-a431d490db9d\") " pod="openshift-image-registry/image-registry-697d97f7c8-254w2" Dec 11 02:03:16 crc kubenswrapper[4824]: E1211 02:03:16.420866 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 02:03:16.920855583 +0000 UTC m=+138.609892972 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-254w2" (UID: "18d74f73-0f04-4f3c-88c4-a431d490db9d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.429419 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-zvcj8" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.438557 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-lfxkq"] Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.440527 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ks7hm"] Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.442602 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c6vtj\" (UniqueName: \"kubernetes.io/projected/09f964e8-942f-4577-983a-bdc08487c0e8-kube-api-access-c6vtj\") pod \"service-ca-9c57cc56f-4r8x8\" (UID: \"09f964e8-942f-4577-983a-bdc08487c0e8\") " pod="openshift-service-ca/service-ca-9c57cc56f-4r8x8" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.455574 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-hr4qs" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.463527 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-6b6br" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.468457 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-qtll9" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.469699 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-d89dq"] Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.486463 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-kkw8v"] Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.489164 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-r89c2" event={"ID":"8c854339-76bb-439d-b267-b2617cdd3f41","Type":"ContainerStarted","Data":"c392ba5ae5d78b17f66f6d0785dfb4964cc9ab73e49093c0cb1ffc7dfd05f4bf"} Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.498442 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-k6kls"] Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.498494 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2vcc6" event={"ID":"29e1953f-668a-4393-82ab-a45fa4b79ec7","Type":"ContainerStarted","Data":"b8d32f6ee41aad7b1cbaa4f27be5659f36b14a694f022c74be8da2449017bd85"} Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.498586 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-krqhc"] Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.500086 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-tfwnz" event={"ID":"f257970f-c1ce-44fb-b207-a78126794cfe","Type":"ContainerStarted","Data":"97ad0d08ed652e86d0282d5643809ac07601045fd2027d5db77455f5e33158e2"} Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.501058 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2zq2n" event={"ID":"247d17b7-7064-4ef1-9aed-c205fec84bc2","Type":"ContainerStarted","Data":"a6850cc65b363bd45c0b0a23c6157fc78d44654df3421df7e750f14227b60770"} Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.502465 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-l9d4f" event={"ID":"1bdf5eb2-fe05-4e24-9e00-43c72afc3583","Type":"ContainerStarted","Data":"71533e4d94be5bde493c3aad37cdc1a4a3ef65e95bc5078dfd4145af73c5a227"} Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.503672 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g692d" event={"ID":"3e0e52dd-503e-47a4-b68e-3f768ccecfc6","Type":"ContainerStarted","Data":"2eb10a007fd07235b5c4598c73979edebc313c28beece9950c904f7b8b4d064e"} Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.504586 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-qr5xz" event={"ID":"9e099899-aaf7-4bd8-b37e-8616b5060e33","Type":"ContainerStarted","Data":"c3b9fdc4b65c7a93573ec7c1b416fa6a98dd8adb41cd979ae09ab5dded0ce475"} Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.505420 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-l88b4" event={"ID":"75da7b9a-7085-492b-954a-6bb920fbb186","Type":"ContainerStarted","Data":"c950a9ea49f11cec9f8d112c07a59b1838f7a82c4c1d175768d3f45567e890d3"} Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.506250 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-msljv" event={"ID":"5407364a-e2b6-43b4-9eb0-deb12a022289","Type":"ContainerStarted","Data":"1311461ca6b530fd2d5299f8f6c5a8f9f353e64c24afe69528bcac3772261b60"} Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.508000 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hxqw8" event={"ID":"fb52538d-050a-4b69-91b2-aa48f9833fe7","Type":"ContainerStarted","Data":"79fc819ce503f26b123feac36b4b68a22525910ae4f200d35414b7f19c714513"} Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.508957 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gndwh" event={"ID":"e87b6d85-2d13-4b72-bb6b-8299b6c72d0a","Type":"ContainerStarted","Data":"81ed72c5bbb0eb15e6ead8da6cb02d66422f0b6559977a4d297bedb01f7f3245"} Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.509910 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lg9s6" event={"ID":"d5fb9349-891c-447a-bb5a-5e18a0ce9f92","Type":"ContainerStarted","Data":"2361f14fc99cba966dee0bc75b4ff0bac589f7a43b5b5d3b24dce2f9ee255359"} Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.517667 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vtdlb" event={"ID":"3404cb6f-6c3a-4e0a-99a4-27530beefa83","Type":"ContainerStarted","Data":"ddb493c1285a31b31f1d23576a42ead1e1e5b19aa08ab68d3b1384e80134fa99"} Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.519629 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-2ng4n" event={"ID":"f185e5b2-54c0-44ef-bdda-118efa89cb5f","Type":"ContainerStarted","Data":"fd79993141b5bf3dce2b09ed4eed9d68f5dcd1414af9c909f1a7def140a0fa8d"} Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.521271 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 02:03:16 crc kubenswrapper[4824]: E1211 02:03:16.521428 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 02:03:17.021401293 +0000 UTC m=+138.710438672 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.521694 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-254w2\" (UID: \"18d74f73-0f04-4f3c-88c4-a431d490db9d\") " pod="openshift-image-registry/image-registry-697d97f7c8-254w2" Dec 11 02:03:16 crc kubenswrapper[4824]: E1211 02:03:16.521974 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 02:03:17.02196748 +0000 UTC m=+138.711004859 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-254w2" (UID: "18d74f73-0f04-4f3c-88c4-a431d490db9d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.538770 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hfh5h\" (UniqueName: \"kubernetes.io/projected/398ffffe-110f-4f78-a784-2ee501ee58a3-kube-api-access-hfh5h\") pod \"olm-operator-6b444d44fb-66mv8\" (UID: \"398ffffe-110f-4f78-a784-2ee501ee58a3\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-66mv8" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.605032 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-8fksd"] Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.624745 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 02:03:16 crc kubenswrapper[4824]: E1211 02:03:16.625320 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 02:03:17.125305833 +0000 UTC m=+138.814343212 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.640669 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-66mv8" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.653669 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-4r8x8" Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.729012 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-254w2\" (UID: \"18d74f73-0f04-4f3c-88c4-a431d490db9d\") " pod="openshift-image-registry/image-registry-697d97f7c8-254w2" Dec 11 02:03:16 crc kubenswrapper[4824]: E1211 02:03:16.729382 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 02:03:17.229368918 +0000 UTC m=+138.918406297 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-254w2" (UID: "18d74f73-0f04-4f3c-88c4-a431d490db9d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.765077 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kdzkc"] Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.802326 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-w6b4l"] Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.820279 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423640-42vd6"] Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.830719 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 02:03:16 crc kubenswrapper[4824]: E1211 02:03:16.831018 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 02:03:17.3310036 +0000 UTC m=+139.020040979 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:03:16 crc kubenswrapper[4824]: W1211 02:03:16.872371 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod877d4a7f_46b0_4f9c_a702_b6b83620c3cb.slice/crio-211bc09b200a5f1cdac291be6cbe311eba7401e2fa98bf8e0ec75e89ec8174e0 WatchSource:0}: Error finding container 211bc09b200a5f1cdac291be6cbe311eba7401e2fa98bf8e0ec75e89ec8174e0: Status 404 returned error can't find the container with id 211bc09b200a5f1cdac291be6cbe311eba7401e2fa98bf8e0ec75e89ec8174e0 Dec 11 02:03:16 crc kubenswrapper[4824]: I1211 02:03:16.932377 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-254w2\" (UID: \"18d74f73-0f04-4f3c-88c4-a431d490db9d\") " pod="openshift-image-registry/image-registry-697d97f7c8-254w2" Dec 11 02:03:16 crc kubenswrapper[4824]: E1211 02:03:16.932675 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 02:03:17.432663872 +0000 UTC m=+139.121701251 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-254w2" (UID: "18d74f73-0f04-4f3c-88c4-a431d490db9d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:03:17 crc kubenswrapper[4824]: I1211 02:03:17.033882 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 02:03:17 crc kubenswrapper[4824]: E1211 02:03:17.034437 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 02:03:17.534410618 +0000 UTC m=+139.223447997 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:03:17 crc kubenswrapper[4824]: I1211 02:03:17.034825 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-254w2\" (UID: \"18d74f73-0f04-4f3c-88c4-a431d490db9d\") " pod="openshift-image-registry/image-registry-697d97f7c8-254w2" Dec 11 02:03:17 crc kubenswrapper[4824]: E1211 02:03:17.035126 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 02:03:17.535095608 +0000 UTC m=+139.224132987 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-254w2" (UID: "18d74f73-0f04-4f3c-88c4-a431d490db9d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:03:17 crc kubenswrapper[4824]: I1211 02:03:17.044514 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-tz2j9"] Dec 11 02:03:17 crc kubenswrapper[4824]: I1211 02:03:17.137030 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 02:03:17 crc kubenswrapper[4824]: E1211 02:03:17.137498 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 02:03:17.637476092 +0000 UTC m=+139.326513471 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:03:17 crc kubenswrapper[4824]: I1211 02:03:17.137685 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-254w2\" (UID: \"18d74f73-0f04-4f3c-88c4-a431d490db9d\") " pod="openshift-image-registry/image-registry-697d97f7c8-254w2" Dec 11 02:03:17 crc kubenswrapper[4824]: E1211 02:03:17.137948 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 02:03:17.637936516 +0000 UTC m=+139.326973895 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-254w2" (UID: "18d74f73-0f04-4f3c-88c4-a431d490db9d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:03:17 crc kubenswrapper[4824]: I1211 02:03:17.176352 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-jlqk5"] Dec 11 02:03:17 crc kubenswrapper[4824]: I1211 02:03:17.238683 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 02:03:17 crc kubenswrapper[4824]: E1211 02:03:17.238804 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 02:03:17.738784444 +0000 UTC m=+139.427821823 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:03:17 crc kubenswrapper[4824]: I1211 02:03:17.239145 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-254w2\" (UID: \"18d74f73-0f04-4f3c-88c4-a431d490db9d\") " pod="openshift-image-registry/image-registry-697d97f7c8-254w2" Dec 11 02:03:17 crc kubenswrapper[4824]: E1211 02:03:17.239504 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 02:03:17.739492046 +0000 UTC m=+139.428529425 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-254w2" (UID: "18d74f73-0f04-4f3c-88c4-a431d490db9d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:03:17 crc kubenswrapper[4824]: I1211 02:03:17.313532 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qqg25"] Dec 11 02:03:17 crc kubenswrapper[4824]: I1211 02:03:17.330622 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p8x2v"] Dec 11 02:03:17 crc kubenswrapper[4824]: I1211 02:03:17.346071 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 02:03:17 crc kubenswrapper[4824]: E1211 02:03:17.346590 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 02:03:17.846571911 +0000 UTC m=+139.535609290 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:03:17 crc kubenswrapper[4824]: I1211 02:03:17.449995 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-254w2\" (UID: \"18d74f73-0f04-4f3c-88c4-a431d490db9d\") " pod="openshift-image-registry/image-registry-697d97f7c8-254w2" Dec 11 02:03:17 crc kubenswrapper[4824]: E1211 02:03:17.450330 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 02:03:17.950318786 +0000 UTC m=+139.639356165 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-254w2" (UID: "18d74f73-0f04-4f3c-88c4-a431d490db9d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:03:17 crc kubenswrapper[4824]: I1211 02:03:17.559369 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 02:03:17 crc kubenswrapper[4824]: E1211 02:03:17.560348 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 02:03:18.060310579 +0000 UTC m=+139.749347958 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:03:17 crc kubenswrapper[4824]: I1211 02:03:17.564231 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-qp2jl"] Dec 11 02:03:17 crc kubenswrapper[4824]: I1211 02:03:17.579148 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-krqhc" event={"ID":"8e2095d8-25c6-407f-aadb-f4603ae44d15","Type":"ContainerStarted","Data":"49522bf90354a25a56ffb4856a3b8710573de70366c57425f229189b3721ed2b"} Dec 11 02:03:17 crc kubenswrapper[4824]: I1211 02:03:17.580642 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-w6b4l" event={"ID":"f5149a23-92a4-42e6-8b20-8411d1d0536f","Type":"ContainerStarted","Data":"73e73368f6f22aa2fb9a7b638597663d54bc66e692db6804a1a47f99ea427f0d"} Dec 11 02:03:17 crc kubenswrapper[4824]: I1211 02:03:17.613058 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-z42g5"] Dec 11 02:03:17 crc kubenswrapper[4824]: I1211 02:03:17.614229 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gndwh" event={"ID":"e87b6d85-2d13-4b72-bb6b-8299b6c72d0a","Type":"ContainerStarted","Data":"04bd2efa2264ede03170cfa1c8bbf2737b00f97adf984a0ce57324a9e72cc9cd"} Dec 11 02:03:17 crc kubenswrapper[4824]: I1211 02:03:17.620501 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-l9d4f" event={"ID":"1bdf5eb2-fe05-4e24-9e00-43c72afc3583","Type":"ContainerStarted","Data":"0121a6485cec61a8b44c16765f66c94217718710b924d2ae72ff108600ccfa3e"} Dec 11 02:03:17 crc kubenswrapper[4824]: I1211 02:03:17.621022 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-l9d4f" Dec 11 02:03:17 crc kubenswrapper[4824]: I1211 02:03:17.636025 4824 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-l9d4f container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.8:6443/healthz\": dial tcp 10.217.0.8:6443: connect: connection refused" start-of-body= Dec 11 02:03:17 crc kubenswrapper[4824]: I1211 02:03:17.636077 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-l9d4f" podUID="1bdf5eb2-fe05-4e24-9e00-43c72afc3583" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.8:6443/healthz\": dial tcp 10.217.0.8:6443: connect: connection refused" Dec 11 02:03:17 crc kubenswrapper[4824]: I1211 02:03:17.638752 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-59bgf"] Dec 11 02:03:17 crc kubenswrapper[4824]: I1211 02:03:17.639525 4824 generic.go:334] "Generic (PLEG): container finished" podID="3e0e52dd-503e-47a4-b68e-3f768ccecfc6" containerID="4940bc48dd652b127bf32bc7a63d802bf0f28470ebd241ba4f0b7aaf018fc0d2" exitCode=0 Dec 11 02:03:17 crc kubenswrapper[4824]: I1211 02:03:17.639602 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g692d" event={"ID":"3e0e52dd-503e-47a4-b68e-3f768ccecfc6","Type":"ContainerDied","Data":"4940bc48dd652b127bf32bc7a63d802bf0f28470ebd241ba4f0b7aaf018fc0d2"} Dec 11 02:03:17 crc kubenswrapper[4824]: I1211 02:03:17.651972 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-pg8rn" event={"ID":"877d4a7f-46b0-4f9c-a702-b6b83620c3cb","Type":"ContainerStarted","Data":"211bc09b200a5f1cdac291be6cbe311eba7401e2fa98bf8e0ec75e89ec8174e0"} Dec 11 02:03:17 crc kubenswrapper[4824]: I1211 02:03:17.662049 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-254w2\" (UID: \"18d74f73-0f04-4f3c-88c4-a431d490db9d\") " pod="openshift-image-registry/image-registry-697d97f7c8-254w2" Dec 11 02:03:17 crc kubenswrapper[4824]: I1211 02:03:17.666059 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-lfxkq" event={"ID":"bb4dde0f-222d-4609-b1c9-fed653668e3a","Type":"ContainerStarted","Data":"905661b5bdb4f5730b8187888a2c9b01a52836e67cbf1b5dec960bbb8d3bcc72"} Dec 11 02:03:17 crc kubenswrapper[4824]: I1211 02:03:17.672255 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-tz2j9" event={"ID":"dbdb9d15-7786-4fef-a38e-e3be94f1bcb2","Type":"ContainerStarted","Data":"47d4d8f563a26a9253e51895dba3372df36c1a1296bbd25499b5316957d0c756"} Dec 11 02:03:17 crc kubenswrapper[4824]: E1211 02:03:17.677812 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 02:03:18.177793786 +0000 UTC m=+139.866831165 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-254w2" (UID: "18d74f73-0f04-4f3c-88c4-a431d490db9d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:03:17 crc kubenswrapper[4824]: I1211 02:03:17.692178 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-kkw8v" event={"ID":"8650bbf0-78d2-410e-a62d-d4adf673cb55","Type":"ContainerStarted","Data":"1349b59f33aff0ce7ebfcb98be7c7e7f4f902045e7b4396bd8f8645901049a8c"} Dec 11 02:03:17 crc kubenswrapper[4824]: I1211 02:03:17.741551 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-vkdbw"] Dec 11 02:03:17 crc kubenswrapper[4824]: I1211 02:03:17.744280 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-tfwnz" event={"ID":"f257970f-c1ce-44fb-b207-a78126794cfe","Type":"ContainerStarted","Data":"2b5f78bebe2660fb7643690e3ed790d90dd044181fe407b156ff528170a03811"} Dec 11 02:03:17 crc kubenswrapper[4824]: I1211 02:03:17.754282 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-qr5xz" event={"ID":"9e099899-aaf7-4bd8-b37e-8616b5060e33","Type":"ContainerStarted","Data":"5a3bbef81765195bec5e85dc7ab9b24eec8b0ff2accaa3e59b2312b6d539413d"} Dec 11 02:03:17 crc kubenswrapper[4824]: I1211 02:03:17.755128 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-qr5xz" Dec 11 02:03:17 crc kubenswrapper[4824]: I1211 02:03:17.771629 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2zq2n" event={"ID":"247d17b7-7064-4ef1-9aed-c205fec84bc2","Type":"ContainerStarted","Data":"582d3a48d11f95db2f06dd91bd1a94e4044c18d1139af3f64052a5556487b1d7"} Dec 11 02:03:17 crc kubenswrapper[4824]: I1211 02:03:17.772818 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 02:03:17 crc kubenswrapper[4824]: E1211 02:03:17.773772 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 02:03:18.273754628 +0000 UTC m=+139.962792007 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:03:17 crc kubenswrapper[4824]: I1211 02:03:17.855140 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-qr5xz" Dec 11 02:03:17 crc kubenswrapper[4824]: I1211 02:03:17.868857 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-zvcj8"] Dec 11 02:03:17 crc kubenswrapper[4824]: I1211 02:03:17.888410 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-254w2\" (UID: \"18d74f73-0f04-4f3c-88c4-a431d490db9d\") " pod="openshift-image-registry/image-registry-697d97f7c8-254w2" Dec 11 02:03:17 crc kubenswrapper[4824]: E1211 02:03:17.890600 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 02:03:18.390584206 +0000 UTC m=+140.079621585 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-254w2" (UID: "18d74f73-0f04-4f3c-88c4-a431d490db9d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:03:17 crc kubenswrapper[4824]: I1211 02:03:17.927627 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lg9s6" event={"ID":"d5fb9349-891c-447a-bb5a-5e18a0ce9f92","Type":"ContainerStarted","Data":"ce3c39dc14d53014b6d9a304bb4ca1c803bd65e24efa9393d03fb807fc008f32"} Dec 11 02:03:17 crc kubenswrapper[4824]: I1211 02:03:17.969681 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-qr5xz" podStartSLOduration=120.96965343 podStartE2EDuration="2m0.96965343s" podCreationTimestamp="2025-12-11 02:01:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:03:17.96828494 +0000 UTC m=+139.657322319" watchObservedRunningTime="2025-12-11 02:03:17.96965343 +0000 UTC m=+139.658690809" Dec 11 02:03:17 crc kubenswrapper[4824]: I1211 02:03:17.972749 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gndwh" podStartSLOduration=120.972740193 podStartE2EDuration="2m0.972740193s" podCreationTimestamp="2025-12-11 02:01:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:03:17.918508555 +0000 UTC m=+139.607545934" watchObservedRunningTime="2025-12-11 02:03:17.972740193 +0000 UTC m=+139.661777572" Dec 11 02:03:17 crc kubenswrapper[4824]: I1211 02:03:17.996260 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 02:03:17 crc kubenswrapper[4824]: E1211 02:03:17.997440 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 02:03:18.497420454 +0000 UTC m=+140.186457833 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:03:18 crc kubenswrapper[4824]: I1211 02:03:18.019239 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2zq2n" podStartSLOduration=121.019220459 podStartE2EDuration="2m1.019220459s" podCreationTimestamp="2025-12-11 02:01:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:03:18.018351243 +0000 UTC m=+139.707388622" watchObservedRunningTime="2025-12-11 02:03:18.019220459 +0000 UTC m=+139.708257838" Dec 11 02:03:18 crc kubenswrapper[4824]: I1211 02:03:18.052627 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-tfwnz" podStartSLOduration=121.052605991 podStartE2EDuration="2m1.052605991s" podCreationTimestamp="2025-12-11 02:01:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:03:18.052095176 +0000 UTC m=+139.741132555" watchObservedRunningTime="2025-12-11 02:03:18.052605991 +0000 UTC m=+139.741643370" Dec 11 02:03:18 crc kubenswrapper[4824]: I1211 02:03:18.088875 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-l9d4f" podStartSLOduration=121.08885338 podStartE2EDuration="2m1.08885338s" podCreationTimestamp="2025-12-11 02:01:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:03:18.083623533 +0000 UTC m=+139.772660912" watchObservedRunningTime="2025-12-11 02:03:18.08885338 +0000 UTC m=+139.777890759" Dec 11 02:03:18 crc kubenswrapper[4824]: I1211 02:03:18.100921 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-254w2\" (UID: \"18d74f73-0f04-4f3c-88c4-a431d490db9d\") " pod="openshift-image-registry/image-registry-697d97f7c8-254w2" Dec 11 02:03:18 crc kubenswrapper[4824]: E1211 02:03:18.101226 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 02:03:18.601214521 +0000 UTC m=+140.290251900 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-254w2" (UID: "18d74f73-0f04-4f3c-88c4-a431d490db9d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:03:18 crc kubenswrapper[4824]: I1211 02:03:18.112896 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kdzkc" event={"ID":"dccdfcce-6b35-40e6-8bdc-7eed59518f24","Type":"ContainerStarted","Data":"00fcc4627656b162f82c19478f57511e8dabc5d729d18e67944719b7b1ba7aed"} Dec 11 02:03:18 crc kubenswrapper[4824]: I1211 02:03:18.141410 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lg9s6" podStartSLOduration=121.141393988 podStartE2EDuration="2m1.141393988s" podCreationTimestamp="2025-12-11 02:01:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:03:18.14016976 +0000 UTC m=+139.829207139" watchObservedRunningTime="2025-12-11 02:03:18.141393988 +0000 UTC m=+139.830431367" Dec 11 02:03:18 crc kubenswrapper[4824]: I1211 02:03:18.145683 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8fksd" event={"ID":"95b90361-2f73-4649-bc1a-1202ba511baf","Type":"ContainerStarted","Data":"1d40a92829d5dcd576f8ad49cba51e38efca2ae85eeedc6906677717370c945c"} Dec 11 02:03:18 crc kubenswrapper[4824]: I1211 02:03:18.181171 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-gp8dt"] Dec 11 02:03:18 crc kubenswrapper[4824]: I1211 02:03:18.202467 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 02:03:18 crc kubenswrapper[4824]: E1211 02:03:18.206841 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 02:03:18.706818592 +0000 UTC m=+140.395855971 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:03:18 crc kubenswrapper[4824]: I1211 02:03:18.210286 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-jlqk5" event={"ID":"04cb3d8f-ccf2-410c-b51c-445add07699c","Type":"ContainerStarted","Data":"0b01ce0860c410698abc23143b1828e0072c0d33a666cb5184ca9cf1bceaa317"} Dec 11 02:03:18 crc kubenswrapper[4824]: I1211 02:03:18.259473 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423640-42vd6" event={"ID":"4c22bf31-ad33-44cd-a495-05d83688f2bd","Type":"ContainerStarted","Data":"4fc4c2da81fdbc8e9a09e760e785feaaaa542605702fe0293f325b9f83602734"} Dec 11 02:03:18 crc kubenswrapper[4824]: I1211 02:03:18.263606 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-r89c2" event={"ID":"8c854339-76bb-439d-b267-b2617cdd3f41","Type":"ContainerStarted","Data":"cd30f2840a0678a55b8c17559175f95daffead0b7a29f7ed3f4806bda7a3ee5a"} Dec 11 02:03:18 crc kubenswrapper[4824]: W1211 02:03:18.267358 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod505f6458_5d5a_4725_a529_1075770afa95.slice/crio-d1cae587ff5c21ff866193ca96dc3884e82c2716e51a821945bc26c54ba13813 WatchSource:0}: Error finding container d1cae587ff5c21ff866193ca96dc3884e82c2716e51a821945bc26c54ba13813: Status 404 returned error can't find the container with id d1cae587ff5c21ff866193ca96dc3884e82c2716e51a821945bc26c54ba13813 Dec 11 02:03:18 crc kubenswrapper[4824]: I1211 02:03:18.267945 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-2ng4n" event={"ID":"f185e5b2-54c0-44ef-bdda-118efa89cb5f","Type":"ContainerStarted","Data":"4bba51830ef805c81c2241f4e328c03b922ab42df8c563c22885e05eb1a9bb3c"} Dec 11 02:03:18 crc kubenswrapper[4824]: I1211 02:03:18.268789 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-2ng4n" Dec 11 02:03:18 crc kubenswrapper[4824]: I1211 02:03:18.270606 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-k6kls" event={"ID":"ad510a18-cd72-4cf0-9b0c-dd3540d491fa","Type":"ContainerStarted","Data":"9f6757a7ca4186844eb08113017138bcde22f84e6d25f11852180354ee173046"} Dec 11 02:03:18 crc kubenswrapper[4824]: I1211 02:03:18.313267 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-254w2\" (UID: \"18d74f73-0f04-4f3c-88c4-a431d490db9d\") " pod="openshift-image-registry/image-registry-697d97f7c8-254w2" Dec 11 02:03:18 crc kubenswrapper[4824]: E1211 02:03:18.313693 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 02:03:18.81365853 +0000 UTC m=+140.502695909 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-254w2" (UID: "18d74f73-0f04-4f3c-88c4-a431d490db9d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:03:18 crc kubenswrapper[4824]: I1211 02:03:18.317413 4824 patch_prober.go:28] interesting pod/downloads-7954f5f757-2ng4n container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Dec 11 02:03:18 crc kubenswrapper[4824]: I1211 02:03:18.317477 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-2ng4n" podUID="f185e5b2-54c0-44ef-bdda-118efa89cb5f" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Dec 11 02:03:18 crc kubenswrapper[4824]: I1211 02:03:18.320326 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-r89c2" podStartSLOduration=121.32031223 podStartE2EDuration="2m1.32031223s" podCreationTimestamp="2025-12-11 02:01:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:03:18.319603158 +0000 UTC m=+140.008640537" watchObservedRunningTime="2025-12-11 02:03:18.32031223 +0000 UTC m=+140.009349609" Dec 11 02:03:18 crc kubenswrapper[4824]: I1211 02:03:18.359478 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-2ng4n" podStartSLOduration=121.359464476 podStartE2EDuration="2m1.359464476s" podCreationTimestamp="2025-12-11 02:01:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:03:18.354379233 +0000 UTC m=+140.043416612" watchObservedRunningTime="2025-12-11 02:03:18.359464476 +0000 UTC m=+140.048501855" Dec 11 02:03:18 crc kubenswrapper[4824]: I1211 02:03:18.361453 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-lpgf5"] Dec 11 02:03:18 crc kubenswrapper[4824]: I1211 02:03:18.368775 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-msljv" event={"ID":"5407364a-e2b6-43b4-9eb0-deb12a022289","Type":"ContainerStarted","Data":"dc7036b2d9bdde3b494cb3e26180f2b717abfa0d317411a55d23e90f4b10e57d"} Dec 11 02:03:18 crc kubenswrapper[4824]: I1211 02:03:18.369911 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-msljv" Dec 11 02:03:18 crc kubenswrapper[4824]: I1211 02:03:18.414900 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 02:03:18 crc kubenswrapper[4824]: E1211 02:03:18.416356 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 02:03:18.916339624 +0000 UTC m=+140.605377003 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:03:18 crc kubenswrapper[4824]: I1211 02:03:18.422624 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-msljv" Dec 11 02:03:18 crc kubenswrapper[4824]: I1211 02:03:18.425029 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-msljv" podStartSLOduration=121.425016044 podStartE2EDuration="2m1.425016044s" podCreationTimestamp="2025-12-11 02:01:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:03:18.423337003 +0000 UTC m=+140.112374382" watchObservedRunningTime="2025-12-11 02:03:18.425016044 +0000 UTC m=+140.114053423" Dec 11 02:03:18 crc kubenswrapper[4824]: I1211 02:03:18.457400 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-d89dq" event={"ID":"7704cded-5ba1-4474-8a31-c0ba947b7679","Type":"ContainerStarted","Data":"6dbed2c2ba853dd9a0e9495b0c6f0ae006417db0135da6683ea4b86f85591507"} Dec 11 02:03:18 crc kubenswrapper[4824]: I1211 02:03:18.459434 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vtdlb" event={"ID":"3404cb6f-6c3a-4e0a-99a4-27530beefa83","Type":"ContainerStarted","Data":"c6726a5fdd7d1e22c5f46bf09362d9a5d714adf091dae29d5687eabe65e5cc12"} Dec 11 02:03:18 crc kubenswrapper[4824]: W1211 02:03:18.462553 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod69814666_b43a_4cab_9864_864862277916.slice/crio-4e59d7c639bef8858829f61b2936575793f5e97fab8e57dd95a24cc79e962150 WatchSource:0}: Error finding container 4e59d7c639bef8858829f61b2936575793f5e97fab8e57dd95a24cc79e962150: Status 404 returned error can't find the container with id 4e59d7c639bef8858829f61b2936575793f5e97fab8e57dd95a24cc79e962150 Dec 11 02:03:18 crc kubenswrapper[4824]: I1211 02:03:18.522136 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-254w2\" (UID: \"18d74f73-0f04-4f3c-88c4-a431d490db9d\") " pod="openshift-image-registry/image-registry-697d97f7c8-254w2" Dec 11 02:03:18 crc kubenswrapper[4824]: E1211 02:03:18.523521 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 02:03:19.023238893 +0000 UTC m=+140.712276272 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-254w2" (UID: "18d74f73-0f04-4f3c-88c4-a431d490db9d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:03:18 crc kubenswrapper[4824]: I1211 02:03:18.552142 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2vcc6" event={"ID":"29e1953f-668a-4393-82ab-a45fa4b79ec7","Type":"ContainerStarted","Data":"a2ced031338e59cdc4318da90ef82fbbe0e2185636e1b2423c11311f5f7db832"} Dec 11 02:03:18 crc kubenswrapper[4824]: I1211 02:03:18.590417 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ks7hm" event={"ID":"fa007484-938d-460e-9c5e-c3f5b31a25d2","Type":"ContainerStarted","Data":"bfe2fa9e70f8b62f824f6b02b8dc7f61eba22d29037290ef380bb0f0f0be2c92"} Dec 11 02:03:18 crc kubenswrapper[4824]: I1211 02:03:18.591888 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hxqw8" event={"ID":"fb52538d-050a-4b69-91b2-aa48f9833fe7","Type":"ContainerStarted","Data":"ab185e39b2d7340bc060b9cf2caad7bcdb6ff400ef7724b9bede014531cec69d"} Dec 11 02:03:18 crc kubenswrapper[4824]: I1211 02:03:18.627578 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 02:03:18 crc kubenswrapper[4824]: E1211 02:03:18.627688 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 02:03:19.127669709 +0000 UTC m=+140.816707088 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:03:18 crc kubenswrapper[4824]: I1211 02:03:18.627955 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-254w2\" (UID: \"18d74f73-0f04-4f3c-88c4-a431d490db9d\") " pod="openshift-image-registry/image-registry-697d97f7c8-254w2" Dec 11 02:03:18 crc kubenswrapper[4824]: E1211 02:03:18.628744 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 02:03:19.128736951 +0000 UTC m=+140.817774330 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-254w2" (UID: "18d74f73-0f04-4f3c-88c4-a431d490db9d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:03:18 crc kubenswrapper[4824]: I1211 02:03:18.695654 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2vcc6" podStartSLOduration=121.69563363 podStartE2EDuration="2m1.69563363s" podCreationTimestamp="2025-12-11 02:01:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:03:18.59238124 +0000 UTC m=+140.281418629" watchObservedRunningTime="2025-12-11 02:03:18.69563363 +0000 UTC m=+140.384671009" Dec 11 02:03:18 crc kubenswrapper[4824]: I1211 02:03:18.708798 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-qtll9"] Dec 11 02:03:18 crc kubenswrapper[4824]: I1211 02:03:18.720575 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-66mv8"] Dec 11 02:03:18 crc kubenswrapper[4824]: I1211 02:03:18.730097 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 02:03:18 crc kubenswrapper[4824]: E1211 02:03:18.731491 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 02:03:19.231474476 +0000 UTC m=+140.920511855 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:03:18 crc kubenswrapper[4824]: I1211 02:03:18.733186 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-hr4qs"] Dec 11 02:03:18 crc kubenswrapper[4824]: I1211 02:03:18.847282 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-254w2\" (UID: \"18d74f73-0f04-4f3c-88c4-a431d490db9d\") " pod="openshift-image-registry/image-registry-697d97f7c8-254w2" Dec 11 02:03:18 crc kubenswrapper[4824]: E1211 02:03:18.847938 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 02:03:19.347925553 +0000 UTC m=+141.036962932 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-254w2" (UID: "18d74f73-0f04-4f3c-88c4-a431d490db9d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:03:18 crc kubenswrapper[4824]: I1211 02:03:18.948017 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-4r8x8"] Dec 11 02:03:18 crc kubenswrapper[4824]: I1211 02:03:18.948464 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 02:03:18 crc kubenswrapper[4824]: E1211 02:03:18.948812 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 02:03:19.448799112 +0000 UTC m=+141.137836491 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:03:19 crc kubenswrapper[4824]: I1211 02:03:19.050341 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-254w2\" (UID: \"18d74f73-0f04-4f3c-88c4-a431d490db9d\") " pod="openshift-image-registry/image-registry-697d97f7c8-254w2" Dec 11 02:03:19 crc kubenswrapper[4824]: E1211 02:03:19.055471 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 02:03:19.555454214 +0000 UTC m=+141.244491583 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-254w2" (UID: "18d74f73-0f04-4f3c-88c4-a431d490db9d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:03:19 crc kubenswrapper[4824]: I1211 02:03:19.158265 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 02:03:19 crc kubenswrapper[4824]: E1211 02:03:19.158803 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 02:03:19.658774887 +0000 UTC m=+141.347812266 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:03:19 crc kubenswrapper[4824]: I1211 02:03:19.263661 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-254w2\" (UID: \"18d74f73-0f04-4f3c-88c4-a431d490db9d\") " pod="openshift-image-registry/image-registry-697d97f7c8-254w2" Dec 11 02:03:19 crc kubenswrapper[4824]: E1211 02:03:19.264105 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 02:03:19.764089249 +0000 UTC m=+141.453126628 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-254w2" (UID: "18d74f73-0f04-4f3c-88c4-a431d490db9d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:03:19 crc kubenswrapper[4824]: I1211 02:03:19.366444 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 02:03:19 crc kubenswrapper[4824]: E1211 02:03:19.367261 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 02:03:19.867242436 +0000 UTC m=+141.556279815 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:03:19 crc kubenswrapper[4824]: I1211 02:03:19.468158 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-254w2\" (UID: \"18d74f73-0f04-4f3c-88c4-a431d490db9d\") " pod="openshift-image-registry/image-registry-697d97f7c8-254w2" Dec 11 02:03:19 crc kubenswrapper[4824]: E1211 02:03:19.468464 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 02:03:19.968452625 +0000 UTC m=+141.657489994 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-254w2" (UID: "18d74f73-0f04-4f3c-88c4-a431d490db9d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:03:19 crc kubenswrapper[4824]: I1211 02:03:19.573101 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 02:03:19 crc kubenswrapper[4824]: E1211 02:03:19.574121 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 02:03:20.074081577 +0000 UTC m=+141.763118956 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:03:19 crc kubenswrapper[4824]: I1211 02:03:19.675954 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-254w2\" (UID: \"18d74f73-0f04-4f3c-88c4-a431d490db9d\") " pod="openshift-image-registry/image-registry-697d97f7c8-254w2" Dec 11 02:03:19 crc kubenswrapper[4824]: E1211 02:03:19.676466 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 02:03:20.176450702 +0000 UTC m=+141.865488081 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-254w2" (UID: "18d74f73-0f04-4f3c-88c4-a431d490db9d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:03:19 crc kubenswrapper[4824]: I1211 02:03:19.725088 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vtdlb" event={"ID":"3404cb6f-6c3a-4e0a-99a4-27530beefa83","Type":"ContainerStarted","Data":"dd863e3d383f756b02120df7f6434d0dd32e5ec28e361ccb890e24bdc2cb46a2"} Dec 11 02:03:19 crc kubenswrapper[4824]: I1211 02:03:19.746524 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-kkw8v" event={"ID":"8650bbf0-78d2-410e-a62d-d4adf673cb55","Type":"ContainerStarted","Data":"15dd9bb0f01d075e71c2e5d4d6e6fa883500380f886eab5a259ccdd95895d8ad"} Dec 11 02:03:19 crc kubenswrapper[4824]: I1211 02:03:19.775375 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vtdlb" podStartSLOduration=122.775351421 podStartE2EDuration="2m2.775351421s" podCreationTimestamp="2025-12-11 02:01:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:03:19.762599068 +0000 UTC m=+141.451636447" watchObservedRunningTime="2025-12-11 02:03:19.775351421 +0000 UTC m=+141.464388800" Dec 11 02:03:19 crc kubenswrapper[4824]: I1211 02:03:19.779999 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 02:03:19 crc kubenswrapper[4824]: E1211 02:03:19.780393 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 02:03:20.280366592 +0000 UTC m=+141.969403971 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:03:19 crc kubenswrapper[4824]: I1211 02:03:19.781616 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g692d" event={"ID":"3e0e52dd-503e-47a4-b68e-3f768ccecfc6","Type":"ContainerStarted","Data":"bf3729d420919535579cb465f22138515efa299d8039f33fced0c8c33018a305"} Dec 11 02:03:19 crc kubenswrapper[4824]: I1211 02:03:19.829546 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-6b6br" event={"ID":"52938126-55df-4106-9bed-ad7cc0f1130d","Type":"ContainerStarted","Data":"e710835edd43d8991a3f7584ed6e8091c3c19bda0d4e0e24954091144131058b"} Dec 11 02:03:19 crc kubenswrapper[4824]: I1211 02:03:19.829603 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-6b6br" event={"ID":"52938126-55df-4106-9bed-ad7cc0f1130d","Type":"ContainerStarted","Data":"ee0d2bc662de13b262d3d924e557e4ed51fec928f2505c7fd7f9d60575c0992c"} Dec 11 02:03:19 crc kubenswrapper[4824]: I1211 02:03:19.868427 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g692d" podStartSLOduration=122.868410255 podStartE2EDuration="2m2.868410255s" podCreationTimestamp="2025-12-11 02:01:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:03:19.86756165 +0000 UTC m=+141.556599039" watchObservedRunningTime="2025-12-11 02:03:19.868410255 +0000 UTC m=+141.557447634" Dec 11 02:03:19 crc kubenswrapper[4824]: I1211 02:03:19.880489 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-d89dq" event={"ID":"7704cded-5ba1-4474-8a31-c0ba947b7679","Type":"ContainerStarted","Data":"223a959a7070dca1ef2682f43b34694a79893f661aea77b1bee3fcf6ed26674d"} Dec 11 02:03:19 crc kubenswrapper[4824]: I1211 02:03:19.880529 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-d89dq" event={"ID":"7704cded-5ba1-4474-8a31-c0ba947b7679","Type":"ContainerStarted","Data":"7edb56e13a207b3456459193ee8823b66ca91c2cf442e3af77465789a5b9ecf8"} Dec 11 02:03:19 crc kubenswrapper[4824]: I1211 02:03:19.886926 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-254w2\" (UID: \"18d74f73-0f04-4f3c-88c4-a431d490db9d\") " pod="openshift-image-registry/image-registry-697d97f7c8-254w2" Dec 11 02:03:19 crc kubenswrapper[4824]: E1211 02:03:19.888729 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 02:03:20.388717465 +0000 UTC m=+142.077754844 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-254w2" (UID: "18d74f73-0f04-4f3c-88c4-a431d490db9d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:03:19 crc kubenswrapper[4824]: I1211 02:03:19.909331 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-6b6br" podStartSLOduration=6.909314624 podStartE2EDuration="6.909314624s" podCreationTimestamp="2025-12-11 02:03:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:03:19.908812619 +0000 UTC m=+141.597849998" watchObservedRunningTime="2025-12-11 02:03:19.909314624 +0000 UTC m=+141.598351993" Dec 11 02:03:19 crc kubenswrapper[4824]: I1211 02:03:19.957076 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423640-42vd6" event={"ID":"4c22bf31-ad33-44cd-a495-05d83688f2bd","Type":"ContainerStarted","Data":"f63691527d0e51385a07c889a507a7c4d8c0a08751f37f06bdd5379d9c81e323"} Dec 11 02:03:19 crc kubenswrapper[4824]: I1211 02:03:19.975358 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-4r8x8" event={"ID":"09f964e8-942f-4577-983a-bdc08487c0e8","Type":"ContainerStarted","Data":"570d82bad638079b9b925fc293a79ce96c178950b58bdc4e7190a0b783884abb"} Dec 11 02:03:19 crc kubenswrapper[4824]: I1211 02:03:19.983337 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-pg8rn" event={"ID":"877d4a7f-46b0-4f9c-a702-b6b83620c3cb","Type":"ContainerStarted","Data":"45515270ca4965616bb8a7584b7c4e9a9d9b4530655876e4b327c812e31b907c"} Dec 11 02:03:19 crc kubenswrapper[4824]: I1211 02:03:19.988760 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 02:03:19 crc kubenswrapper[4824]: E1211 02:03:19.989814 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 02:03:20.48979956 +0000 UTC m=+142.178836929 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:03:20 crc kubenswrapper[4824]: I1211 02:03:20.059381 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-z42g5" event={"ID":"ffb4acd4-543f-4509-b993-249e3656048d","Type":"ContainerStarted","Data":"60693fc1fe023dcf8dbae48f9d5d679fb604f2a41ce2a4e0c21987b053e0c4f1"} Dec 11 02:03:20 crc kubenswrapper[4824]: I1211 02:03:20.059424 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-z42g5" event={"ID":"ffb4acd4-543f-4509-b993-249e3656048d","Type":"ContainerStarted","Data":"08bb8882304545419153546c5cbe5f2043cb256198be52a6208a7a5118123a30"} Dec 11 02:03:20 crc kubenswrapper[4824]: I1211 02:03:20.060321 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-z42g5" Dec 11 02:03:20 crc kubenswrapper[4824]: I1211 02:03:20.092515 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-254w2\" (UID: \"18d74f73-0f04-4f3c-88c4-a431d490db9d\") " pod="openshift-image-registry/image-registry-697d97f7c8-254w2" Dec 11 02:03:20 crc kubenswrapper[4824]: I1211 02:03:20.095031 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-tz2j9" event={"ID":"dbdb9d15-7786-4fef-a38e-e3be94f1bcb2","Type":"ContainerStarted","Data":"77eb52b0c2f094834a36648355ceba3dae342b063a542caa37569111691e89f4"} Dec 11 02:03:20 crc kubenswrapper[4824]: E1211 02:03:20.097019 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 02:03:20.59700442 +0000 UTC m=+142.286041799 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-254w2" (UID: "18d74f73-0f04-4f3c-88c4-a431d490db9d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:03:20 crc kubenswrapper[4824]: I1211 02:03:20.113146 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-lpgf5" event={"ID":"69814666-b43a-4cab-9864-864862277916","Type":"ContainerStarted","Data":"4e59d7c639bef8858829f61b2936575793f5e97fab8e57dd95a24cc79e962150"} Dec 11 02:03:20 crc kubenswrapper[4824]: I1211 02:03:20.115469 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-gp8dt" event={"ID":"c70805eb-e095-4462-a202-e01860559471","Type":"ContainerStarted","Data":"68987c1734849800583b84780107abe140aef558da5e1f6aca7dafb2264ecd53"} Dec 11 02:03:20 crc kubenswrapper[4824]: I1211 02:03:20.102327 4824 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-z42g5 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.20:8080/healthz\": dial tcp 10.217.0.20:8080: connect: connection refused" start-of-body= Dec 11 02:03:20 crc kubenswrapper[4824]: I1211 02:03:20.121327 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-z42g5" podUID="ffb4acd4-543f-4509-b993-249e3656048d" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.20:8080/healthz\": dial tcp 10.217.0.20:8080: connect: connection refused" Dec 11 02:03:20 crc kubenswrapper[4824]: I1211 02:03:20.124965 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-d89dq" podStartSLOduration=123.124947469 podStartE2EDuration="2m3.124947469s" podCreationTimestamp="2025-12-11 02:01:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:03:20.046999638 +0000 UTC m=+141.736037027" watchObservedRunningTime="2025-12-11 02:03:20.124947469 +0000 UTC m=+141.813984848" Dec 11 02:03:20 crc kubenswrapper[4824]: I1211 02:03:20.125195 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29423640-42vd6" podStartSLOduration=122.125191186 podStartE2EDuration="2m2.125191186s" podCreationTimestamp="2025-12-11 02:01:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:03:20.121725672 +0000 UTC m=+141.810763051" watchObservedRunningTime="2025-12-11 02:03:20.125191186 +0000 UTC m=+141.814228565" Dec 11 02:03:20 crc kubenswrapper[4824]: I1211 02:03:20.150527 4824 generic.go:334] "Generic (PLEG): container finished" podID="f5149a23-92a4-42e6-8b20-8411d1d0536f" containerID="9dbbdb970e50e5ee155d702f28fc4d369baedd4f51ff81ec4d1181ef4c039221" exitCode=0 Dec 11 02:03:20 crc kubenswrapper[4824]: I1211 02:03:20.150760 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-w6b4l" event={"ID":"f5149a23-92a4-42e6-8b20-8411d1d0536f","Type":"ContainerDied","Data":"9dbbdb970e50e5ee155d702f28fc4d369baedd4f51ff81ec4d1181ef4c039221"} Dec 11 02:03:20 crc kubenswrapper[4824]: I1211 02:03:20.190732 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-pg8rn" podStartSLOduration=123.176479006 podStartE2EDuration="2m3.176479006s" podCreationTimestamp="2025-12-11 02:01:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:03:20.175200658 +0000 UTC m=+141.864238037" watchObservedRunningTime="2025-12-11 02:03:20.176479006 +0000 UTC m=+141.865516385" Dec 11 02:03:20 crc kubenswrapper[4824]: I1211 02:03:20.192143 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-66mv8" event={"ID":"398ffffe-110f-4f78-a784-2ee501ee58a3","Type":"ContainerStarted","Data":"e07101f3d41eb50abf22105a1d567ffb58f963fb56b1790f428da9c4213fcbc1"} Dec 11 02:03:20 crc kubenswrapper[4824]: I1211 02:03:20.192692 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-66mv8" Dec 11 02:03:20 crc kubenswrapper[4824]: I1211 02:03:20.193548 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 02:03:20 crc kubenswrapper[4824]: E1211 02:03:20.195218 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 02:03:20.695196968 +0000 UTC m=+142.384234347 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:03:20 crc kubenswrapper[4824]: I1211 02:03:20.204080 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-qp2jl" event={"ID":"1c402f2f-f79a-4a1a-bf40-07da4223059e","Type":"ContainerStarted","Data":"677d9dc615d78dbebb539e50763f4887aa447fdf0a1d9275eb48c62f89b8ef1e"} Dec 11 02:03:20 crc kubenswrapper[4824]: I1211 02:03:20.229387 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vkdbw" event={"ID":"505f6458-5d5a-4725-a529-1075770afa95","Type":"ContainerStarted","Data":"f0b8f2b556ca946063909e233d0a601b250e6117bd5fd8f5e741e5736c3f10f4"} Dec 11 02:03:20 crc kubenswrapper[4824]: I1211 02:03:20.229863 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vkdbw" event={"ID":"505f6458-5d5a-4725-a529-1075770afa95","Type":"ContainerStarted","Data":"d1cae587ff5c21ff866193ca96dc3884e82c2716e51a821945bc26c54ba13813"} Dec 11 02:03:20 crc kubenswrapper[4824]: I1211 02:03:20.244303 4824 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-66mv8 container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.34:8443/healthz\": dial tcp 10.217.0.34:8443: connect: connection refused" start-of-body= Dec 11 02:03:20 crc kubenswrapper[4824]: I1211 02:03:20.244364 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-66mv8" podUID="398ffffe-110f-4f78-a784-2ee501ee58a3" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.34:8443/healthz\": dial tcp 10.217.0.34:8443: connect: connection refused" Dec 11 02:03:20 crc kubenswrapper[4824]: I1211 02:03:20.245611 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-z42g5" podStartSLOduration=123.245601702 podStartE2EDuration="2m3.245601702s" podCreationTimestamp="2025-12-11 02:01:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:03:20.229490708 +0000 UTC m=+141.918528087" watchObservedRunningTime="2025-12-11 02:03:20.245601702 +0000 UTC m=+141.934639081" Dec 11 02:03:20 crc kubenswrapper[4824]: I1211 02:03:20.277264 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-krqhc" event={"ID":"8e2095d8-25c6-407f-aadb-f4603ae44d15","Type":"ContainerStarted","Data":"9ae04f9c275025207d028b464f6907319b74aad5f9bffbc323ed836eafadb66e"} Dec 11 02:03:20 crc kubenswrapper[4824]: I1211 02:03:20.277309 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-krqhc" event={"ID":"8e2095d8-25c6-407f-aadb-f4603ae44d15","Type":"ContainerStarted","Data":"c24d602951e5bb83757effd6a511d8204e88efeb8daa39a8a913d225f5f7d490"} Dec 11 02:03:20 crc kubenswrapper[4824]: I1211 02:03:20.291486 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-gp8dt" podStartSLOduration=123.291469949 podStartE2EDuration="2m3.291469949s" podCreationTimestamp="2025-12-11 02:01:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:03:20.290780538 +0000 UTC m=+141.979817917" watchObservedRunningTime="2025-12-11 02:03:20.291469949 +0000 UTC m=+141.980507328" Dec 11 02:03:20 crc kubenswrapper[4824]: I1211 02:03:20.294857 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-254w2\" (UID: \"18d74f73-0f04-4f3c-88c4-a431d490db9d\") " pod="openshift-image-registry/image-registry-697d97f7c8-254w2" Dec 11 02:03:20 crc kubenswrapper[4824]: E1211 02:03:20.295143 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 02:03:20.795130808 +0000 UTC m=+142.484168187 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-254w2" (UID: "18d74f73-0f04-4f3c-88c4-a431d490db9d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:03:20 crc kubenswrapper[4824]: I1211 02:03:20.336399 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-pg8rn" Dec 11 02:03:20 crc kubenswrapper[4824]: I1211 02:03:20.341000 4824 patch_prober.go:28] interesting pod/router-default-5444994796-pg8rn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 11 02:03:20 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Dec 11 02:03:20 crc kubenswrapper[4824]: [+]process-running ok Dec 11 02:03:20 crc kubenswrapper[4824]: healthz check failed Dec 11 02:03:20 crc kubenswrapper[4824]: I1211 02:03:20.341231 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pg8rn" podUID="877d4a7f-46b0-4f9c-a702-b6b83620c3cb" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 11 02:03:20 crc kubenswrapper[4824]: I1211 02:03:20.354944 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-qtll9" event={"ID":"a4f0d12d-4a03-4ffd-9058-f311bf0041fe","Type":"ContainerStarted","Data":"de6be3f8bc74602e39c7477db1b3c7e5625bc21f14ef7854b05eaa940b161e4a"} Dec 11 02:03:20 crc kubenswrapper[4824]: I1211 02:03:20.389216 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g692d" Dec 11 02:03:20 crc kubenswrapper[4824]: I1211 02:03:20.389643 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g692d" Dec 11 02:03:20 crc kubenswrapper[4824]: I1211 02:03:20.402186 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 02:03:20 crc kubenswrapper[4824]: E1211 02:03:20.402592 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 02:03:20.902569955 +0000 UTC m=+142.591607334 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:03:20 crc kubenswrapper[4824]: I1211 02:03:20.413036 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-lfxkq" event={"ID":"bb4dde0f-222d-4609-b1c9-fed653668e3a","Type":"ContainerStarted","Data":"69116b65bc90ffa19352f33f5cb2ec7b0968ac3382e0de3cf418c6899f74192f"} Dec 11 02:03:20 crc kubenswrapper[4824]: I1211 02:03:20.413582 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-lfxkq" Dec 11 02:03:20 crc kubenswrapper[4824]: I1211 02:03:20.441378 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-tz2j9" podStartSLOduration=123.4413601 podStartE2EDuration="2m3.4413601s" podCreationTimestamp="2025-12-11 02:01:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:03:20.429659219 +0000 UTC m=+142.118696588" watchObservedRunningTime="2025-12-11 02:03:20.4413601 +0000 UTC m=+142.130397479" Dec 11 02:03:20 crc kubenswrapper[4824]: I1211 02:03:20.457457 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-lfxkq" Dec 11 02:03:20 crc kubenswrapper[4824]: I1211 02:03:20.478937 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p8x2v" event={"ID":"59dd9dac-e58a-4869-b1ee-e0c5e8a78974","Type":"ContainerStarted","Data":"77efe92b2e03afc63d2e19bdfa2eb09d2228ba4acb506370e0a9f771818d1fe8"} Dec 11 02:03:20 crc kubenswrapper[4824]: I1211 02:03:20.479005 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p8x2v" event={"ID":"59dd9dac-e58a-4869-b1ee-e0c5e8a78974","Type":"ContainerStarted","Data":"2c1e3146b63611eda444b9268d5fcc3ff007071c0845bbc2c7785a4c731f5a5c"} Dec 11 02:03:20 crc kubenswrapper[4824]: I1211 02:03:20.480143 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p8x2v" Dec 11 02:03:20 crc kubenswrapper[4824]: I1211 02:03:20.509457 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-jlqk5" event={"ID":"04cb3d8f-ccf2-410c-b51c-445add07699c","Type":"ContainerStarted","Data":"9df34f10a0e0f4881a29d271546d3b0bb1ac3b2027a91d909e09352ce21c3a42"} Dec 11 02:03:20 crc kubenswrapper[4824]: I1211 02:03:20.514999 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-254w2\" (UID: \"18d74f73-0f04-4f3c-88c4-a431d490db9d\") " pod="openshift-image-registry/image-registry-697d97f7c8-254w2" Dec 11 02:03:20 crc kubenswrapper[4824]: E1211 02:03:20.515960 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 02:03:21.015944009 +0000 UTC m=+142.704981388 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-254w2" (UID: "18d74f73-0f04-4f3c-88c4-a431d490db9d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:03:20 crc kubenswrapper[4824]: I1211 02:03:20.521927 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-krqhc" podStartSLOduration=123.521908228 podStartE2EDuration="2m3.521908228s" podCreationTimestamp="2025-12-11 02:01:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:03:20.479386611 +0000 UTC m=+142.168423990" watchObservedRunningTime="2025-12-11 02:03:20.521908228 +0000 UTC m=+142.210945597" Dec 11 02:03:20 crc kubenswrapper[4824]: I1211 02:03:20.522841 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-lfxkq" podStartSLOduration=123.522834566 podStartE2EDuration="2m3.522834566s" podCreationTimestamp="2025-12-11 02:01:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:03:20.521792144 +0000 UTC m=+142.210829513" watchObservedRunningTime="2025-12-11 02:03:20.522834566 +0000 UTC m=+142.211871945" Dec 11 02:03:20 crc kubenswrapper[4824]: I1211 02:03:20.552706 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-66mv8" podStartSLOduration=123.552687882 podStartE2EDuration="2m3.552687882s" podCreationTimestamp="2025-12-11 02:01:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:03:20.552394913 +0000 UTC m=+142.241432292" watchObservedRunningTime="2025-12-11 02:03:20.552687882 +0000 UTC m=+142.241725261" Dec 11 02:03:20 crc kubenswrapper[4824]: I1211 02:03:20.556298 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kdzkc" event={"ID":"dccdfcce-6b35-40e6-8bdc-7eed59518f24","Type":"ContainerStarted","Data":"dd14e9d3a45f086cd6313e3f7cd6e923a8721f2413f5b883cdf7ad7bfbbb1645"} Dec 11 02:03:20 crc kubenswrapper[4824]: I1211 02:03:20.582134 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vkdbw" podStartSLOduration=123.582096416 podStartE2EDuration="2m3.582096416s" podCreationTimestamp="2025-12-11 02:01:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:03:20.581649152 +0000 UTC m=+142.270686531" watchObservedRunningTime="2025-12-11 02:03:20.582096416 +0000 UTC m=+142.271133795" Dec 11 02:03:20 crc kubenswrapper[4824]: I1211 02:03:20.583335 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qqg25" event={"ID":"56378af0-da53-4ffc-a66c-a381083e1ba6","Type":"ContainerStarted","Data":"f4544579f0e3ff059a4fbce5b50b70c19fcf6767d53753cfaf0f79c4f22c6c66"} Dec 11 02:03:20 crc kubenswrapper[4824]: I1211 02:03:20.583380 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qqg25" event={"ID":"56378af0-da53-4ffc-a66c-a381083e1ba6","Type":"ContainerStarted","Data":"18d6361d70a9d427d7f7187b42d1125e1ad87e42141c830e9ed7d634728e417e"} Dec 11 02:03:20 crc kubenswrapper[4824]: I1211 02:03:20.584243 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qqg25" Dec 11 02:03:20 crc kubenswrapper[4824]: I1211 02:03:20.612615 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-59bgf" event={"ID":"09112ee7-18d2-4a6d-83ce-eddfce56b083","Type":"ContainerStarted","Data":"e2368fd35335a17c7537ff33fd743e0eaa35c854b12d977ebf6bc48e4808b8d5"} Dec 11 02:03:20 crc kubenswrapper[4824]: I1211 02:03:20.612666 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-59bgf" event={"ID":"09112ee7-18d2-4a6d-83ce-eddfce56b083","Type":"ContainerStarted","Data":"3e5810b93c0bb25259966f9e1e34d80dde1d2b50c741d81584ae411114996d55"} Dec 11 02:03:20 crc kubenswrapper[4824]: I1211 02:03:20.615985 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 02:03:20 crc kubenswrapper[4824]: E1211 02:03:20.616158 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 02:03:21.116090256 +0000 UTC m=+142.805127635 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:03:20 crc kubenswrapper[4824]: I1211 02:03:20.616458 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-254w2\" (UID: \"18d74f73-0f04-4f3c-88c4-a431d490db9d\") " pod="openshift-image-registry/image-registry-697d97f7c8-254w2" Dec 11 02:03:20 crc kubenswrapper[4824]: E1211 02:03:20.617052 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 02:03:21.117031884 +0000 UTC m=+142.806069263 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-254w2" (UID: "18d74f73-0f04-4f3c-88c4-a431d490db9d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:03:20 crc kubenswrapper[4824]: I1211 02:03:20.632429 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-k6kls" event={"ID":"ad510a18-cd72-4cf0-9b0c-dd3540d491fa","Type":"ContainerStarted","Data":"c7eb4f774e5a9e63b3626818ebe56f13753bed99b12c3e95972fb1b6876fc303"} Dec 11 02:03:20 crc kubenswrapper[4824]: I1211 02:03:20.665952 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qqg25" Dec 11 02:03:20 crc kubenswrapper[4824]: I1211 02:03:20.670537 4824 generic.go:334] "Generic (PLEG): container finished" podID="75da7b9a-7085-492b-954a-6bb920fbb186" containerID="520b201096449f33d7e7045d8ae412fe38a72aacf68358da316186cab9e996bb" exitCode=0 Dec 11 02:03:20 crc kubenswrapper[4824]: I1211 02:03:20.670659 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-l88b4" event={"ID":"75da7b9a-7085-492b-954a-6bb920fbb186","Type":"ContainerDied","Data":"520b201096449f33d7e7045d8ae412fe38a72aacf68358da316186cab9e996bb"} Dec 11 02:03:20 crc kubenswrapper[4824]: I1211 02:03:20.677584 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-hr4qs" event={"ID":"335b007e-ea5c-4e20-9402-5599da94ef22","Type":"ContainerStarted","Data":"e57e311992557d5193dfdc003184eb256e9222aa65595f25e084c0361844d23a"} Dec 11 02:03:20 crc kubenswrapper[4824]: I1211 02:03:20.704720 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kdzkc" podStartSLOduration=123.704702387 podStartE2EDuration="2m3.704702387s" podCreationTimestamp="2025-12-11 02:01:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:03:20.703566313 +0000 UTC m=+142.392603692" watchObservedRunningTime="2025-12-11 02:03:20.704702387 +0000 UTC m=+142.393739766" Dec 11 02:03:20 crc kubenswrapper[4824]: I1211 02:03:20.706011 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p8x2v" podStartSLOduration=123.706006736 podStartE2EDuration="2m3.706006736s" podCreationTimestamp="2025-12-11 02:01:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:03:20.647409087 +0000 UTC m=+142.336446466" watchObservedRunningTime="2025-12-11 02:03:20.706006736 +0000 UTC m=+142.395044115" Dec 11 02:03:20 crc kubenswrapper[4824]: I1211 02:03:20.707222 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hxqw8" event={"ID":"fb52538d-050a-4b69-91b2-aa48f9833fe7","Type":"ContainerStarted","Data":"893f903fbbf3a8fed690a4b8c02913601a405f1dda5a86e5b2afb1189e341685"} Dec 11 02:03:20 crc kubenswrapper[4824]: I1211 02:03:20.718554 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 02:03:20 crc kubenswrapper[4824]: E1211 02:03:20.720591 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 02:03:21.220569683 +0000 UTC m=+142.909607062 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:03:20 crc kubenswrapper[4824]: I1211 02:03:20.730452 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8fksd" event={"ID":"95b90361-2f73-4649-bc1a-1202ba511baf","Type":"ContainerStarted","Data":"a00aa0c146963ad61e23315d869ab7387c79dc574c7e879d0dc2329fc0922c1e"} Dec 11 02:03:20 crc kubenswrapper[4824]: I1211 02:03:20.730495 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8fksd" event={"ID":"95b90361-2f73-4649-bc1a-1202ba511baf","Type":"ContainerStarted","Data":"80ffe2ed32aa22f9def0dd22147654c34754bb9f89e1ce07e456ebaaae6afe7f"} Dec 11 02:03:20 crc kubenswrapper[4824]: I1211 02:03:20.742903 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-jlqk5" podStartSLOduration=123.742888263 podStartE2EDuration="2m3.742888263s" podCreationTimestamp="2025-12-11 02:01:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:03:20.738133791 +0000 UTC m=+142.427171170" watchObservedRunningTime="2025-12-11 02:03:20.742888263 +0000 UTC m=+142.431925642" Dec 11 02:03:20 crc kubenswrapper[4824]: I1211 02:03:20.743664 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ks7hm" event={"ID":"fa007484-938d-460e-9c5e-c3f5b31a25d2","Type":"ContainerStarted","Data":"4780cfed5b3a7364131ced21f9311553a63651c0d76bce414c7f93e5c2d079ea"} Dec 11 02:03:20 crc kubenswrapper[4824]: I1211 02:03:20.750325 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-zvcj8" event={"ID":"63fe488c-de20-435f-83fd-47a4179341e6","Type":"ContainerStarted","Data":"dbf998c2e5b8f20f9592b1ac565d98a8cb679be3a283a0e2eb492f91ada0bb03"} Dec 11 02:03:20 crc kubenswrapper[4824]: I1211 02:03:20.750366 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-zvcj8" event={"ID":"63fe488c-de20-435f-83fd-47a4179341e6","Type":"ContainerStarted","Data":"c9357b482dc8ccb15708b54330fe01c953402502a002cbc0be43642da84298c1"} Dec 11 02:03:20 crc kubenswrapper[4824]: I1211 02:03:20.753310 4824 patch_prober.go:28] interesting pod/downloads-7954f5f757-2ng4n container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Dec 11 02:03:20 crc kubenswrapper[4824]: I1211 02:03:20.753368 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-2ng4n" podUID="f185e5b2-54c0-44ef-bdda-118efa89cb5f" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Dec 11 02:03:20 crc kubenswrapper[4824]: I1211 02:03:20.787510 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-l9d4f" Dec 11 02:03:20 crc kubenswrapper[4824]: I1211 02:03:20.790685 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qqg25" podStartSLOduration=123.790675478 podStartE2EDuration="2m3.790675478s" podCreationTimestamp="2025-12-11 02:01:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:03:20.783666598 +0000 UTC m=+142.472703977" watchObservedRunningTime="2025-12-11 02:03:20.790675478 +0000 UTC m=+142.479712857" Dec 11 02:03:20 crc kubenswrapper[4824]: I1211 02:03:20.829277 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-254w2\" (UID: \"18d74f73-0f04-4f3c-88c4-a431d490db9d\") " pod="openshift-image-registry/image-registry-697d97f7c8-254w2" Dec 11 02:03:20 crc kubenswrapper[4824]: E1211 02:03:20.857020 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 02:03:21.35700422 +0000 UTC m=+143.046041599 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-254w2" (UID: "18d74f73-0f04-4f3c-88c4-a431d490db9d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:03:20 crc kubenswrapper[4824]: I1211 02:03:20.886024 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8fksd" podStartSLOduration=123.886002241 podStartE2EDuration="2m3.886002241s" podCreationTimestamp="2025-12-11 02:01:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:03:20.864778254 +0000 UTC m=+142.553815633" watchObservedRunningTime="2025-12-11 02:03:20.886002241 +0000 UTC m=+142.575039620" Dec 11 02:03:20 crc kubenswrapper[4824]: I1211 02:03:20.940132 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 02:03:20 crc kubenswrapper[4824]: E1211 02:03:20.940703 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 02:03:21.440686833 +0000 UTC m=+143.129724212 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:03:21 crc kubenswrapper[4824]: I1211 02:03:21.014799 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hxqw8" podStartSLOduration=124.014782238 podStartE2EDuration="2m4.014782238s" podCreationTimestamp="2025-12-11 02:01:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:03:20.936728085 +0000 UTC m=+142.625765464" watchObservedRunningTime="2025-12-11 02:03:21.014782238 +0000 UTC m=+142.703819617" Dec 11 02:03:21 crc kubenswrapper[4824]: I1211 02:03:21.042970 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-254w2\" (UID: \"18d74f73-0f04-4f3c-88c4-a431d490db9d\") " pod="openshift-image-registry/image-registry-697d97f7c8-254w2" Dec 11 02:03:21 crc kubenswrapper[4824]: E1211 02:03:21.046417 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 02:03:21.546401278 +0000 UTC m=+143.235438657 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-254w2" (UID: "18d74f73-0f04-4f3c-88c4-a431d490db9d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:03:21 crc kubenswrapper[4824]: I1211 02:03:21.061742 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-59bgf" podStartSLOduration=124.061726098 podStartE2EDuration="2m4.061726098s" podCreationTimestamp="2025-12-11 02:01:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:03:21.061679606 +0000 UTC m=+142.750716985" watchObservedRunningTime="2025-12-11 02:03:21.061726098 +0000 UTC m=+142.750763467" Dec 11 02:03:21 crc kubenswrapper[4824]: I1211 02:03:21.138225 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-k6kls" podStartSLOduration=124.138204114 podStartE2EDuration="2m4.138204114s" podCreationTimestamp="2025-12-11 02:01:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:03:21.109450561 +0000 UTC m=+142.798487940" watchObservedRunningTime="2025-12-11 02:03:21.138204114 +0000 UTC m=+142.827241493" Dec 11 02:03:21 crc kubenswrapper[4824]: I1211 02:03:21.140439 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-zvcj8" podStartSLOduration=8.140429061 podStartE2EDuration="8.140429061s" podCreationTimestamp="2025-12-11 02:03:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:03:21.136688539 +0000 UTC m=+142.825725918" watchObservedRunningTime="2025-12-11 02:03:21.140429061 +0000 UTC m=+142.829466440" Dec 11 02:03:21 crc kubenswrapper[4824]: I1211 02:03:21.144614 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 02:03:21 crc kubenswrapper[4824]: E1211 02:03:21.144880 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 02:03:21.644849694 +0000 UTC m=+143.333887073 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:03:21 crc kubenswrapper[4824]: I1211 02:03:21.144948 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-254w2\" (UID: \"18d74f73-0f04-4f3c-88c4-a431d490db9d\") " pod="openshift-image-registry/image-registry-697d97f7c8-254w2" Dec 11 02:03:21 crc kubenswrapper[4824]: E1211 02:03:21.145451 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 02:03:21.645444902 +0000 UTC m=+143.334482281 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-254w2" (UID: "18d74f73-0f04-4f3c-88c4-a431d490db9d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:03:21 crc kubenswrapper[4824]: I1211 02:03:21.228522 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p8x2v" Dec 11 02:03:21 crc kubenswrapper[4824]: I1211 02:03:21.251572 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 02:03:21 crc kubenswrapper[4824]: E1211 02:03:21.251978 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 02:03:21.7519638 +0000 UTC m=+143.441001179 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:03:21 crc kubenswrapper[4824]: I1211 02:03:21.295353 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ks7hm" podStartSLOduration=124.295309842 podStartE2EDuration="2m4.295309842s" podCreationTimestamp="2025-12-11 02:01:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:03:21.212420143 +0000 UTC m=+142.901457552" watchObservedRunningTime="2025-12-11 02:03:21.295309842 +0000 UTC m=+142.984347381" Dec 11 02:03:21 crc kubenswrapper[4824]: I1211 02:03:21.312607 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g692d" Dec 11 02:03:21 crc kubenswrapper[4824]: I1211 02:03:21.338283 4824 patch_prober.go:28] interesting pod/router-default-5444994796-pg8rn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 11 02:03:21 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Dec 11 02:03:21 crc kubenswrapper[4824]: [+]process-running ok Dec 11 02:03:21 crc kubenswrapper[4824]: healthz check failed Dec 11 02:03:21 crc kubenswrapper[4824]: I1211 02:03:21.338340 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pg8rn" podUID="877d4a7f-46b0-4f9c-a702-b6b83620c3cb" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 11 02:03:21 crc kubenswrapper[4824]: I1211 02:03:21.353229 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-254w2\" (UID: \"18d74f73-0f04-4f3c-88c4-a431d490db9d\") " pod="openshift-image-registry/image-registry-697d97f7c8-254w2" Dec 11 02:03:21 crc kubenswrapper[4824]: E1211 02:03:21.353724 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 02:03:21.853707435 +0000 UTC m=+143.542744814 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-254w2" (UID: "18d74f73-0f04-4f3c-88c4-a431d490db9d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:03:21 crc kubenswrapper[4824]: I1211 02:03:21.454522 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 02:03:21 crc kubenswrapper[4824]: E1211 02:03:21.454944 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 02:03:21.954928024 +0000 UTC m=+143.643965403 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:03:21 crc kubenswrapper[4824]: I1211 02:03:21.555732 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-254w2\" (UID: \"18d74f73-0f04-4f3c-88c4-a431d490db9d\") " pod="openshift-image-registry/image-registry-697d97f7c8-254w2" Dec 11 02:03:21 crc kubenswrapper[4824]: E1211 02:03:21.556363 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 02:03:22.05635235 +0000 UTC m=+143.745389729 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-254w2" (UID: "18d74f73-0f04-4f3c-88c4-a431d490db9d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:03:21 crc kubenswrapper[4824]: I1211 02:03:21.657662 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 02:03:21 crc kubenswrapper[4824]: E1211 02:03:21.658032 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 02:03:22.158017563 +0000 UTC m=+143.847054942 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:03:21 crc kubenswrapper[4824]: I1211 02:03:21.759223 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-254w2\" (UID: \"18d74f73-0f04-4f3c-88c4-a431d490db9d\") " pod="openshift-image-registry/image-registry-697d97f7c8-254w2" Dec 11 02:03:21 crc kubenswrapper[4824]: E1211 02:03:21.759532 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 02:03:22.2595187 +0000 UTC m=+143.948556079 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-254w2" (UID: "18d74f73-0f04-4f3c-88c4-a431d490db9d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:03:21 crc kubenswrapper[4824]: I1211 02:03:21.798662 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-lpgf5" event={"ID":"69814666-b43a-4cab-9864-864862277916","Type":"ContainerStarted","Data":"75353b87130fa9faad21cc1544748785355a91fc8719b009128e0b4123cbc95e"} Dec 11 02:03:21 crc kubenswrapper[4824]: I1211 02:03:21.798708 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-lpgf5" event={"ID":"69814666-b43a-4cab-9864-864862277916","Type":"ContainerStarted","Data":"55952cca48b3d24ddb75f69578527497fb9bcac7b432106bb43f1e08a4d6fa24"} Dec 11 02:03:21 crc kubenswrapper[4824]: I1211 02:03:21.799180 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-lpgf5" Dec 11 02:03:21 crc kubenswrapper[4824]: I1211 02:03:21.824048 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-gp8dt" event={"ID":"c70805eb-e095-4462-a202-e01860559471","Type":"ContainerStarted","Data":"1ac72fc3c135fe5ebed110ab3371f874c8a5d3a77ca1827f45c956ac56e4fdd7"} Dec 11 02:03:21 crc kubenswrapper[4824]: I1211 02:03:21.847631 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vkdbw" event={"ID":"505f6458-5d5a-4725-a529-1075770afa95","Type":"ContainerStarted","Data":"76173b0c79ac35f860bc29619e079559e590a1c29818976dcbccd1dc28ad4b05"} Dec 11 02:03:21 crc kubenswrapper[4824]: I1211 02:03:21.858637 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-lpgf5" podStartSLOduration=124.858618897 podStartE2EDuration="2m4.858618897s" podCreationTimestamp="2025-12-11 02:01:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:03:21.857011558 +0000 UTC m=+143.546048937" watchObservedRunningTime="2025-12-11 02:03:21.858618897 +0000 UTC m=+143.547656276" Dec 11 02:03:21 crc kubenswrapper[4824]: I1211 02:03:21.860395 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-l88b4" event={"ID":"75da7b9a-7085-492b-954a-6bb920fbb186","Type":"ContainerStarted","Data":"a56466b5337e17cfae8b8f62323ab85697b5ade2caf5ef91875accc745c05a89"} Dec 11 02:03:21 crc kubenswrapper[4824]: I1211 02:03:21.861549 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 02:03:21 crc kubenswrapper[4824]: E1211 02:03:21.862652 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 02:03:22.362629417 +0000 UTC m=+144.051666796 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:03:21 crc kubenswrapper[4824]: I1211 02:03:21.881409 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-qtll9" event={"ID":"a4f0d12d-4a03-4ffd-9058-f311bf0041fe","Type":"ContainerStarted","Data":"939df0b72cced386171fa4ad8dfcc46a446acdc0d0df7b96792c21388e6faeb5"} Dec 11 02:03:21 crc kubenswrapper[4824]: I1211 02:03:21.881459 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-qtll9" event={"ID":"a4f0d12d-4a03-4ffd-9058-f311bf0041fe","Type":"ContainerStarted","Data":"8d28e2b0f7e4576034c2df18bd7224198c395ac81cd5be620757b39c17846569"} Dec 11 02:03:21 crc kubenswrapper[4824]: I1211 02:03:21.882791 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-qtll9" Dec 11 02:03:21 crc kubenswrapper[4824]: I1211 02:03:21.913066 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-hr4qs" event={"ID":"335b007e-ea5c-4e20-9402-5599da94ef22","Type":"ContainerStarted","Data":"1362c3da39adffd3e314ae3d32112c241a85eb0ce8ce008aa413dcc6b21b69c9"} Dec 11 02:03:21 crc kubenswrapper[4824]: I1211 02:03:21.955308 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-kkw8v" event={"ID":"8650bbf0-78d2-410e-a62d-d4adf673cb55","Type":"ContainerStarted","Data":"0f3e570c58153964bb5f310fc50c7fd2d156aaa8264a067f5fcb04fcfec64e10"} Dec 11 02:03:21 crc kubenswrapper[4824]: I1211 02:03:21.967021 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-254w2\" (UID: \"18d74f73-0f04-4f3c-88c4-a431d490db9d\") " pod="openshift-image-registry/image-registry-697d97f7c8-254w2" Dec 11 02:03:21 crc kubenswrapper[4824]: E1211 02:03:21.967612 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 02:03:22.467593698 +0000 UTC m=+144.156631077 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-254w2" (UID: "18d74f73-0f04-4f3c-88c4-a431d490db9d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:03:21 crc kubenswrapper[4824]: I1211 02:03:21.988521 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-qp2jl" event={"ID":"1c402f2f-f79a-4a1a-bf40-07da4223059e","Type":"ContainerStarted","Data":"92da61cf69d6c6b14b05914a250660ef127abb7f34a33836810979abf0e48ff0"} Dec 11 02:03:21 crc kubenswrapper[4824]: I1211 02:03:21.988942 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-qp2jl" event={"ID":"1c402f2f-f79a-4a1a-bf40-07da4223059e","Type":"ContainerStarted","Data":"3ee28d9d116ffa0a6a64e0c78466b928e3114fcc268986e7464ae4643cca727d"} Dec 11 02:03:21 crc kubenswrapper[4824]: I1211 02:03:21.996065 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-qtll9" podStartSLOduration=8.996046613 podStartE2EDuration="8.996046613s" podCreationTimestamp="2025-12-11 02:03:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:03:21.934418713 +0000 UTC m=+143.623456092" watchObservedRunningTime="2025-12-11 02:03:21.996046613 +0000 UTC m=+143.685083992" Dec 11 02:03:21 crc kubenswrapper[4824]: I1211 02:03:21.996463 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-kkw8v" podStartSLOduration=124.996457196 podStartE2EDuration="2m4.996457196s" podCreationTimestamp="2025-12-11 02:01:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:03:21.994543008 +0000 UTC m=+143.683580377" watchObservedRunningTime="2025-12-11 02:03:21.996457196 +0000 UTC m=+143.685494575" Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.002891 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-4r8x8" event={"ID":"09f964e8-942f-4577-983a-bdc08487c0e8","Type":"ContainerStarted","Data":"2d75a8bc149c49b7b3f9ff0e1180ca2a42c945f2ac2b807ce47ca9215e773d16"} Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.023402 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-qp2jl" podStartSLOduration=125.023384323 podStartE2EDuration="2m5.023384323s" podCreationTimestamp="2025-12-11 02:01:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:03:22.021302312 +0000 UTC m=+143.710339691" watchObservedRunningTime="2025-12-11 02:03:22.023384323 +0000 UTC m=+143.712421702" Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.026749 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-w6b4l" event={"ID":"f5149a23-92a4-42e6-8b20-8411d1d0536f","Type":"ContainerStarted","Data":"9b609d4ec17623b886e308b6b21d09741c004ff0fe66dafa8e23ca822758078a"} Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.026852 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-w6b4l" Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.045459 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-66mv8" event={"ID":"398ffffe-110f-4f78-a784-2ee501ee58a3","Type":"ContainerStarted","Data":"91416ea39a50f48a7160371ad7ae0a606fecc71db11b19fa02e6a7059481d926"} Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.050743 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-66mv8" Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.056719 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-jlqk5" event={"ID":"04cb3d8f-ccf2-410c-b51c-445add07699c","Type":"ContainerStarted","Data":"6278e4ac5d0aca922621826e6d5ea0629a619d065a27ca9d6ca626e5ec72fef9"} Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.068101 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-4r8x8" podStartSLOduration=125.068085996 podStartE2EDuration="2m5.068085996s" podCreationTimestamp="2025-12-11 02:01:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:03:22.066642993 +0000 UTC m=+143.755680372" watchObservedRunningTime="2025-12-11 02:03:22.068085996 +0000 UTC m=+143.757123375" Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.068754 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 02:03:22 crc kubenswrapper[4824]: E1211 02:03:22.069025 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 02:03:22.568980273 +0000 UTC m=+144.258017652 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.069658 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-254w2\" (UID: \"18d74f73-0f04-4f3c-88c4-a431d490db9d\") " pod="openshift-image-registry/image-registry-697d97f7c8-254w2" Dec 11 02:03:22 crc kubenswrapper[4824]: E1211 02:03:22.070136 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 02:03:22.570103037 +0000 UTC m=+144.259140636 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-254w2" (UID: "18d74f73-0f04-4f3c-88c4-a431d490db9d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.080675 4824 generic.go:334] "Generic (PLEG): container finished" podID="4c22bf31-ad33-44cd-a495-05d83688f2bd" containerID="f63691527d0e51385a07c889a507a7c4d8c0a08751f37f06bdd5379d9c81e323" exitCode=0 Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.081416 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423640-42vd6" event={"ID":"4c22bf31-ad33-44cd-a495-05d83688f2bd","Type":"ContainerDied","Data":"f63691527d0e51385a07c889a507a7c4d8c0a08751f37f06bdd5379d9c81e323"} Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.083006 4824 patch_prober.go:28] interesting pod/downloads-7954f5f757-2ng4n container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.083069 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-2ng4n" podUID="f185e5b2-54c0-44ef-bdda-118efa89cb5f" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.095463 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g692d" Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.099073 4824 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-z42g5 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.20:8080/healthz\": dial tcp 10.217.0.20:8080: connect: connection refused" start-of-body= Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.127026 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-z42g5" podUID="ffb4acd4-543f-4509-b993-249e3656048d" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.20:8080/healthz\": dial tcp 10.217.0.20:8080: connect: connection refused" Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.142276 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-w6b4l" podStartSLOduration=125.142247473 podStartE2EDuration="2m5.142247473s" podCreationTimestamp="2025-12-11 02:01:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:03:22.098348085 +0000 UTC m=+143.787385464" watchObservedRunningTime="2025-12-11 02:03:22.142247473 +0000 UTC m=+143.831284852" Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.154025 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-nqjh5"] Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.165632 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nqjh5" Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.169901 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.170361 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nqjh5"] Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.171024 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 02:03:22 crc kubenswrapper[4824]: E1211 02:03:22.172776 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 02:03:22.672749899 +0000 UTC m=+144.361787278 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.277085 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9806a3c-cb3d-4b3e-8889-11f8cf49c030-utilities\") pod \"certified-operators-nqjh5\" (UID: \"d9806a3c-cb3d-4b3e-8889-11f8cf49c030\") " pod="openshift-marketplace/certified-operators-nqjh5" Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.277180 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-254w2\" (UID: \"18d74f73-0f04-4f3c-88c4-a431d490db9d\") " pod="openshift-image-registry/image-registry-697d97f7c8-254w2" Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.277204 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4xhtb\" (UniqueName: \"kubernetes.io/projected/d9806a3c-cb3d-4b3e-8889-11f8cf49c030-kube-api-access-4xhtb\") pod \"certified-operators-nqjh5\" (UID: \"d9806a3c-cb3d-4b3e-8889-11f8cf49c030\") " pod="openshift-marketplace/certified-operators-nqjh5" Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.277300 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9806a3c-cb3d-4b3e-8889-11f8cf49c030-catalog-content\") pod \"certified-operators-nqjh5\" (UID: \"d9806a3c-cb3d-4b3e-8889-11f8cf49c030\") " pod="openshift-marketplace/certified-operators-nqjh5" Dec 11 02:03:22 crc kubenswrapper[4824]: E1211 02:03:22.277617 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 02:03:22.777603477 +0000 UTC m=+144.466640856 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-254w2" (UID: "18d74f73-0f04-4f3c-88c4-a431d490db9d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.328067 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-86mf9"] Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.333699 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-86mf9" Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.339930 4824 patch_prober.go:28] interesting pod/router-default-5444994796-pg8rn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 11 02:03:22 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Dec 11 02:03:22 crc kubenswrapper[4824]: [+]process-running ok Dec 11 02:03:22 crc kubenswrapper[4824]: healthz check failed Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.340315 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pg8rn" podUID="877d4a7f-46b0-4f9c-a702-b6b83620c3cb" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.341124 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.347064 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-86mf9"] Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.378820 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.379051 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4xhtb\" (UniqueName: \"kubernetes.io/projected/d9806a3c-cb3d-4b3e-8889-11f8cf49c030-kube-api-access-4xhtb\") pod \"certified-operators-nqjh5\" (UID: \"d9806a3c-cb3d-4b3e-8889-11f8cf49c030\") " pod="openshift-marketplace/certified-operators-nqjh5" Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.379088 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2b7c06f-4350-41be-b472-c75d21558911-catalog-content\") pod \"community-operators-86mf9\" (UID: \"b2b7c06f-4350-41be-b472-c75d21558911\") " pod="openshift-marketplace/community-operators-86mf9" Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.379173 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9806a3c-cb3d-4b3e-8889-11f8cf49c030-catalog-content\") pod \"certified-operators-nqjh5\" (UID: \"d9806a3c-cb3d-4b3e-8889-11f8cf49c030\") " pod="openshift-marketplace/certified-operators-nqjh5" Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.379213 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2b7c06f-4350-41be-b472-c75d21558911-utilities\") pod \"community-operators-86mf9\" (UID: \"b2b7c06f-4350-41be-b472-c75d21558911\") " pod="openshift-marketplace/community-operators-86mf9" Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.379232 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2gdpn\" (UniqueName: \"kubernetes.io/projected/b2b7c06f-4350-41be-b472-c75d21558911-kube-api-access-2gdpn\") pod \"community-operators-86mf9\" (UID: \"b2b7c06f-4350-41be-b472-c75d21558911\") " pod="openshift-marketplace/community-operators-86mf9" Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.379250 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9806a3c-cb3d-4b3e-8889-11f8cf49c030-utilities\") pod \"certified-operators-nqjh5\" (UID: \"d9806a3c-cb3d-4b3e-8889-11f8cf49c030\") " pod="openshift-marketplace/certified-operators-nqjh5" Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.379858 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9806a3c-cb3d-4b3e-8889-11f8cf49c030-utilities\") pod \"certified-operators-nqjh5\" (UID: \"d9806a3c-cb3d-4b3e-8889-11f8cf49c030\") " pod="openshift-marketplace/certified-operators-nqjh5" Dec 11 02:03:22 crc kubenswrapper[4824]: E1211 02:03:22.379925 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 02:03:22.87990999 +0000 UTC m=+144.568947369 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.380396 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9806a3c-cb3d-4b3e-8889-11f8cf49c030-catalog-content\") pod \"certified-operators-nqjh5\" (UID: \"d9806a3c-cb3d-4b3e-8889-11f8cf49c030\") " pod="openshift-marketplace/certified-operators-nqjh5" Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.396405 4824 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.430781 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4xhtb\" (UniqueName: \"kubernetes.io/projected/d9806a3c-cb3d-4b3e-8889-11f8cf49c030-kube-api-access-4xhtb\") pod \"certified-operators-nqjh5\" (UID: \"d9806a3c-cb3d-4b3e-8889-11f8cf49c030\") " pod="openshift-marketplace/certified-operators-nqjh5" Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.479998 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2b7c06f-4350-41be-b472-c75d21558911-utilities\") pod \"community-operators-86mf9\" (UID: \"b2b7c06f-4350-41be-b472-c75d21558911\") " pod="openshift-marketplace/community-operators-86mf9" Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.480033 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2gdpn\" (UniqueName: \"kubernetes.io/projected/b2b7c06f-4350-41be-b472-c75d21558911-kube-api-access-2gdpn\") pod \"community-operators-86mf9\" (UID: \"b2b7c06f-4350-41be-b472-c75d21558911\") " pod="openshift-marketplace/community-operators-86mf9" Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.480074 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-254w2\" (UID: \"18d74f73-0f04-4f3c-88c4-a431d490db9d\") " pod="openshift-image-registry/image-registry-697d97f7c8-254w2" Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.480126 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2b7c06f-4350-41be-b472-c75d21558911-catalog-content\") pod \"community-operators-86mf9\" (UID: \"b2b7c06f-4350-41be-b472-c75d21558911\") " pod="openshift-marketplace/community-operators-86mf9" Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.480468 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2b7c06f-4350-41be-b472-c75d21558911-utilities\") pod \"community-operators-86mf9\" (UID: \"b2b7c06f-4350-41be-b472-c75d21558911\") " pod="openshift-marketplace/community-operators-86mf9" Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.480575 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2b7c06f-4350-41be-b472-c75d21558911-catalog-content\") pod \"community-operators-86mf9\" (UID: \"b2b7c06f-4350-41be-b472-c75d21558911\") " pod="openshift-marketplace/community-operators-86mf9" Dec 11 02:03:22 crc kubenswrapper[4824]: E1211 02:03:22.480732 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 02:03:22.980718487 +0000 UTC m=+144.669755866 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-254w2" (UID: "18d74f73-0f04-4f3c-88c4-a431d490db9d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.500855 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nqjh5" Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.501995 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2gdpn\" (UniqueName: \"kubernetes.io/projected/b2b7c06f-4350-41be-b472-c75d21558911-kube-api-access-2gdpn\") pod \"community-operators-86mf9\" (UID: \"b2b7c06f-4350-41be-b472-c75d21558911\") " pod="openshift-marketplace/community-operators-86mf9" Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.524912 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-ptkck"] Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.525833 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ptkck" Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.544044 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ptkck"] Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.581375 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.581543 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f40e0dc-3513-431e-95b3-af69c5cca852-utilities\") pod \"certified-operators-ptkck\" (UID: \"9f40e0dc-3513-431e-95b3-af69c5cca852\") " pod="openshift-marketplace/certified-operators-ptkck" Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.581578 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t6z6w\" (UniqueName: \"kubernetes.io/projected/9f40e0dc-3513-431e-95b3-af69c5cca852-kube-api-access-t6z6w\") pod \"certified-operators-ptkck\" (UID: \"9f40e0dc-3513-431e-95b3-af69c5cca852\") " pod="openshift-marketplace/certified-operators-ptkck" Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.581629 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f40e0dc-3513-431e-95b3-af69c5cca852-catalog-content\") pod \"certified-operators-ptkck\" (UID: \"9f40e0dc-3513-431e-95b3-af69c5cca852\") " pod="openshift-marketplace/certified-operators-ptkck" Dec 11 02:03:22 crc kubenswrapper[4824]: E1211 02:03:22.581747 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 02:03:23.08173318 +0000 UTC m=+144.770770549 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.657224 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-86mf9" Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.682848 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f40e0dc-3513-431e-95b3-af69c5cca852-utilities\") pod \"certified-operators-ptkck\" (UID: \"9f40e0dc-3513-431e-95b3-af69c5cca852\") " pod="openshift-marketplace/certified-operators-ptkck" Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.682886 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-254w2\" (UID: \"18d74f73-0f04-4f3c-88c4-a431d490db9d\") " pod="openshift-image-registry/image-registry-697d97f7c8-254w2" Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.682906 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t6z6w\" (UniqueName: \"kubernetes.io/projected/9f40e0dc-3513-431e-95b3-af69c5cca852-kube-api-access-t6z6w\") pod \"certified-operators-ptkck\" (UID: \"9f40e0dc-3513-431e-95b3-af69c5cca852\") " pod="openshift-marketplace/certified-operators-ptkck" Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.682954 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f40e0dc-3513-431e-95b3-af69c5cca852-catalog-content\") pod \"certified-operators-ptkck\" (UID: \"9f40e0dc-3513-431e-95b3-af69c5cca852\") " pod="openshift-marketplace/certified-operators-ptkck" Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.683680 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f40e0dc-3513-431e-95b3-af69c5cca852-catalog-content\") pod \"certified-operators-ptkck\" (UID: \"9f40e0dc-3513-431e-95b3-af69c5cca852\") " pod="openshift-marketplace/certified-operators-ptkck" Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.683871 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f40e0dc-3513-431e-95b3-af69c5cca852-utilities\") pod \"certified-operators-ptkck\" (UID: \"9f40e0dc-3513-431e-95b3-af69c5cca852\") " pod="openshift-marketplace/certified-operators-ptkck" Dec 11 02:03:22 crc kubenswrapper[4824]: E1211 02:03:22.684126 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 02:03:23.184101983 +0000 UTC m=+144.873139362 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-254w2" (UID: "18d74f73-0f04-4f3c-88c4-a431d490db9d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.715541 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t6z6w\" (UniqueName: \"kubernetes.io/projected/9f40e0dc-3513-431e-95b3-af69c5cca852-kube-api-access-t6z6w\") pod \"certified-operators-ptkck\" (UID: \"9f40e0dc-3513-431e-95b3-af69c5cca852\") " pod="openshift-marketplace/certified-operators-ptkck" Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.715635 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zfgl8"] Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.716880 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zfgl8" Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.731909 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zfgl8"] Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.786998 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.787236 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lb5d7\" (UniqueName: \"kubernetes.io/projected/a9bd6e85-2ed9-4525-9308-f6dd2f4eacda-kube-api-access-lb5d7\") pod \"community-operators-zfgl8\" (UID: \"a9bd6e85-2ed9-4525-9308-f6dd2f4eacda\") " pod="openshift-marketplace/community-operators-zfgl8" Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.787303 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9bd6e85-2ed9-4525-9308-f6dd2f4eacda-catalog-content\") pod \"community-operators-zfgl8\" (UID: \"a9bd6e85-2ed9-4525-9308-f6dd2f4eacda\") " pod="openshift-marketplace/community-operators-zfgl8" Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.787352 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9bd6e85-2ed9-4525-9308-f6dd2f4eacda-utilities\") pod \"community-operators-zfgl8\" (UID: \"a9bd6e85-2ed9-4525-9308-f6dd2f4eacda\") " pod="openshift-marketplace/community-operators-zfgl8" Dec 11 02:03:22 crc kubenswrapper[4824]: E1211 02:03:22.787459 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 02:03:23.287443246 +0000 UTC m=+144.976480625 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.788088 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nqjh5"] Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.816453 4824 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-12-11T02:03:22.396432286Z","Handler":null,"Name":""} Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.851284 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ptkck" Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.856255 4824 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.856284 4824 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.888694 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9bd6e85-2ed9-4525-9308-f6dd2f4eacda-utilities\") pod \"community-operators-zfgl8\" (UID: \"a9bd6e85-2ed9-4525-9308-f6dd2f4eacda\") " pod="openshift-marketplace/community-operators-zfgl8" Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.888762 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lb5d7\" (UniqueName: \"kubernetes.io/projected/a9bd6e85-2ed9-4525-9308-f6dd2f4eacda-kube-api-access-lb5d7\") pod \"community-operators-zfgl8\" (UID: \"a9bd6e85-2ed9-4525-9308-f6dd2f4eacda\") " pod="openshift-marketplace/community-operators-zfgl8" Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.888813 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-254w2\" (UID: \"18d74f73-0f04-4f3c-88c4-a431d490db9d\") " pod="openshift-image-registry/image-registry-697d97f7c8-254w2" Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.888838 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9bd6e85-2ed9-4525-9308-f6dd2f4eacda-catalog-content\") pod \"community-operators-zfgl8\" (UID: \"a9bd6e85-2ed9-4525-9308-f6dd2f4eacda\") " pod="openshift-marketplace/community-operators-zfgl8" Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.889227 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9bd6e85-2ed9-4525-9308-f6dd2f4eacda-catalog-content\") pod \"community-operators-zfgl8\" (UID: \"a9bd6e85-2ed9-4525-9308-f6dd2f4eacda\") " pod="openshift-marketplace/community-operators-zfgl8" Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.890784 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9bd6e85-2ed9-4525-9308-f6dd2f4eacda-utilities\") pod \"community-operators-zfgl8\" (UID: \"a9bd6e85-2ed9-4525-9308-f6dd2f4eacda\") " pod="openshift-marketplace/community-operators-zfgl8" Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.912004 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-86mf9"] Dec 11 02:03:22 crc kubenswrapper[4824]: I1211 02:03:22.915926 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lb5d7\" (UniqueName: \"kubernetes.io/projected/a9bd6e85-2ed9-4525-9308-f6dd2f4eacda-kube-api-access-lb5d7\") pod \"community-operators-zfgl8\" (UID: \"a9bd6e85-2ed9-4525-9308-f6dd2f4eacda\") " pod="openshift-marketplace/community-operators-zfgl8" Dec 11 02:03:23 crc kubenswrapper[4824]: I1211 02:03:23.047178 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zfgl8" Dec 11 02:03:23 crc kubenswrapper[4824]: I1211 02:03:23.054598 4824 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 11 02:03:23 crc kubenswrapper[4824]: I1211 02:03:23.054665 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-254w2\" (UID: \"18d74f73-0f04-4f3c-88c4-a431d490db9d\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-254w2" Dec 11 02:03:23 crc kubenswrapper[4824]: I1211 02:03:23.065942 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ptkck"] Dec 11 02:03:23 crc kubenswrapper[4824]: I1211 02:03:23.109573 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nqjh5" event={"ID":"d9806a3c-cb3d-4b3e-8889-11f8cf49c030","Type":"ContainerStarted","Data":"84b098c9bd03d479948fdfc8972a1077980e5c30874d36863df78bfcb45acfb1"} Dec 11 02:03:23 crc kubenswrapper[4824]: I1211 02:03:23.109616 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nqjh5" event={"ID":"d9806a3c-cb3d-4b3e-8889-11f8cf49c030","Type":"ContainerStarted","Data":"e5c7e4ed64f3273f14c292ac3b3d8ff1470dbb633aa74733de56db3daab5bde2"} Dec 11 02:03:23 crc kubenswrapper[4824]: I1211 02:03:23.113415 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-86mf9" event={"ID":"b2b7c06f-4350-41be-b472-c75d21558911","Type":"ContainerStarted","Data":"b5fe6fcc461fc458da49c42d8a8cc74b4e28b5c1a118a0a667331b5d36b1cc04"} Dec 11 02:03:23 crc kubenswrapper[4824]: I1211 02:03:23.119296 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-l88b4" event={"ID":"75da7b9a-7085-492b-954a-6bb920fbb186","Type":"ContainerStarted","Data":"6f9e011b08cdd19c98eda0931cfd48c764912e4e16e5be9a966e7447c6e5ebb6"} Dec 11 02:03:23 crc kubenswrapper[4824]: I1211 02:03:23.132156 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-254w2\" (UID: \"18d74f73-0f04-4f3c-88c4-a431d490db9d\") " pod="openshift-image-registry/image-registry-697d97f7c8-254w2" Dec 11 02:03:23 crc kubenswrapper[4824]: I1211 02:03:23.133616 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-hr4qs" event={"ID":"335b007e-ea5c-4e20-9402-5599da94ef22","Type":"ContainerStarted","Data":"afc644568c9ad72db66080afed29e01200033cf48b058247c64ab9155acfbed3"} Dec 11 02:03:23 crc kubenswrapper[4824]: I1211 02:03:23.133662 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-hr4qs" event={"ID":"335b007e-ea5c-4e20-9402-5599da94ef22","Type":"ContainerStarted","Data":"eea395245e15242b80d73e8bf9bff4800bebc5aa36cf873847b8efb127ba1b1d"} Dec 11 02:03:23 crc kubenswrapper[4824]: I1211 02:03:23.154065 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-z42g5" Dec 11 02:03:23 crc kubenswrapper[4824]: I1211 02:03:23.156305 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-l88b4" podStartSLOduration=126.156281222 podStartE2EDuration="2m6.156281222s" podCreationTimestamp="2025-12-11 02:01:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:03:23.143516828 +0000 UTC m=+144.832554227" watchObservedRunningTime="2025-12-11 02:03:23.156281222 +0000 UTC m=+144.845318601" Dec 11 02:03:23 crc kubenswrapper[4824]: I1211 02:03:23.156698 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-w6b4l" Dec 11 02:03:23 crc kubenswrapper[4824]: I1211 02:03:23.158539 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-254w2" Dec 11 02:03:23 crc kubenswrapper[4824]: I1211 02:03:23.198172 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 02:03:23 crc kubenswrapper[4824]: I1211 02:03:23.246546 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 11 02:03:23 crc kubenswrapper[4824]: I1211 02:03:23.338922 4824 patch_prober.go:28] interesting pod/router-default-5444994796-pg8rn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 11 02:03:23 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Dec 11 02:03:23 crc kubenswrapper[4824]: [+]process-running ok Dec 11 02:03:23 crc kubenswrapper[4824]: healthz check failed Dec 11 02:03:23 crc kubenswrapper[4824]: I1211 02:03:23.339307 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pg8rn" podUID="877d4a7f-46b0-4f9c-a702-b6b83620c3cb" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 11 02:03:23 crc kubenswrapper[4824]: I1211 02:03:23.459686 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zfgl8"] Dec 11 02:03:23 crc kubenswrapper[4824]: I1211 02:03:23.464847 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423640-42vd6" Dec 11 02:03:23 crc kubenswrapper[4824]: I1211 02:03:23.512625 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4c22bf31-ad33-44cd-a495-05d83688f2bd-config-volume\") pod \"4c22bf31-ad33-44cd-a495-05d83688f2bd\" (UID: \"4c22bf31-ad33-44cd-a495-05d83688f2bd\") " Dec 11 02:03:23 crc kubenswrapper[4824]: I1211 02:03:23.512673 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4c22bf31-ad33-44cd-a495-05d83688f2bd-secret-volume\") pod \"4c22bf31-ad33-44cd-a495-05d83688f2bd\" (UID: \"4c22bf31-ad33-44cd-a495-05d83688f2bd\") " Dec 11 02:03:23 crc kubenswrapper[4824]: I1211 02:03:23.512746 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-drmvw\" (UniqueName: \"kubernetes.io/projected/4c22bf31-ad33-44cd-a495-05d83688f2bd-kube-api-access-drmvw\") pod \"4c22bf31-ad33-44cd-a495-05d83688f2bd\" (UID: \"4c22bf31-ad33-44cd-a495-05d83688f2bd\") " Dec 11 02:03:23 crc kubenswrapper[4824]: I1211 02:03:23.518531 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c22bf31-ad33-44cd-a495-05d83688f2bd-config-volume" (OuterVolumeSpecName: "config-volume") pod "4c22bf31-ad33-44cd-a495-05d83688f2bd" (UID: "4c22bf31-ad33-44cd-a495-05d83688f2bd"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:03:23 crc kubenswrapper[4824]: I1211 02:03:23.526077 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c22bf31-ad33-44cd-a495-05d83688f2bd-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "4c22bf31-ad33-44cd-a495-05d83688f2bd" (UID: "4c22bf31-ad33-44cd-a495-05d83688f2bd"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:03:23 crc kubenswrapper[4824]: I1211 02:03:23.527200 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c22bf31-ad33-44cd-a495-05d83688f2bd-kube-api-access-drmvw" (OuterVolumeSpecName: "kube-api-access-drmvw") pod "4c22bf31-ad33-44cd-a495-05d83688f2bd" (UID: "4c22bf31-ad33-44cd-a495-05d83688f2bd"). InnerVolumeSpecName "kube-api-access-drmvw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:03:23 crc kubenswrapper[4824]: I1211 02:03:23.535922 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-254w2"] Dec 11 02:03:23 crc kubenswrapper[4824]: I1211 02:03:23.614648 4824 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4c22bf31-ad33-44cd-a495-05d83688f2bd-config-volume\") on node \"crc\" DevicePath \"\"" Dec 11 02:03:23 crc kubenswrapper[4824]: I1211 02:03:23.615086 4824 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4c22bf31-ad33-44cd-a495-05d83688f2bd-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 11 02:03:23 crc kubenswrapper[4824]: I1211 02:03:23.615100 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-drmvw\" (UniqueName: \"kubernetes.io/projected/4c22bf31-ad33-44cd-a495-05d83688f2bd-kube-api-access-drmvw\") on node \"crc\" DevicePath \"\"" Dec 11 02:03:24 crc kubenswrapper[4824]: I1211 02:03:24.125211 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-jc49g"] Dec 11 02:03:24 crc kubenswrapper[4824]: E1211 02:03:24.125454 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c22bf31-ad33-44cd-a495-05d83688f2bd" containerName="collect-profiles" Dec 11 02:03:24 crc kubenswrapper[4824]: I1211 02:03:24.125471 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c22bf31-ad33-44cd-a495-05d83688f2bd" containerName="collect-profiles" Dec 11 02:03:24 crc kubenswrapper[4824]: I1211 02:03:24.125609 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c22bf31-ad33-44cd-a495-05d83688f2bd" containerName="collect-profiles" Dec 11 02:03:24 crc kubenswrapper[4824]: I1211 02:03:24.126500 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jc49g" Dec 11 02:03:24 crc kubenswrapper[4824]: I1211 02:03:24.132173 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 11 02:03:24 crc kubenswrapper[4824]: I1211 02:03:24.151782 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jc49g"] Dec 11 02:03:24 crc kubenswrapper[4824]: I1211 02:03:24.154810 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-hr4qs" event={"ID":"335b007e-ea5c-4e20-9402-5599da94ef22","Type":"ContainerStarted","Data":"8834771dddc3a14e10c5af6c1b1921948a347f012504950ed1d097c66dbb20de"} Dec 11 02:03:24 crc kubenswrapper[4824]: I1211 02:03:24.178880 4824 generic.go:334] "Generic (PLEG): container finished" podID="9f40e0dc-3513-431e-95b3-af69c5cca852" containerID="18ec7c455c85f9cc7aad983ed5ae6aff2d9a0109928101d53120cb25c21cce9b" exitCode=0 Dec 11 02:03:24 crc kubenswrapper[4824]: I1211 02:03:24.178995 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ptkck" event={"ID":"9f40e0dc-3513-431e-95b3-af69c5cca852","Type":"ContainerDied","Data":"18ec7c455c85f9cc7aad983ed5ae6aff2d9a0109928101d53120cb25c21cce9b"} Dec 11 02:03:24 crc kubenswrapper[4824]: I1211 02:03:24.179028 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ptkck" event={"ID":"9f40e0dc-3513-431e-95b3-af69c5cca852","Type":"ContainerStarted","Data":"01a2f59c747f884738fabe124af1ffc22217be02033720fc6a5f7ff385165b96"} Dec 11 02:03:24 crc kubenswrapper[4824]: I1211 02:03:24.188328 4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 11 02:03:24 crc kubenswrapper[4824]: I1211 02:03:24.190152 4824 generic.go:334] "Generic (PLEG): container finished" podID="a9bd6e85-2ed9-4525-9308-f6dd2f4eacda" containerID="e1aad0c20b9c4d8ce6d5efd2b9be643373b60da8cce4695bf039ea40a61a66b8" exitCode=0 Dec 11 02:03:24 crc kubenswrapper[4824]: I1211 02:03:24.190200 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zfgl8" event={"ID":"a9bd6e85-2ed9-4525-9308-f6dd2f4eacda","Type":"ContainerDied","Data":"e1aad0c20b9c4d8ce6d5efd2b9be643373b60da8cce4695bf039ea40a61a66b8"} Dec 11 02:03:24 crc kubenswrapper[4824]: I1211 02:03:24.190248 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zfgl8" event={"ID":"a9bd6e85-2ed9-4525-9308-f6dd2f4eacda","Type":"ContainerStarted","Data":"67802c46fd3d9710d00d5a844f1f2f945cfa31af7c9b0a31c56e462c01d34d40"} Dec 11 02:03:24 crc kubenswrapper[4824]: I1211 02:03:24.193015 4824 generic.go:334] "Generic (PLEG): container finished" podID="d9806a3c-cb3d-4b3e-8889-11f8cf49c030" containerID="84b098c9bd03d479948fdfc8972a1077980e5c30874d36863df78bfcb45acfb1" exitCode=0 Dec 11 02:03:24 crc kubenswrapper[4824]: I1211 02:03:24.193066 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nqjh5" event={"ID":"d9806a3c-cb3d-4b3e-8889-11f8cf49c030","Type":"ContainerDied","Data":"84b098c9bd03d479948fdfc8972a1077980e5c30874d36863df78bfcb45acfb1"} Dec 11 02:03:24 crc kubenswrapper[4824]: I1211 02:03:24.196171 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423640-42vd6" event={"ID":"4c22bf31-ad33-44cd-a495-05d83688f2bd","Type":"ContainerDied","Data":"4fc4c2da81fdbc8e9a09e760e785feaaaa542605702fe0293f325b9f83602734"} Dec 11 02:03:24 crc kubenswrapper[4824]: I1211 02:03:24.196195 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4fc4c2da81fdbc8e9a09e760e785feaaaa542605702fe0293f325b9f83602734" Dec 11 02:03:24 crc kubenswrapper[4824]: I1211 02:03:24.196233 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423640-42vd6" Dec 11 02:03:24 crc kubenswrapper[4824]: I1211 02:03:24.203043 4824 generic.go:334] "Generic (PLEG): container finished" podID="b2b7c06f-4350-41be-b472-c75d21558911" containerID="143b00c17a13c20a70e1b91e408ede84b97addff9037deb82bc4f487dbcc0a69" exitCode=0 Dec 11 02:03:24 crc kubenswrapper[4824]: I1211 02:03:24.203103 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-86mf9" event={"ID":"b2b7c06f-4350-41be-b472-c75d21558911","Type":"ContainerDied","Data":"143b00c17a13c20a70e1b91e408ede84b97addff9037deb82bc4f487dbcc0a69"} Dec 11 02:03:24 crc kubenswrapper[4824]: I1211 02:03:24.209223 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-254w2" event={"ID":"18d74f73-0f04-4f3c-88c4-a431d490db9d","Type":"ContainerStarted","Data":"3f56af012e4c19f54832d8b45fdab91a708a1cf9f8d8ad236c76736800cf9a4e"} Dec 11 02:03:24 crc kubenswrapper[4824]: I1211 02:03:24.209263 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-254w2" event={"ID":"18d74f73-0f04-4f3c-88c4-a431d490db9d","Type":"ContainerStarted","Data":"f6e961ed0d22d4ebf17f47d28193643c29f600b6086ac18952ba7840db02ccc3"} Dec 11 02:03:24 crc kubenswrapper[4824]: I1211 02:03:24.226982 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6127fb96-7f67-453d-8d16-40fac34e6482-utilities\") pod \"redhat-marketplace-jc49g\" (UID: \"6127fb96-7f67-453d-8d16-40fac34e6482\") " pod="openshift-marketplace/redhat-marketplace-jc49g" Dec 11 02:03:24 crc kubenswrapper[4824]: I1211 02:03:24.227125 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6127fb96-7f67-453d-8d16-40fac34e6482-catalog-content\") pod \"redhat-marketplace-jc49g\" (UID: \"6127fb96-7f67-453d-8d16-40fac34e6482\") " pod="openshift-marketplace/redhat-marketplace-jc49g" Dec 11 02:03:24 crc kubenswrapper[4824]: I1211 02:03:24.227230 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-64k6f\" (UniqueName: \"kubernetes.io/projected/6127fb96-7f67-453d-8d16-40fac34e6482-kube-api-access-64k6f\") pod \"redhat-marketplace-jc49g\" (UID: \"6127fb96-7f67-453d-8d16-40fac34e6482\") " pod="openshift-marketplace/redhat-marketplace-jc49g" Dec 11 02:03:24 crc kubenswrapper[4824]: I1211 02:03:24.260021 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-hr4qs" podStartSLOduration=11.260005344 podStartE2EDuration="11.260005344s" podCreationTimestamp="2025-12-11 02:03:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:03:24.234011384 +0000 UTC m=+145.923048763" watchObservedRunningTime="2025-12-11 02:03:24.260005344 +0000 UTC m=+145.949042723" Dec 11 02:03:24 crc kubenswrapper[4824]: I1211 02:03:24.328534 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6127fb96-7f67-453d-8d16-40fac34e6482-utilities\") pod \"redhat-marketplace-jc49g\" (UID: \"6127fb96-7f67-453d-8d16-40fac34e6482\") " pod="openshift-marketplace/redhat-marketplace-jc49g" Dec 11 02:03:24 crc kubenswrapper[4824]: I1211 02:03:24.328679 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6127fb96-7f67-453d-8d16-40fac34e6482-catalog-content\") pod \"redhat-marketplace-jc49g\" (UID: \"6127fb96-7f67-453d-8d16-40fac34e6482\") " pod="openshift-marketplace/redhat-marketplace-jc49g" Dec 11 02:03:24 crc kubenswrapper[4824]: I1211 02:03:24.328978 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-64k6f\" (UniqueName: \"kubernetes.io/projected/6127fb96-7f67-453d-8d16-40fac34e6482-kube-api-access-64k6f\") pod \"redhat-marketplace-jc49g\" (UID: \"6127fb96-7f67-453d-8d16-40fac34e6482\") " pod="openshift-marketplace/redhat-marketplace-jc49g" Dec 11 02:03:24 crc kubenswrapper[4824]: I1211 02:03:24.329095 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6127fb96-7f67-453d-8d16-40fac34e6482-utilities\") pod \"redhat-marketplace-jc49g\" (UID: \"6127fb96-7f67-453d-8d16-40fac34e6482\") " pod="openshift-marketplace/redhat-marketplace-jc49g" Dec 11 02:03:24 crc kubenswrapper[4824]: I1211 02:03:24.329761 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6127fb96-7f67-453d-8d16-40fac34e6482-catalog-content\") pod \"redhat-marketplace-jc49g\" (UID: \"6127fb96-7f67-453d-8d16-40fac34e6482\") " pod="openshift-marketplace/redhat-marketplace-jc49g" Dec 11 02:03:24 crc kubenswrapper[4824]: I1211 02:03:24.337142 4824 patch_prober.go:28] interesting pod/router-default-5444994796-pg8rn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 11 02:03:24 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Dec 11 02:03:24 crc kubenswrapper[4824]: [+]process-running ok Dec 11 02:03:24 crc kubenswrapper[4824]: healthz check failed Dec 11 02:03:24 crc kubenswrapper[4824]: I1211 02:03:24.337411 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pg8rn" podUID="877d4a7f-46b0-4f9c-a702-b6b83620c3cb" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 11 02:03:24 crc kubenswrapper[4824]: I1211 02:03:24.345919 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-254w2" podStartSLOduration=127.345899843 podStartE2EDuration="2m7.345899843s" podCreationTimestamp="2025-12-11 02:01:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:03:24.341856712 +0000 UTC m=+146.030894081" watchObservedRunningTime="2025-12-11 02:03:24.345899843 +0000 UTC m=+146.034937222" Dec 11 02:03:24 crc kubenswrapper[4824]: I1211 02:03:24.349739 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-64k6f\" (UniqueName: \"kubernetes.io/projected/6127fb96-7f67-453d-8d16-40fac34e6482-kube-api-access-64k6f\") pod \"redhat-marketplace-jc49g\" (UID: \"6127fb96-7f67-453d-8d16-40fac34e6482\") " pod="openshift-marketplace/redhat-marketplace-jc49g" Dec 11 02:03:24 crc kubenswrapper[4824]: I1211 02:03:24.442573 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jc49g" Dec 11 02:03:24 crc kubenswrapper[4824]: I1211 02:03:24.522056 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-lbp88"] Dec 11 02:03:24 crc kubenswrapper[4824]: I1211 02:03:24.523027 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lbp88" Dec 11 02:03:24 crc kubenswrapper[4824]: I1211 02:03:24.542424 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lbp88"] Dec 11 02:03:24 crc kubenswrapper[4824]: I1211 02:03:24.636081 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2t5k\" (UniqueName: \"kubernetes.io/projected/0a7ce62c-21db-428f-9b80-29ed781821e4-kube-api-access-j2t5k\") pod \"redhat-marketplace-lbp88\" (UID: \"0a7ce62c-21db-428f-9b80-29ed781821e4\") " pod="openshift-marketplace/redhat-marketplace-lbp88" Dec 11 02:03:24 crc kubenswrapper[4824]: I1211 02:03:24.636156 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a7ce62c-21db-428f-9b80-29ed781821e4-utilities\") pod \"redhat-marketplace-lbp88\" (UID: \"0a7ce62c-21db-428f-9b80-29ed781821e4\") " pod="openshift-marketplace/redhat-marketplace-lbp88" Dec 11 02:03:24 crc kubenswrapper[4824]: I1211 02:03:24.636254 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a7ce62c-21db-428f-9b80-29ed781821e4-catalog-content\") pod \"redhat-marketplace-lbp88\" (UID: \"0a7ce62c-21db-428f-9b80-29ed781821e4\") " pod="openshift-marketplace/redhat-marketplace-lbp88" Dec 11 02:03:24 crc kubenswrapper[4824]: I1211 02:03:24.649125 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Dec 11 02:03:24 crc kubenswrapper[4824]: I1211 02:03:24.737770 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a7ce62c-21db-428f-9b80-29ed781821e4-catalog-content\") pod \"redhat-marketplace-lbp88\" (UID: \"0a7ce62c-21db-428f-9b80-29ed781821e4\") " pod="openshift-marketplace/redhat-marketplace-lbp88" Dec 11 02:03:24 crc kubenswrapper[4824]: I1211 02:03:24.737909 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2t5k\" (UniqueName: \"kubernetes.io/projected/0a7ce62c-21db-428f-9b80-29ed781821e4-kube-api-access-j2t5k\") pod \"redhat-marketplace-lbp88\" (UID: \"0a7ce62c-21db-428f-9b80-29ed781821e4\") " pod="openshift-marketplace/redhat-marketplace-lbp88" Dec 11 02:03:24 crc kubenswrapper[4824]: I1211 02:03:24.737945 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a7ce62c-21db-428f-9b80-29ed781821e4-utilities\") pod \"redhat-marketplace-lbp88\" (UID: \"0a7ce62c-21db-428f-9b80-29ed781821e4\") " pod="openshift-marketplace/redhat-marketplace-lbp88" Dec 11 02:03:24 crc kubenswrapper[4824]: I1211 02:03:24.738905 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a7ce62c-21db-428f-9b80-29ed781821e4-utilities\") pod \"redhat-marketplace-lbp88\" (UID: \"0a7ce62c-21db-428f-9b80-29ed781821e4\") " pod="openshift-marketplace/redhat-marketplace-lbp88" Dec 11 02:03:24 crc kubenswrapper[4824]: I1211 02:03:24.739201 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a7ce62c-21db-428f-9b80-29ed781821e4-catalog-content\") pod \"redhat-marketplace-lbp88\" (UID: \"0a7ce62c-21db-428f-9b80-29ed781821e4\") " pod="openshift-marketplace/redhat-marketplace-lbp88" Dec 11 02:03:24 crc kubenswrapper[4824]: I1211 02:03:24.743069 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jc49g"] Dec 11 02:03:24 crc kubenswrapper[4824]: W1211 02:03:24.748655 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6127fb96_7f67_453d_8d16_40fac34e6482.slice/crio-f0492d0355db6825fede043db8c18469767b36b313a1a0718163d8e0c2a9baa2 WatchSource:0}: Error finding container f0492d0355db6825fede043db8c18469767b36b313a1a0718163d8e0c2a9baa2: Status 404 returned error can't find the container with id f0492d0355db6825fede043db8c18469767b36b313a1a0718163d8e0c2a9baa2 Dec 11 02:03:24 crc kubenswrapper[4824]: I1211 02:03:24.761734 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2t5k\" (UniqueName: \"kubernetes.io/projected/0a7ce62c-21db-428f-9b80-29ed781821e4-kube-api-access-j2t5k\") pod \"redhat-marketplace-lbp88\" (UID: \"0a7ce62c-21db-428f-9b80-29ed781821e4\") " pod="openshift-marketplace/redhat-marketplace-lbp88" Dec 11 02:03:24 crc kubenswrapper[4824]: I1211 02:03:24.836824 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lbp88" Dec 11 02:03:25 crc kubenswrapper[4824]: I1211 02:03:25.070390 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lbp88"] Dec 11 02:03:25 crc kubenswrapper[4824]: W1211 02:03:25.081984 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0a7ce62c_21db_428f_9b80_29ed781821e4.slice/crio-49ccf53fae791aec1bb7e69c206eb84414e85056ede6ef8701888f633879bfa4 WatchSource:0}: Error finding container 49ccf53fae791aec1bb7e69c206eb84414e85056ede6ef8701888f633879bfa4: Status 404 returned error can't find the container with id 49ccf53fae791aec1bb7e69c206eb84414e85056ede6ef8701888f633879bfa4 Dec 11 02:03:25 crc kubenswrapper[4824]: I1211 02:03:25.149505 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 11 02:03:25 crc kubenswrapper[4824]: I1211 02:03:25.150464 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 11 02:03:25 crc kubenswrapper[4824]: I1211 02:03:25.154188 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Dec 11 02:03:25 crc kubenswrapper[4824]: I1211 02:03:25.168315 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 11 02:03:25 crc kubenswrapper[4824]: I1211 02:03:25.170673 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Dec 11 02:03:25 crc kubenswrapper[4824]: I1211 02:03:25.218339 4824 generic.go:334] "Generic (PLEG): container finished" podID="6127fb96-7f67-453d-8d16-40fac34e6482" containerID="7513e3598adb92ec0a08cf216ce6210836909455aaddb7f47d6ab801d14e1588" exitCode=0 Dec 11 02:03:25 crc kubenswrapper[4824]: I1211 02:03:25.218425 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jc49g" event={"ID":"6127fb96-7f67-453d-8d16-40fac34e6482","Type":"ContainerDied","Data":"7513e3598adb92ec0a08cf216ce6210836909455aaddb7f47d6ab801d14e1588"} Dec 11 02:03:25 crc kubenswrapper[4824]: I1211 02:03:25.218474 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jc49g" event={"ID":"6127fb96-7f67-453d-8d16-40fac34e6482","Type":"ContainerStarted","Data":"f0492d0355db6825fede043db8c18469767b36b313a1a0718163d8e0c2a9baa2"} Dec 11 02:03:25 crc kubenswrapper[4824]: I1211 02:03:25.221028 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lbp88" event={"ID":"0a7ce62c-21db-428f-9b80-29ed781821e4","Type":"ContainerStarted","Data":"49ccf53fae791aec1bb7e69c206eb84414e85056ede6ef8701888f633879bfa4"} Dec 11 02:03:25 crc kubenswrapper[4824]: I1211 02:03:25.229996 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-254w2" Dec 11 02:03:25 crc kubenswrapper[4824]: I1211 02:03:25.248395 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/660b5718-4b95-498e-88ae-1b96258cc871-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"660b5718-4b95-498e-88ae-1b96258cc871\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 11 02:03:25 crc kubenswrapper[4824]: I1211 02:03:25.248534 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/660b5718-4b95-498e-88ae-1b96258cc871-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"660b5718-4b95-498e-88ae-1b96258cc871\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 11 02:03:25 crc kubenswrapper[4824]: I1211 02:03:25.336977 4824 patch_prober.go:28] interesting pod/router-default-5444994796-pg8rn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 11 02:03:25 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Dec 11 02:03:25 crc kubenswrapper[4824]: [+]process-running ok Dec 11 02:03:25 crc kubenswrapper[4824]: healthz check failed Dec 11 02:03:25 crc kubenswrapper[4824]: I1211 02:03:25.337070 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pg8rn" podUID="877d4a7f-46b0-4f9c-a702-b6b83620c3cb" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 11 02:03:25 crc kubenswrapper[4824]: I1211 02:03:25.349797 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/660b5718-4b95-498e-88ae-1b96258cc871-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"660b5718-4b95-498e-88ae-1b96258cc871\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 11 02:03:25 crc kubenswrapper[4824]: I1211 02:03:25.349918 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/660b5718-4b95-498e-88ae-1b96258cc871-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"660b5718-4b95-498e-88ae-1b96258cc871\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 11 02:03:25 crc kubenswrapper[4824]: I1211 02:03:25.350719 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/660b5718-4b95-498e-88ae-1b96258cc871-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"660b5718-4b95-498e-88ae-1b96258cc871\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 11 02:03:25 crc kubenswrapper[4824]: I1211 02:03:25.369418 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/660b5718-4b95-498e-88ae-1b96258cc871-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"660b5718-4b95-498e-88ae-1b96258cc871\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 11 02:03:25 crc kubenswrapper[4824]: I1211 02:03:25.494693 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 11 02:03:25 crc kubenswrapper[4824]: I1211 02:03:25.517317 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-rnmtp"] Dec 11 02:03:25 crc kubenswrapper[4824]: I1211 02:03:25.518383 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rnmtp" Dec 11 02:03:25 crc kubenswrapper[4824]: I1211 02:03:25.520398 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 11 02:03:25 crc kubenswrapper[4824]: I1211 02:03:25.545415 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rnmtp"] Dec 11 02:03:25 crc kubenswrapper[4824]: I1211 02:03:25.553399 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b3b02dc-e8f5-4629-b1bf-72febdb40843-utilities\") pod \"redhat-operators-rnmtp\" (UID: \"0b3b02dc-e8f5-4629-b1bf-72febdb40843\") " pod="openshift-marketplace/redhat-operators-rnmtp" Dec 11 02:03:25 crc kubenswrapper[4824]: I1211 02:03:25.553582 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b3b02dc-e8f5-4629-b1bf-72febdb40843-catalog-content\") pod \"redhat-operators-rnmtp\" (UID: \"0b3b02dc-e8f5-4629-b1bf-72febdb40843\") " pod="openshift-marketplace/redhat-operators-rnmtp" Dec 11 02:03:25 crc kubenswrapper[4824]: I1211 02:03:25.553669 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zgcxd\" (UniqueName: \"kubernetes.io/projected/0b3b02dc-e8f5-4629-b1bf-72febdb40843-kube-api-access-zgcxd\") pod \"redhat-operators-rnmtp\" (UID: \"0b3b02dc-e8f5-4629-b1bf-72febdb40843\") " pod="openshift-marketplace/redhat-operators-rnmtp" Dec 11 02:03:25 crc kubenswrapper[4824]: I1211 02:03:25.621564 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-2ng4n" Dec 11 02:03:25 crc kubenswrapper[4824]: I1211 02:03:25.655045 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b3b02dc-e8f5-4629-b1bf-72febdb40843-catalog-content\") pod \"redhat-operators-rnmtp\" (UID: \"0b3b02dc-e8f5-4629-b1bf-72febdb40843\") " pod="openshift-marketplace/redhat-operators-rnmtp" Dec 11 02:03:25 crc kubenswrapper[4824]: I1211 02:03:25.655153 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zgcxd\" (UniqueName: \"kubernetes.io/projected/0b3b02dc-e8f5-4629-b1bf-72febdb40843-kube-api-access-zgcxd\") pod \"redhat-operators-rnmtp\" (UID: \"0b3b02dc-e8f5-4629-b1bf-72febdb40843\") " pod="openshift-marketplace/redhat-operators-rnmtp" Dec 11 02:03:25 crc kubenswrapper[4824]: I1211 02:03:25.655235 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b3b02dc-e8f5-4629-b1bf-72febdb40843-utilities\") pod \"redhat-operators-rnmtp\" (UID: \"0b3b02dc-e8f5-4629-b1bf-72febdb40843\") " pod="openshift-marketplace/redhat-operators-rnmtp" Dec 11 02:03:25 crc kubenswrapper[4824]: I1211 02:03:25.655768 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b3b02dc-e8f5-4629-b1bf-72febdb40843-utilities\") pod \"redhat-operators-rnmtp\" (UID: \"0b3b02dc-e8f5-4629-b1bf-72febdb40843\") " pod="openshift-marketplace/redhat-operators-rnmtp" Dec 11 02:03:25 crc kubenswrapper[4824]: I1211 02:03:25.656831 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b3b02dc-e8f5-4629-b1bf-72febdb40843-catalog-content\") pod \"redhat-operators-rnmtp\" (UID: \"0b3b02dc-e8f5-4629-b1bf-72febdb40843\") " pod="openshift-marketplace/redhat-operators-rnmtp" Dec 11 02:03:25 crc kubenswrapper[4824]: I1211 02:03:25.700337 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zgcxd\" (UniqueName: \"kubernetes.io/projected/0b3b02dc-e8f5-4629-b1bf-72febdb40843-kube-api-access-zgcxd\") pod \"redhat-operators-rnmtp\" (UID: \"0b3b02dc-e8f5-4629-b1bf-72febdb40843\") " pod="openshift-marketplace/redhat-operators-rnmtp" Dec 11 02:03:25 crc kubenswrapper[4824]: I1211 02:03:25.756931 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 02:03:25 crc kubenswrapper[4824]: I1211 02:03:25.756978 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 02:03:25 crc kubenswrapper[4824]: I1211 02:03:25.757049 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 02:03:25 crc kubenswrapper[4824]: I1211 02:03:25.757069 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 02:03:25 crc kubenswrapper[4824]: I1211 02:03:25.764501 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 02:03:25 crc kubenswrapper[4824]: I1211 02:03:25.765630 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 02:03:25 crc kubenswrapper[4824]: I1211 02:03:25.769379 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 02:03:25 crc kubenswrapper[4824]: I1211 02:03:25.787711 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 02:03:25 crc kubenswrapper[4824]: I1211 02:03:25.827381 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 11 02:03:25 crc kubenswrapper[4824]: I1211 02:03:25.854471 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 02:03:25 crc kubenswrapper[4824]: I1211 02:03:25.855150 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rnmtp" Dec 11 02:03:25 crc kubenswrapper[4824]: W1211 02:03:25.859661 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod660b5718_4b95_498e_88ae_1b96258cc871.slice/crio-f20a572d9a624de9899414c490b667c570bebc00b362ccf089e03c2f0850e2de WatchSource:0}: Error finding container f20a572d9a624de9899414c490b667c570bebc00b362ccf089e03c2f0850e2de: Status 404 returned error can't find the container with id f20a572d9a624de9899414c490b667c570bebc00b362ccf089e03c2f0850e2de Dec 11 02:03:25 crc kubenswrapper[4824]: I1211 02:03:25.863659 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-k6kls" Dec 11 02:03:25 crc kubenswrapper[4824]: I1211 02:03:25.863704 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-k6kls" Dec 11 02:03:25 crc kubenswrapper[4824]: I1211 02:03:25.865773 4824 patch_prober.go:28] interesting pod/console-f9d7485db-k6kls container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.21:8443/health\": dial tcp 10.217.0.21:8443: connect: connection refused" start-of-body= Dec 11 02:03:25 crc kubenswrapper[4824]: I1211 02:03:25.865818 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-k6kls" podUID="ad510a18-cd72-4cf0-9b0c-dd3540d491fa" containerName="console" probeResult="failure" output="Get \"https://10.217.0.21:8443/health\": dial tcp 10.217.0.21:8443: connect: connection refused" Dec 11 02:03:25 crc kubenswrapper[4824]: I1211 02:03:25.869969 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 02:03:25 crc kubenswrapper[4824]: I1211 02:03:25.883137 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 02:03:25 crc kubenswrapper[4824]: I1211 02:03:25.918536 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-ktsvl"] Dec 11 02:03:25 crc kubenswrapper[4824]: I1211 02:03:25.920485 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ktsvl" Dec 11 02:03:25 crc kubenswrapper[4824]: I1211 02:03:25.922270 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-l88b4" Dec 11 02:03:25 crc kubenswrapper[4824]: I1211 02:03:25.924447 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-l88b4" Dec 11 02:03:25 crc kubenswrapper[4824]: I1211 02:03:25.929364 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ktsvl"] Dec 11 02:03:25 crc kubenswrapper[4824]: I1211 02:03:25.931934 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-l88b4" Dec 11 02:03:25 crc kubenswrapper[4824]: I1211 02:03:25.959615 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/36403c25-41c3-4014-b64d-5eea981691a4-catalog-content\") pod \"redhat-operators-ktsvl\" (UID: \"36403c25-41c3-4014-b64d-5eea981691a4\") " pod="openshift-marketplace/redhat-operators-ktsvl" Dec 11 02:03:25 crc kubenswrapper[4824]: I1211 02:03:25.960075 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/36403c25-41c3-4014-b64d-5eea981691a4-utilities\") pod \"redhat-operators-ktsvl\" (UID: \"36403c25-41c3-4014-b64d-5eea981691a4\") " pod="openshift-marketplace/redhat-operators-ktsvl" Dec 11 02:03:25 crc kubenswrapper[4824]: I1211 02:03:25.960101 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-968xz\" (UniqueName: \"kubernetes.io/projected/36403c25-41c3-4014-b64d-5eea981691a4-kube-api-access-968xz\") pod \"redhat-operators-ktsvl\" (UID: \"36403c25-41c3-4014-b64d-5eea981691a4\") " pod="openshift-marketplace/redhat-operators-ktsvl" Dec 11 02:03:26 crc kubenswrapper[4824]: I1211 02:03:26.060983 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/36403c25-41c3-4014-b64d-5eea981691a4-catalog-content\") pod \"redhat-operators-ktsvl\" (UID: \"36403c25-41c3-4014-b64d-5eea981691a4\") " pod="openshift-marketplace/redhat-operators-ktsvl" Dec 11 02:03:26 crc kubenswrapper[4824]: I1211 02:03:26.061048 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/36403c25-41c3-4014-b64d-5eea981691a4-utilities\") pod \"redhat-operators-ktsvl\" (UID: \"36403c25-41c3-4014-b64d-5eea981691a4\") " pod="openshift-marketplace/redhat-operators-ktsvl" Dec 11 02:03:26 crc kubenswrapper[4824]: I1211 02:03:26.061074 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-968xz\" (UniqueName: \"kubernetes.io/projected/36403c25-41c3-4014-b64d-5eea981691a4-kube-api-access-968xz\") pod \"redhat-operators-ktsvl\" (UID: \"36403c25-41c3-4014-b64d-5eea981691a4\") " pod="openshift-marketplace/redhat-operators-ktsvl" Dec 11 02:03:26 crc kubenswrapper[4824]: I1211 02:03:26.061917 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/36403c25-41c3-4014-b64d-5eea981691a4-utilities\") pod \"redhat-operators-ktsvl\" (UID: \"36403c25-41c3-4014-b64d-5eea981691a4\") " pod="openshift-marketplace/redhat-operators-ktsvl" Dec 11 02:03:26 crc kubenswrapper[4824]: I1211 02:03:26.068717 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/36403c25-41c3-4014-b64d-5eea981691a4-catalog-content\") pod \"redhat-operators-ktsvl\" (UID: \"36403c25-41c3-4014-b64d-5eea981691a4\") " pod="openshift-marketplace/redhat-operators-ktsvl" Dec 11 02:03:26 crc kubenswrapper[4824]: I1211 02:03:26.087767 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-968xz\" (UniqueName: \"kubernetes.io/projected/36403c25-41c3-4014-b64d-5eea981691a4-kube-api-access-968xz\") pod \"redhat-operators-ktsvl\" (UID: \"36403c25-41c3-4014-b64d-5eea981691a4\") " pod="openshift-marketplace/redhat-operators-ktsvl" Dec 11 02:03:26 crc kubenswrapper[4824]: I1211 02:03:26.255069 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"660b5718-4b95-498e-88ae-1b96258cc871","Type":"ContainerStarted","Data":"f20a572d9a624de9899414c490b667c570bebc00b362ccf089e03c2f0850e2de"} Dec 11 02:03:26 crc kubenswrapper[4824]: I1211 02:03:26.259290 4824 generic.go:334] "Generic (PLEG): container finished" podID="0a7ce62c-21db-428f-9b80-29ed781821e4" containerID="d74812b216e037d8f413c4e0aba5661d8b148c305340c78600f17fb29a79386b" exitCode=0 Dec 11 02:03:26 crc kubenswrapper[4824]: I1211 02:03:26.259886 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lbp88" event={"ID":"0a7ce62c-21db-428f-9b80-29ed781821e4","Type":"ContainerDied","Data":"d74812b216e037d8f413c4e0aba5661d8b148c305340c78600f17fb29a79386b"} Dec 11 02:03:26 crc kubenswrapper[4824]: I1211 02:03:26.265510 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-l88b4" Dec 11 02:03:26 crc kubenswrapper[4824]: I1211 02:03:26.271463 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ktsvl" Dec 11 02:03:26 crc kubenswrapper[4824]: I1211 02:03:26.293321 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rnmtp"] Dec 11 02:03:26 crc kubenswrapper[4824]: I1211 02:03:26.339261 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-pg8rn" Dec 11 02:03:26 crc kubenswrapper[4824]: I1211 02:03:26.367819 4824 patch_prober.go:28] interesting pod/router-default-5444994796-pg8rn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 11 02:03:26 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Dec 11 02:03:26 crc kubenswrapper[4824]: [+]process-running ok Dec 11 02:03:26 crc kubenswrapper[4824]: healthz check failed Dec 11 02:03:26 crc kubenswrapper[4824]: I1211 02:03:26.368915 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pg8rn" podUID="877d4a7f-46b0-4f9c-a702-b6b83620c3cb" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 11 02:03:26 crc kubenswrapper[4824]: I1211 02:03:26.378422 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 11 02:03:26 crc kubenswrapper[4824]: I1211 02:03:26.379583 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 11 02:03:26 crc kubenswrapper[4824]: I1211 02:03:26.395178 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 11 02:03:26 crc kubenswrapper[4824]: I1211 02:03:26.397795 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 11 02:03:26 crc kubenswrapper[4824]: I1211 02:03:26.399171 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 11 02:03:26 crc kubenswrapper[4824]: W1211 02:03:26.544921 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-00bcfe476fa8a026f16eb4333aa0380663867460151849374878f6f986233ff9 WatchSource:0}: Error finding container 00bcfe476fa8a026f16eb4333aa0380663867460151849374878f6f986233ff9: Status 404 returned error can't find the container with id 00bcfe476fa8a026f16eb4333aa0380663867460151849374878f6f986233ff9 Dec 11 02:03:26 crc kubenswrapper[4824]: I1211 02:03:26.569186 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/027539a3-0709-498e-9136-5a6e61dcdab4-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"027539a3-0709-498e-9136-5a6e61dcdab4\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 11 02:03:26 crc kubenswrapper[4824]: I1211 02:03:26.569987 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/027539a3-0709-498e-9136-5a6e61dcdab4-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"027539a3-0709-498e-9136-5a6e61dcdab4\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 11 02:03:26 crc kubenswrapper[4824]: I1211 02:03:26.605410 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ktsvl"] Dec 11 02:03:26 crc kubenswrapper[4824]: I1211 02:03:26.670957 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/027539a3-0709-498e-9136-5a6e61dcdab4-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"027539a3-0709-498e-9136-5a6e61dcdab4\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 11 02:03:26 crc kubenswrapper[4824]: I1211 02:03:26.670999 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/027539a3-0709-498e-9136-5a6e61dcdab4-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"027539a3-0709-498e-9136-5a6e61dcdab4\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 11 02:03:26 crc kubenswrapper[4824]: I1211 02:03:26.671338 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/027539a3-0709-498e-9136-5a6e61dcdab4-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"027539a3-0709-498e-9136-5a6e61dcdab4\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 11 02:03:26 crc kubenswrapper[4824]: W1211 02:03:26.683321 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-cced04740c75e00cf876c8d468b759c4540f3c874361958271196149cb3e7a37 WatchSource:0}: Error finding container cced04740c75e00cf876c8d468b759c4540f3c874361958271196149cb3e7a37: Status 404 returned error can't find the container with id cced04740c75e00cf876c8d468b759c4540f3c874361958271196149cb3e7a37 Dec 11 02:03:26 crc kubenswrapper[4824]: I1211 02:03:26.687354 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/027539a3-0709-498e-9136-5a6e61dcdab4-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"027539a3-0709-498e-9136-5a6e61dcdab4\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 11 02:03:26 crc kubenswrapper[4824]: I1211 02:03:26.746322 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 11 02:03:26 crc kubenswrapper[4824]: W1211 02:03:26.830377 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-07b9a6448abc0c2d08c18b480177cd861236163afda441d53e546c8af5952ee9 WatchSource:0}: Error finding container 07b9a6448abc0c2d08c18b480177cd861236163afda441d53e546c8af5952ee9: Status 404 returned error can't find the container with id 07b9a6448abc0c2d08c18b480177cd861236163afda441d53e546c8af5952ee9 Dec 11 02:03:27 crc kubenswrapper[4824]: I1211 02:03:27.075630 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 11 02:03:27 crc kubenswrapper[4824]: W1211 02:03:27.107758 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod027539a3_0709_498e_9136_5a6e61dcdab4.slice/crio-a6e3783bee238cb6375a0cb5826bb9a94fd9b9935ad8b3a9009670004689c2e7 WatchSource:0}: Error finding container a6e3783bee238cb6375a0cb5826bb9a94fd9b9935ad8b3a9009670004689c2e7: Status 404 returned error can't find the container with id a6e3783bee238cb6375a0cb5826bb9a94fd9b9935ad8b3a9009670004689c2e7 Dec 11 02:03:27 crc kubenswrapper[4824]: I1211 02:03:27.338655 4824 patch_prober.go:28] interesting pod/router-default-5444994796-pg8rn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 11 02:03:27 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Dec 11 02:03:27 crc kubenswrapper[4824]: [+]process-running ok Dec 11 02:03:27 crc kubenswrapper[4824]: healthz check failed Dec 11 02:03:27 crc kubenswrapper[4824]: I1211 02:03:27.338837 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pg8rn" podUID="877d4a7f-46b0-4f9c-a702-b6b83620c3cb" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 11 02:03:27 crc kubenswrapper[4824]: I1211 02:03:27.357967 4824 generic.go:334] "Generic (PLEG): container finished" podID="0b3b02dc-e8f5-4629-b1bf-72febdb40843" containerID="57a96e8e6e5d272df43e6d59a414c321970a47564eb5dde745abddba9723e046" exitCode=0 Dec 11 02:03:27 crc kubenswrapper[4824]: I1211 02:03:27.358080 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rnmtp" event={"ID":"0b3b02dc-e8f5-4629-b1bf-72febdb40843","Type":"ContainerDied","Data":"57a96e8e6e5d272df43e6d59a414c321970a47564eb5dde745abddba9723e046"} Dec 11 02:03:27 crc kubenswrapper[4824]: I1211 02:03:27.358118 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rnmtp" event={"ID":"0b3b02dc-e8f5-4629-b1bf-72febdb40843","Type":"ContainerStarted","Data":"aced56de9a90fc5c8e1f819e7e3e6eb8fa5199de5104caef53f1d721f12fa3f1"} Dec 11 02:03:27 crc kubenswrapper[4824]: I1211 02:03:27.368571 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"027539a3-0709-498e-9136-5a6e61dcdab4","Type":"ContainerStarted","Data":"a6e3783bee238cb6375a0cb5826bb9a94fd9b9935ad8b3a9009670004689c2e7"} Dec 11 02:03:27 crc kubenswrapper[4824]: I1211 02:03:27.389281 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"660b5718-4b95-498e-88ae-1b96258cc871","Type":"ContainerStarted","Data":"263361fe09d4cb2764a99c13cddd319b69bd5b54338b29f4b157e900c6694c7b"} Dec 11 02:03:27 crc kubenswrapper[4824]: I1211 02:03:27.407800 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ktsvl" event={"ID":"36403c25-41c3-4014-b64d-5eea981691a4","Type":"ContainerStarted","Data":"6137d9e1d216ff1ffb575a5aa83ea1f1c9c8cf6cdc6df5a7e3d222a01d5a97d7"} Dec 11 02:03:27 crc kubenswrapper[4824]: I1211 02:03:27.416229 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"3a57bf32b084ce014a3e975171e5d523b3eacbb3e1a989f4c5ac651f35f497a2"} Dec 11 02:03:27 crc kubenswrapper[4824]: I1211 02:03:27.416275 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"00bcfe476fa8a026f16eb4333aa0380663867460151849374878f6f986233ff9"} Dec 11 02:03:27 crc kubenswrapper[4824]: I1211 02:03:27.418393 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"07b9a6448abc0c2d08c18b480177cd861236163afda441d53e546c8af5952ee9"} Dec 11 02:03:27 crc kubenswrapper[4824]: I1211 02:03:27.435820 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"34ac47d2c01d4032a4ef9f096a33be534ed0352d3aa6451f184c6e899c89aaa1"} Dec 11 02:03:27 crc kubenswrapper[4824]: I1211 02:03:27.435878 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"cced04740c75e00cf876c8d468b759c4540f3c874361958271196149cb3e7a37"} Dec 11 02:03:27 crc kubenswrapper[4824]: I1211 02:03:27.436139 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 02:03:28 crc kubenswrapper[4824]: I1211 02:03:28.338930 4824 patch_prober.go:28] interesting pod/router-default-5444994796-pg8rn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 11 02:03:28 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Dec 11 02:03:28 crc kubenswrapper[4824]: [+]process-running ok Dec 11 02:03:28 crc kubenswrapper[4824]: healthz check failed Dec 11 02:03:28 crc kubenswrapper[4824]: I1211 02:03:28.339310 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pg8rn" podUID="877d4a7f-46b0-4f9c-a702-b6b83620c3cb" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 11 02:03:28 crc kubenswrapper[4824]: I1211 02:03:28.467328 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"027539a3-0709-498e-9136-5a6e61dcdab4","Type":"ContainerStarted","Data":"3409d84cb887cb19c460534279ab50d430ebf36f9baf929d2777f8442468b90e"} Dec 11 02:03:28 crc kubenswrapper[4824]: I1211 02:03:28.481310 4824 generic.go:334] "Generic (PLEG): container finished" podID="660b5718-4b95-498e-88ae-1b96258cc871" containerID="263361fe09d4cb2764a99c13cddd319b69bd5b54338b29f4b157e900c6694c7b" exitCode=0 Dec 11 02:03:28 crc kubenswrapper[4824]: I1211 02:03:28.481405 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"660b5718-4b95-498e-88ae-1b96258cc871","Type":"ContainerDied","Data":"263361fe09d4cb2764a99c13cddd319b69bd5b54338b29f4b157e900c6694c7b"} Dec 11 02:03:28 crc kubenswrapper[4824]: I1211 02:03:28.494685 4824 generic.go:334] "Generic (PLEG): container finished" podID="36403c25-41c3-4014-b64d-5eea981691a4" containerID="d9ab1d347c29c4ec85f3a7faef64e75bf9d9e4d8b1b8422f530ca91921e902c0" exitCode=0 Dec 11 02:03:28 crc kubenswrapper[4824]: I1211 02:03:28.494753 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ktsvl" event={"ID":"36403c25-41c3-4014-b64d-5eea981691a4","Type":"ContainerDied","Data":"d9ab1d347c29c4ec85f3a7faef64e75bf9d9e4d8b1b8422f530ca91921e902c0"} Dec 11 02:03:28 crc kubenswrapper[4824]: I1211 02:03:28.499683 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"e8eb7961b193b064287d8925321dcf1a403eb54c0290623b6330bf869874b56f"} Dec 11 02:03:28 crc kubenswrapper[4824]: I1211 02:03:28.845011 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 11 02:03:29 crc kubenswrapper[4824]: I1211 02:03:29.015139 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/660b5718-4b95-498e-88ae-1b96258cc871-kube-api-access\") pod \"660b5718-4b95-498e-88ae-1b96258cc871\" (UID: \"660b5718-4b95-498e-88ae-1b96258cc871\") " Dec 11 02:03:29 crc kubenswrapper[4824]: I1211 02:03:29.015195 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/660b5718-4b95-498e-88ae-1b96258cc871-kubelet-dir\") pod \"660b5718-4b95-498e-88ae-1b96258cc871\" (UID: \"660b5718-4b95-498e-88ae-1b96258cc871\") " Dec 11 02:03:29 crc kubenswrapper[4824]: I1211 02:03:29.015459 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/660b5718-4b95-498e-88ae-1b96258cc871-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "660b5718-4b95-498e-88ae-1b96258cc871" (UID: "660b5718-4b95-498e-88ae-1b96258cc871"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 02:03:29 crc kubenswrapper[4824]: I1211 02:03:29.034684 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/660b5718-4b95-498e-88ae-1b96258cc871-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "660b5718-4b95-498e-88ae-1b96258cc871" (UID: "660b5718-4b95-498e-88ae-1b96258cc871"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:03:29 crc kubenswrapper[4824]: I1211 02:03:29.117047 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/660b5718-4b95-498e-88ae-1b96258cc871-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 11 02:03:29 crc kubenswrapper[4824]: I1211 02:03:29.117078 4824 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/660b5718-4b95-498e-88ae-1b96258cc871-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 11 02:03:29 crc kubenswrapper[4824]: I1211 02:03:29.337301 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-pg8rn" Dec 11 02:03:29 crc kubenswrapper[4824]: I1211 02:03:29.344692 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-pg8rn" Dec 11 02:03:29 crc kubenswrapper[4824]: I1211 02:03:29.528958 4824 generic.go:334] "Generic (PLEG): container finished" podID="027539a3-0709-498e-9136-5a6e61dcdab4" containerID="3409d84cb887cb19c460534279ab50d430ebf36f9baf929d2777f8442468b90e" exitCode=0 Dec 11 02:03:29 crc kubenswrapper[4824]: I1211 02:03:29.529058 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"027539a3-0709-498e-9136-5a6e61dcdab4","Type":"ContainerDied","Data":"3409d84cb887cb19c460534279ab50d430ebf36f9baf929d2777f8442468b90e"} Dec 11 02:03:29 crc kubenswrapper[4824]: I1211 02:03:29.535075 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 11 02:03:29 crc kubenswrapper[4824]: I1211 02:03:29.535703 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"660b5718-4b95-498e-88ae-1b96258cc871","Type":"ContainerDied","Data":"f20a572d9a624de9899414c490b667c570bebc00b362ccf089e03c2f0850e2de"} Dec 11 02:03:29 crc kubenswrapper[4824]: I1211 02:03:29.535730 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f20a572d9a624de9899414c490b667c570bebc00b362ccf089e03c2f0850e2de" Dec 11 02:03:31 crc kubenswrapper[4824]: I1211 02:03:31.475763 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-qtll9" Dec 11 02:03:35 crc kubenswrapper[4824]: I1211 02:03:35.898281 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-k6kls" Dec 11 02:03:35 crc kubenswrapper[4824]: I1211 02:03:35.904072 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-k6kls" Dec 11 02:03:36 crc kubenswrapper[4824]: I1211 02:03:36.410075 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 11 02:03:36 crc kubenswrapper[4824]: I1211 02:03:36.588661 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/027539a3-0709-498e-9136-5a6e61dcdab4-kubelet-dir\") pod \"027539a3-0709-498e-9136-5a6e61dcdab4\" (UID: \"027539a3-0709-498e-9136-5a6e61dcdab4\") " Dec 11 02:03:36 crc kubenswrapper[4824]: I1211 02:03:36.588712 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/027539a3-0709-498e-9136-5a6e61dcdab4-kube-api-access\") pod \"027539a3-0709-498e-9136-5a6e61dcdab4\" (UID: \"027539a3-0709-498e-9136-5a6e61dcdab4\") " Dec 11 02:03:36 crc kubenswrapper[4824]: I1211 02:03:36.588760 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/027539a3-0709-498e-9136-5a6e61dcdab4-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "027539a3-0709-498e-9136-5a6e61dcdab4" (UID: "027539a3-0709-498e-9136-5a6e61dcdab4"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 02:03:36 crc kubenswrapper[4824]: I1211 02:03:36.589125 4824 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/027539a3-0709-498e-9136-5a6e61dcdab4-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 11 02:03:36 crc kubenswrapper[4824]: I1211 02:03:36.592886 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 11 02:03:36 crc kubenswrapper[4824]: I1211 02:03:36.592910 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"027539a3-0709-498e-9136-5a6e61dcdab4","Type":"ContainerDied","Data":"a6e3783bee238cb6375a0cb5826bb9a94fd9b9935ad8b3a9009670004689c2e7"} Dec 11 02:03:36 crc kubenswrapper[4824]: I1211 02:03:36.593360 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a6e3783bee238cb6375a0cb5826bb9a94fd9b9935ad8b3a9009670004689c2e7" Dec 11 02:03:36 crc kubenswrapper[4824]: I1211 02:03:36.595594 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/027539a3-0709-498e-9136-5a6e61dcdab4-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "027539a3-0709-498e-9136-5a6e61dcdab4" (UID: "027539a3-0709-498e-9136-5a6e61dcdab4"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:03:36 crc kubenswrapper[4824]: I1211 02:03:36.691002 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/027539a3-0709-498e-9136-5a6e61dcdab4-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 11 02:03:37 crc kubenswrapper[4824]: I1211 02:03:37.251873 4824 patch_prober.go:28] interesting pod/machine-config-daemon-gx6xt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 02:03:37 crc kubenswrapper[4824]: I1211 02:03:37.252358 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 02:03:39 crc kubenswrapper[4824]: I1211 02:03:39.832280 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/09a52948-0404-45ff-85b1-51479268bf71-metrics-certs\") pod \"network-metrics-daemon-grnpt\" (UID: \"09a52948-0404-45ff-85b1-51479268bf71\") " pod="openshift-multus/network-metrics-daemon-grnpt" Dec 11 02:03:39 crc kubenswrapper[4824]: I1211 02:03:39.849254 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/09a52948-0404-45ff-85b1-51479268bf71-metrics-certs\") pod \"network-metrics-daemon-grnpt\" (UID: \"09a52948-0404-45ff-85b1-51479268bf71\") " pod="openshift-multus/network-metrics-daemon-grnpt" Dec 11 02:03:40 crc kubenswrapper[4824]: I1211 02:03:40.060743 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-grnpt" Dec 11 02:03:43 crc kubenswrapper[4824]: I1211 02:03:43.173943 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-254w2" Dec 11 02:03:52 crc kubenswrapper[4824]: E1211 02:03:52.251090 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 11 02:03:52 crc kubenswrapper[4824]: E1211 02:03:52.251799 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-t6z6w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-ptkck_openshift-marketplace(9f40e0dc-3513-431e-95b3-af69c5cca852): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 11 02:03:52 crc kubenswrapper[4824]: E1211 02:03:52.253019 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-ptkck" podUID="9f40e0dc-3513-431e-95b3-af69c5cca852" Dec 11 02:03:54 crc kubenswrapper[4824]: E1211 02:03:54.650412 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-ptkck" podUID="9f40e0dc-3513-431e-95b3-af69c5cca852" Dec 11 02:03:54 crc kubenswrapper[4824]: E1211 02:03:54.712885 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 11 02:03:54 crc kubenswrapper[4824]: E1211 02:03:54.713043 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4xhtb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-nqjh5_openshift-marketplace(d9806a3c-cb3d-4b3e-8889-11f8cf49c030): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 11 02:03:54 crc kubenswrapper[4824]: E1211 02:03:54.714572 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-nqjh5" podUID="d9806a3c-cb3d-4b3e-8889-11f8cf49c030" Dec 11 02:03:56 crc kubenswrapper[4824]: E1211 02:03:56.142894 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-nqjh5" podUID="d9806a3c-cb3d-4b3e-8889-11f8cf49c030" Dec 11 02:03:56 crc kubenswrapper[4824]: E1211 02:03:56.217134 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 11 02:03:56 crc kubenswrapper[4824]: E1211 02:03:56.217318 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2gdpn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-86mf9_openshift-marketplace(b2b7c06f-4350-41be-b472-c75d21558911): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 11 02:03:56 crc kubenswrapper[4824]: E1211 02:03:56.218461 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-86mf9" podUID="b2b7c06f-4350-41be-b472-c75d21558911" Dec 11 02:03:56 crc kubenswrapper[4824]: E1211 02:03:56.242309 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 11 02:03:56 crc kubenswrapper[4824]: E1211 02:03:56.242508 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lb5d7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-zfgl8_openshift-marketplace(a9bd6e85-2ed9-4525-9308-f6dd2f4eacda): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 11 02:03:56 crc kubenswrapper[4824]: E1211 02:03:56.243615 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-zfgl8" podUID="a9bd6e85-2ed9-4525-9308-f6dd2f4eacda" Dec 11 02:03:56 crc kubenswrapper[4824]: I1211 02:03:56.405698 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-lpgf5" Dec 11 02:03:58 crc kubenswrapper[4824]: E1211 02:03:58.971577 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-zfgl8" podUID="a9bd6e85-2ed9-4525-9308-f6dd2f4eacda" Dec 11 02:03:58 crc kubenswrapper[4824]: E1211 02:03:58.971605 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-86mf9" podUID="b2b7c06f-4350-41be-b472-c75d21558911" Dec 11 02:04:01 crc kubenswrapper[4824]: I1211 02:04:01.798149 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 11 02:04:01 crc kubenswrapper[4824]: E1211 02:04:01.798555 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="027539a3-0709-498e-9136-5a6e61dcdab4" containerName="pruner" Dec 11 02:04:01 crc kubenswrapper[4824]: I1211 02:04:01.798574 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="027539a3-0709-498e-9136-5a6e61dcdab4" containerName="pruner" Dec 11 02:04:01 crc kubenswrapper[4824]: E1211 02:04:01.798603 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="660b5718-4b95-498e-88ae-1b96258cc871" containerName="pruner" Dec 11 02:04:01 crc kubenswrapper[4824]: I1211 02:04:01.798617 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="660b5718-4b95-498e-88ae-1b96258cc871" containerName="pruner" Dec 11 02:04:01 crc kubenswrapper[4824]: I1211 02:04:01.798753 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="027539a3-0709-498e-9136-5a6e61dcdab4" containerName="pruner" Dec 11 02:04:01 crc kubenswrapper[4824]: I1211 02:04:01.798775 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="660b5718-4b95-498e-88ae-1b96258cc871" containerName="pruner" Dec 11 02:04:01 crc kubenswrapper[4824]: I1211 02:04:01.799444 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 11 02:04:01 crc kubenswrapper[4824]: I1211 02:04:01.803502 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 11 02:04:01 crc kubenswrapper[4824]: I1211 02:04:01.804971 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 11 02:04:01 crc kubenswrapper[4824]: I1211 02:04:01.808942 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 11 02:04:01 crc kubenswrapper[4824]: I1211 02:04:01.923627 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e702020b-27da-448e-8e87-fdcb85fc09be-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"e702020b-27da-448e-8e87-fdcb85fc09be\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 11 02:04:01 crc kubenswrapper[4824]: I1211 02:04:01.923955 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e702020b-27da-448e-8e87-fdcb85fc09be-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"e702020b-27da-448e-8e87-fdcb85fc09be\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 11 02:04:01 crc kubenswrapper[4824]: E1211 02:04:01.958246 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 11 02:04:01 crc kubenswrapper[4824]: E1211 02:04:01.958401 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-64k6f,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-jc49g_openshift-marketplace(6127fb96-7f67-453d-8d16-40fac34e6482): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 11 02:04:01 crc kubenswrapper[4824]: E1211 02:04:01.959753 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-jc49g" podUID="6127fb96-7f67-453d-8d16-40fac34e6482" Dec 11 02:04:01 crc kubenswrapper[4824]: E1211 02:04:01.998252 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 11 02:04:01 crc kubenswrapper[4824]: E1211 02:04:01.998408 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-968xz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-ktsvl_openshift-marketplace(36403c25-41c3-4014-b64d-5eea981691a4): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 11 02:04:01 crc kubenswrapper[4824]: E1211 02:04:01.999765 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-ktsvl" podUID="36403c25-41c3-4014-b64d-5eea981691a4" Dec 11 02:04:02 crc kubenswrapper[4824]: E1211 02:04:02.016363 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 11 02:04:02 crc kubenswrapper[4824]: E1211 02:04:02.016526 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-j2t5k,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-lbp88_openshift-marketplace(0a7ce62c-21db-428f-9b80-29ed781821e4): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 11 02:04:02 crc kubenswrapper[4824]: E1211 02:04:02.017652 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-lbp88" podUID="0a7ce62c-21db-428f-9b80-29ed781821e4" Dec 11 02:04:02 crc kubenswrapper[4824]: I1211 02:04:02.025474 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e702020b-27da-448e-8e87-fdcb85fc09be-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"e702020b-27da-448e-8e87-fdcb85fc09be\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 11 02:04:02 crc kubenswrapper[4824]: I1211 02:04:02.025523 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e702020b-27da-448e-8e87-fdcb85fc09be-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"e702020b-27da-448e-8e87-fdcb85fc09be\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 11 02:04:02 crc kubenswrapper[4824]: I1211 02:04:02.025715 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e702020b-27da-448e-8e87-fdcb85fc09be-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"e702020b-27da-448e-8e87-fdcb85fc09be\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 11 02:04:02 crc kubenswrapper[4824]: I1211 02:04:02.048367 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e702020b-27da-448e-8e87-fdcb85fc09be-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"e702020b-27da-448e-8e87-fdcb85fc09be\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 11 02:04:02 crc kubenswrapper[4824]: I1211 02:04:02.118024 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-grnpt"] Dec 11 02:04:02 crc kubenswrapper[4824]: I1211 02:04:02.130087 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 11 02:04:02 crc kubenswrapper[4824]: I1211 02:04:02.348690 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 11 02:04:02 crc kubenswrapper[4824]: W1211 02:04:02.373358 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pode702020b_27da_448e_8e87_fdcb85fc09be.slice/crio-a433c456dbe94d56d12165c47dca3a15608335801b85aeaa805fc7e09769a900 WatchSource:0}: Error finding container a433c456dbe94d56d12165c47dca3a15608335801b85aeaa805fc7e09769a900: Status 404 returned error can't find the container with id a433c456dbe94d56d12165c47dca3a15608335801b85aeaa805fc7e09769a900 Dec 11 02:04:02 crc kubenswrapper[4824]: I1211 02:04:02.736184 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"e702020b-27da-448e-8e87-fdcb85fc09be","Type":"ContainerStarted","Data":"a433c456dbe94d56d12165c47dca3a15608335801b85aeaa805fc7e09769a900"} Dec 11 02:04:02 crc kubenswrapper[4824]: I1211 02:04:02.738321 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-grnpt" event={"ID":"09a52948-0404-45ff-85b1-51479268bf71","Type":"ContainerStarted","Data":"c936ba83d4ffc863a4b477e13783c23e5499e4fdf8fec5705686c6ca2f44f066"} Dec 11 02:04:02 crc kubenswrapper[4824]: I1211 02:04:02.738392 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-grnpt" event={"ID":"09a52948-0404-45ff-85b1-51479268bf71","Type":"ContainerStarted","Data":"6f461e0051afebf66fb2385678dfb838b098c1c8e876a28f01ad7943966c1ba0"} Dec 11 02:04:02 crc kubenswrapper[4824]: I1211 02:04:02.738404 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-grnpt" event={"ID":"09a52948-0404-45ff-85b1-51479268bf71","Type":"ContainerStarted","Data":"4e64b76c9861bb9f28bd74aeaa25da5cb43501adb098323db095c2d107c15fc7"} Dec 11 02:04:02 crc kubenswrapper[4824]: I1211 02:04:02.747511 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rnmtp" event={"ID":"0b3b02dc-e8f5-4629-b1bf-72febdb40843","Type":"ContainerStarted","Data":"2b69d5e0865f31b41962f4290865c9d2a83767b4752253b77987606cc86a9bb3"} Dec 11 02:04:02 crc kubenswrapper[4824]: E1211 02:04:02.752574 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-lbp88" podUID="0a7ce62c-21db-428f-9b80-29ed781821e4" Dec 11 02:04:02 crc kubenswrapper[4824]: E1211 02:04:02.757211 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-jc49g" podUID="6127fb96-7f67-453d-8d16-40fac34e6482" Dec 11 02:04:02 crc kubenswrapper[4824]: E1211 02:04:02.761870 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-ktsvl" podUID="36403c25-41c3-4014-b64d-5eea981691a4" Dec 11 02:04:02 crc kubenswrapper[4824]: I1211 02:04:02.766138 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-grnpt" podStartSLOduration=165.766100683 podStartE2EDuration="2m45.766100683s" podCreationTimestamp="2025-12-11 02:01:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:04:02.763984559 +0000 UTC m=+184.453021938" watchObservedRunningTime="2025-12-11 02:04:02.766100683 +0000 UTC m=+184.455138062" Dec 11 02:04:03 crc kubenswrapper[4824]: I1211 02:04:03.748598 4824 generic.go:334] "Generic (PLEG): container finished" podID="e702020b-27da-448e-8e87-fdcb85fc09be" containerID="da317fe02193e5343dc1f2a5a65f152f898e95822025432b46e6cd0353267f88" exitCode=0 Dec 11 02:04:03 crc kubenswrapper[4824]: I1211 02:04:03.748644 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"e702020b-27da-448e-8e87-fdcb85fc09be","Type":"ContainerDied","Data":"da317fe02193e5343dc1f2a5a65f152f898e95822025432b46e6cd0353267f88"} Dec 11 02:04:03 crc kubenswrapper[4824]: I1211 02:04:03.752093 4824 generic.go:334] "Generic (PLEG): container finished" podID="0b3b02dc-e8f5-4629-b1bf-72febdb40843" containerID="2b69d5e0865f31b41962f4290865c9d2a83767b4752253b77987606cc86a9bb3" exitCode=0 Dec 11 02:04:03 crc kubenswrapper[4824]: I1211 02:04:03.752135 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rnmtp" event={"ID":"0b3b02dc-e8f5-4629-b1bf-72febdb40843","Type":"ContainerDied","Data":"2b69d5e0865f31b41962f4290865c9d2a83767b4752253b77987606cc86a9bb3"} Dec 11 02:04:04 crc kubenswrapper[4824]: I1211 02:04:04.761443 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rnmtp" event={"ID":"0b3b02dc-e8f5-4629-b1bf-72febdb40843","Type":"ContainerStarted","Data":"844683dffb7882a5cc33ed4c458f332f8ffa61080cfdae10d2ce9819e505dc34"} Dec 11 02:04:04 crc kubenswrapper[4824]: I1211 02:04:04.789817 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-rnmtp" podStartSLOduration=2.983277925 podStartE2EDuration="39.78979832s" podCreationTimestamp="2025-12-11 02:03:25 +0000 UTC" firstStartedPulling="2025-12-11 02:03:27.367385101 +0000 UTC m=+149.056422480" lastFinishedPulling="2025-12-11 02:04:04.173905476 +0000 UTC m=+185.862942875" observedRunningTime="2025-12-11 02:04:04.787208122 +0000 UTC m=+186.476245511" watchObservedRunningTime="2025-12-11 02:04:04.78979832 +0000 UTC m=+186.478835699" Dec 11 02:04:04 crc kubenswrapper[4824]: I1211 02:04:04.981430 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 11 02:04:05 crc kubenswrapper[4824]: I1211 02:04:05.063686 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e702020b-27da-448e-8e87-fdcb85fc09be-kube-api-access\") pod \"e702020b-27da-448e-8e87-fdcb85fc09be\" (UID: \"e702020b-27da-448e-8e87-fdcb85fc09be\") " Dec 11 02:04:05 crc kubenswrapper[4824]: I1211 02:04:05.063827 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e702020b-27da-448e-8e87-fdcb85fc09be-kubelet-dir\") pod \"e702020b-27da-448e-8e87-fdcb85fc09be\" (UID: \"e702020b-27da-448e-8e87-fdcb85fc09be\") " Dec 11 02:04:05 crc kubenswrapper[4824]: I1211 02:04:05.063890 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e702020b-27da-448e-8e87-fdcb85fc09be-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "e702020b-27da-448e-8e87-fdcb85fc09be" (UID: "e702020b-27da-448e-8e87-fdcb85fc09be"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 02:04:05 crc kubenswrapper[4824]: I1211 02:04:05.064217 4824 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e702020b-27da-448e-8e87-fdcb85fc09be-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 11 02:04:05 crc kubenswrapper[4824]: I1211 02:04:05.073239 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e702020b-27da-448e-8e87-fdcb85fc09be-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e702020b-27da-448e-8e87-fdcb85fc09be" (UID: "e702020b-27da-448e-8e87-fdcb85fc09be"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:04:05 crc kubenswrapper[4824]: I1211 02:04:05.165474 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e702020b-27da-448e-8e87-fdcb85fc09be-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 11 02:04:05 crc kubenswrapper[4824]: I1211 02:04:05.767921 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"e702020b-27da-448e-8e87-fdcb85fc09be","Type":"ContainerDied","Data":"a433c456dbe94d56d12165c47dca3a15608335801b85aeaa805fc7e09769a900"} Dec 11 02:04:05 crc kubenswrapper[4824]: I1211 02:04:05.768214 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a433c456dbe94d56d12165c47dca3a15608335801b85aeaa805fc7e09769a900" Dec 11 02:04:05 crc kubenswrapper[4824]: I1211 02:04:05.767995 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 11 02:04:05 crc kubenswrapper[4824]: I1211 02:04:05.855544 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-rnmtp" Dec 11 02:04:05 crc kubenswrapper[4824]: I1211 02:04:05.855610 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-rnmtp" Dec 11 02:04:05 crc kubenswrapper[4824]: I1211 02:04:05.873885 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 02:04:06 crc kubenswrapper[4824]: I1211 02:04:06.948508 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-rnmtp" podUID="0b3b02dc-e8f5-4629-b1bf-72febdb40843" containerName="registry-server" probeResult="failure" output=< Dec 11 02:04:06 crc kubenswrapper[4824]: timeout: failed to connect service ":50051" within 1s Dec 11 02:04:06 crc kubenswrapper[4824]: > Dec 11 02:04:07 crc kubenswrapper[4824]: I1211 02:04:07.251981 4824 patch_prober.go:28] interesting pod/machine-config-daemon-gx6xt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 02:04:07 crc kubenswrapper[4824]: I1211 02:04:07.252046 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 02:04:08 crc kubenswrapper[4824]: I1211 02:04:08.365914 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 11 02:04:08 crc kubenswrapper[4824]: E1211 02:04:08.366382 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e702020b-27da-448e-8e87-fdcb85fc09be" containerName="pruner" Dec 11 02:04:08 crc kubenswrapper[4824]: I1211 02:04:08.366413 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="e702020b-27da-448e-8e87-fdcb85fc09be" containerName="pruner" Dec 11 02:04:08 crc kubenswrapper[4824]: I1211 02:04:08.366663 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="e702020b-27da-448e-8e87-fdcb85fc09be" containerName="pruner" Dec 11 02:04:08 crc kubenswrapper[4824]: I1211 02:04:08.367667 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 11 02:04:08 crc kubenswrapper[4824]: I1211 02:04:08.370237 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 11 02:04:08 crc kubenswrapper[4824]: I1211 02:04:08.370399 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 11 02:04:08 crc kubenswrapper[4824]: I1211 02:04:08.397434 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 11 02:04:08 crc kubenswrapper[4824]: I1211 02:04:08.513177 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/2b27ed82-f069-42db-a3ae-907b1283d49e-var-lock\") pod \"installer-9-crc\" (UID: \"2b27ed82-f069-42db-a3ae-907b1283d49e\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 11 02:04:08 crc kubenswrapper[4824]: I1211 02:04:08.513329 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2b27ed82-f069-42db-a3ae-907b1283d49e-kubelet-dir\") pod \"installer-9-crc\" (UID: \"2b27ed82-f069-42db-a3ae-907b1283d49e\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 11 02:04:08 crc kubenswrapper[4824]: I1211 02:04:08.513426 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2b27ed82-f069-42db-a3ae-907b1283d49e-kube-api-access\") pod \"installer-9-crc\" (UID: \"2b27ed82-f069-42db-a3ae-907b1283d49e\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 11 02:04:08 crc kubenswrapper[4824]: I1211 02:04:08.615095 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2b27ed82-f069-42db-a3ae-907b1283d49e-kubelet-dir\") pod \"installer-9-crc\" (UID: \"2b27ed82-f069-42db-a3ae-907b1283d49e\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 11 02:04:08 crc kubenswrapper[4824]: I1211 02:04:08.615255 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2b27ed82-f069-42db-a3ae-907b1283d49e-kube-api-access\") pod \"installer-9-crc\" (UID: \"2b27ed82-f069-42db-a3ae-907b1283d49e\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 11 02:04:08 crc kubenswrapper[4824]: I1211 02:04:08.615342 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/2b27ed82-f069-42db-a3ae-907b1283d49e-var-lock\") pod \"installer-9-crc\" (UID: \"2b27ed82-f069-42db-a3ae-907b1283d49e\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 11 02:04:08 crc kubenswrapper[4824]: I1211 02:04:08.615482 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/2b27ed82-f069-42db-a3ae-907b1283d49e-var-lock\") pod \"installer-9-crc\" (UID: \"2b27ed82-f069-42db-a3ae-907b1283d49e\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 11 02:04:08 crc kubenswrapper[4824]: I1211 02:04:08.615617 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2b27ed82-f069-42db-a3ae-907b1283d49e-kubelet-dir\") pod \"installer-9-crc\" (UID: \"2b27ed82-f069-42db-a3ae-907b1283d49e\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 11 02:04:08 crc kubenswrapper[4824]: I1211 02:04:08.633057 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2b27ed82-f069-42db-a3ae-907b1283d49e-kube-api-access\") pod \"installer-9-crc\" (UID: \"2b27ed82-f069-42db-a3ae-907b1283d49e\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 11 02:04:08 crc kubenswrapper[4824]: I1211 02:04:08.688530 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 11 02:04:08 crc kubenswrapper[4824]: I1211 02:04:08.875374 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 11 02:04:09 crc kubenswrapper[4824]: I1211 02:04:09.805006 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"2b27ed82-f069-42db-a3ae-907b1283d49e","Type":"ContainerStarted","Data":"be38bc8fafc68eb0e4533c1738ee94821a0bd5ccad6b1145759cd08a07865e4c"} Dec 11 02:04:09 crc kubenswrapper[4824]: I1211 02:04:09.806054 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"2b27ed82-f069-42db-a3ae-907b1283d49e","Type":"ContainerStarted","Data":"43bac705fc828f2b56abaacd88464c66825a19d5f559a14a0d2f005c5bd0b7c3"} Dec 11 02:04:10 crc kubenswrapper[4824]: I1211 02:04:10.836910 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=2.836889528 podStartE2EDuration="2.836889528s" podCreationTimestamp="2025-12-11 02:04:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:04:10.834310411 +0000 UTC m=+192.523347830" watchObservedRunningTime="2025-12-11 02:04:10.836889528 +0000 UTC m=+192.525926907" Dec 11 02:04:15 crc kubenswrapper[4824]: I1211 02:04:15.927716 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-rnmtp" Dec 11 02:04:15 crc kubenswrapper[4824]: I1211 02:04:15.972284 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-rnmtp" Dec 11 02:04:18 crc kubenswrapper[4824]: I1211 02:04:18.872535 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ptkck" event={"ID":"9f40e0dc-3513-431e-95b3-af69c5cca852","Type":"ContainerStarted","Data":"a86925c4f82414c9d36757ccd7e92a47e9da8ea449470b638454774d0a02240e"} Dec 11 02:04:18 crc kubenswrapper[4824]: I1211 02:04:18.879834 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zfgl8" event={"ID":"a9bd6e85-2ed9-4525-9308-f6dd2f4eacda","Type":"ContainerStarted","Data":"e817d7850b869ef56a3c1e5bfddd7e9e1fcb4618675c57e8d7b7ef9f30fe668c"} Dec 11 02:04:18 crc kubenswrapper[4824]: I1211 02:04:18.882037 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jc49g" event={"ID":"6127fb96-7f67-453d-8d16-40fac34e6482","Type":"ContainerStarted","Data":"e284d9b7c962ec7af9fd503606c0088dbf83b7e2646b38c7bbb32a9ecf3b1445"} Dec 11 02:04:18 crc kubenswrapper[4824]: I1211 02:04:18.885384 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nqjh5" event={"ID":"d9806a3c-cb3d-4b3e-8889-11f8cf49c030","Type":"ContainerStarted","Data":"2a640e46ccf1141c27a92fc5e06a0e9ee5df1690b4df836c0188d0f31938d90d"} Dec 11 02:04:18 crc kubenswrapper[4824]: I1211 02:04:18.899098 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-86mf9" event={"ID":"b2b7c06f-4350-41be-b472-c75d21558911","Type":"ContainerStarted","Data":"bdab89470dc1e06cdad086566602a8595a319b4bed0607fc722d7a808b90c2f8"} Dec 11 02:04:19 crc kubenswrapper[4824]: I1211 02:04:19.907580 4824 generic.go:334] "Generic (PLEG): container finished" podID="a9bd6e85-2ed9-4525-9308-f6dd2f4eacda" containerID="e817d7850b869ef56a3c1e5bfddd7e9e1fcb4618675c57e8d7b7ef9f30fe668c" exitCode=0 Dec 11 02:04:19 crc kubenswrapper[4824]: I1211 02:04:19.907699 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zfgl8" event={"ID":"a9bd6e85-2ed9-4525-9308-f6dd2f4eacda","Type":"ContainerDied","Data":"e817d7850b869ef56a3c1e5bfddd7e9e1fcb4618675c57e8d7b7ef9f30fe668c"} Dec 11 02:04:19 crc kubenswrapper[4824]: I1211 02:04:19.910239 4824 generic.go:334] "Generic (PLEG): container finished" podID="6127fb96-7f67-453d-8d16-40fac34e6482" containerID="e284d9b7c962ec7af9fd503606c0088dbf83b7e2646b38c7bbb32a9ecf3b1445" exitCode=0 Dec 11 02:04:19 crc kubenswrapper[4824]: I1211 02:04:19.910323 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jc49g" event={"ID":"6127fb96-7f67-453d-8d16-40fac34e6482","Type":"ContainerDied","Data":"e284d9b7c962ec7af9fd503606c0088dbf83b7e2646b38c7bbb32a9ecf3b1445"} Dec 11 02:04:19 crc kubenswrapper[4824]: I1211 02:04:19.914172 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ktsvl" event={"ID":"36403c25-41c3-4014-b64d-5eea981691a4","Type":"ContainerStarted","Data":"b7aa8fbd046cac1564f79857c8e0309f8b729f197e5be40f5baf20a10e5f0b2b"} Dec 11 02:04:19 crc kubenswrapper[4824]: I1211 02:04:19.918935 4824 generic.go:334] "Generic (PLEG): container finished" podID="d9806a3c-cb3d-4b3e-8889-11f8cf49c030" containerID="2a640e46ccf1141c27a92fc5e06a0e9ee5df1690b4df836c0188d0f31938d90d" exitCode=0 Dec 11 02:04:19 crc kubenswrapper[4824]: I1211 02:04:19.919096 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nqjh5" event={"ID":"d9806a3c-cb3d-4b3e-8889-11f8cf49c030","Type":"ContainerDied","Data":"2a640e46ccf1141c27a92fc5e06a0e9ee5df1690b4df836c0188d0f31938d90d"} Dec 11 02:04:19 crc kubenswrapper[4824]: I1211 02:04:19.924179 4824 generic.go:334] "Generic (PLEG): container finished" podID="b2b7c06f-4350-41be-b472-c75d21558911" containerID="bdab89470dc1e06cdad086566602a8595a319b4bed0607fc722d7a808b90c2f8" exitCode=0 Dec 11 02:04:19 crc kubenswrapper[4824]: I1211 02:04:19.924717 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-86mf9" event={"ID":"b2b7c06f-4350-41be-b472-c75d21558911","Type":"ContainerDied","Data":"bdab89470dc1e06cdad086566602a8595a319b4bed0607fc722d7a808b90c2f8"} Dec 11 02:04:19 crc kubenswrapper[4824]: I1211 02:04:19.932164 4824 generic.go:334] "Generic (PLEG): container finished" podID="0a7ce62c-21db-428f-9b80-29ed781821e4" containerID="d64886df414a3ee9d5be209f89962d72eaab7c62bdb9e3ce0f32efbab607e1e2" exitCode=0 Dec 11 02:04:19 crc kubenswrapper[4824]: I1211 02:04:19.932307 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lbp88" event={"ID":"0a7ce62c-21db-428f-9b80-29ed781821e4","Type":"ContainerDied","Data":"d64886df414a3ee9d5be209f89962d72eaab7c62bdb9e3ce0f32efbab607e1e2"} Dec 11 02:04:19 crc kubenswrapper[4824]: I1211 02:04:19.944012 4824 generic.go:334] "Generic (PLEG): container finished" podID="9f40e0dc-3513-431e-95b3-af69c5cca852" containerID="a86925c4f82414c9d36757ccd7e92a47e9da8ea449470b638454774d0a02240e" exitCode=0 Dec 11 02:04:19 crc kubenswrapper[4824]: I1211 02:04:19.944082 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ptkck" event={"ID":"9f40e0dc-3513-431e-95b3-af69c5cca852","Type":"ContainerDied","Data":"a86925c4f82414c9d36757ccd7e92a47e9da8ea449470b638454774d0a02240e"} Dec 11 02:04:20 crc kubenswrapper[4824]: I1211 02:04:20.958208 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zfgl8" event={"ID":"a9bd6e85-2ed9-4525-9308-f6dd2f4eacda","Type":"ContainerStarted","Data":"b87a66e84175026b9e684782162b19fdd46a3c4b46e44e828dc4cf9e85d8d209"} Dec 11 02:04:20 crc kubenswrapper[4824]: I1211 02:04:20.960721 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jc49g" event={"ID":"6127fb96-7f67-453d-8d16-40fac34e6482","Type":"ContainerStarted","Data":"bd6609faa0c263d1acb6b286455bfc86a0fb4b396c42cf8eeb5c1a12469e1ec6"} Dec 11 02:04:20 crc kubenswrapper[4824]: I1211 02:04:20.962757 4824 generic.go:334] "Generic (PLEG): container finished" podID="36403c25-41c3-4014-b64d-5eea981691a4" containerID="b7aa8fbd046cac1564f79857c8e0309f8b729f197e5be40f5baf20a10e5f0b2b" exitCode=0 Dec 11 02:04:20 crc kubenswrapper[4824]: I1211 02:04:20.962806 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ktsvl" event={"ID":"36403c25-41c3-4014-b64d-5eea981691a4","Type":"ContainerDied","Data":"b7aa8fbd046cac1564f79857c8e0309f8b729f197e5be40f5baf20a10e5f0b2b"} Dec 11 02:04:20 crc kubenswrapper[4824]: I1211 02:04:20.966208 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nqjh5" event={"ID":"d9806a3c-cb3d-4b3e-8889-11f8cf49c030","Type":"ContainerStarted","Data":"0f6c6d568fd4ead23e24d85c0ced23037ff04686489fd1752516ab279b1e63b3"} Dec 11 02:04:20 crc kubenswrapper[4824]: I1211 02:04:20.971004 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-86mf9" event={"ID":"b2b7c06f-4350-41be-b472-c75d21558911","Type":"ContainerStarted","Data":"15910645eb6869cc7816a7bae3d43c849e26e86197c452fe3b6a9e0df7d31b82"} Dec 11 02:04:20 crc kubenswrapper[4824]: I1211 02:04:20.973090 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lbp88" event={"ID":"0a7ce62c-21db-428f-9b80-29ed781821e4","Type":"ContainerStarted","Data":"838ae2849505bbca80d522066bec9a30f20aeadc5bec9dc52e61d93579ae6688"} Dec 11 02:04:20 crc kubenswrapper[4824]: I1211 02:04:20.975297 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ptkck" event={"ID":"9f40e0dc-3513-431e-95b3-af69c5cca852","Type":"ContainerStarted","Data":"1c10a0cb8f6d6f5e190de4653f01335ff8a29b49ee4d32ca6652cc9a23584f0f"} Dec 11 02:04:20 crc kubenswrapper[4824]: I1211 02:04:20.977991 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zfgl8" podStartSLOduration=2.747434399 podStartE2EDuration="58.977976857s" podCreationTimestamp="2025-12-11 02:03:22 +0000 UTC" firstStartedPulling="2025-12-11 02:03:24.192168947 +0000 UTC m=+145.881206326" lastFinishedPulling="2025-12-11 02:04:20.422711395 +0000 UTC m=+202.111748784" observedRunningTime="2025-12-11 02:04:20.977345985 +0000 UTC m=+202.666383364" watchObservedRunningTime="2025-12-11 02:04:20.977976857 +0000 UTC m=+202.667014236" Dec 11 02:04:20 crc kubenswrapper[4824]: I1211 02:04:20.995568 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-ptkck" podStartSLOduration=2.676180038 podStartE2EDuration="58.995550702s" podCreationTimestamp="2025-12-11 02:03:22 +0000 UTC" firstStartedPulling="2025-12-11 02:03:24.188024833 +0000 UTC m=+145.877062212" lastFinishedPulling="2025-12-11 02:04:20.507395487 +0000 UTC m=+202.196432876" observedRunningTime="2025-12-11 02:04:20.991582858 +0000 UTC m=+202.680620237" watchObservedRunningTime="2025-12-11 02:04:20.995550702 +0000 UTC m=+202.684588081" Dec 11 02:04:21 crc kubenswrapper[4824]: I1211 02:04:21.041835 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-jc49g" podStartSLOduration=1.8681109 podStartE2EDuration="57.041805712s" podCreationTimestamp="2025-12-11 02:03:24 +0000 UTC" firstStartedPulling="2025-12-11 02:03:25.220385982 +0000 UTC m=+146.909423361" lastFinishedPulling="2025-12-11 02:04:20.394080784 +0000 UTC m=+202.083118173" observedRunningTime="2025-12-11 02:04:21.039561405 +0000 UTC m=+202.728598784" watchObservedRunningTime="2025-12-11 02:04:21.041805712 +0000 UTC m=+202.730843091" Dec 11 02:04:21 crc kubenswrapper[4824]: I1211 02:04:21.063432 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-nqjh5" podStartSLOduration=2.916191772 podStartE2EDuration="59.063414885s" podCreationTimestamp="2025-12-11 02:03:22 +0000 UTC" firstStartedPulling="2025-12-11 02:03:24.194964151 +0000 UTC m=+145.884001530" lastFinishedPulling="2025-12-11 02:04:20.342187254 +0000 UTC m=+202.031224643" observedRunningTime="2025-12-11 02:04:21.061414457 +0000 UTC m=+202.750451836" watchObservedRunningTime="2025-12-11 02:04:21.063414885 +0000 UTC m=+202.752452264" Dec 11 02:04:21 crc kubenswrapper[4824]: I1211 02:04:21.079410 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-86mf9" podStartSLOduration=2.733333493 podStartE2EDuration="59.079391596s" podCreationTimestamp="2025-12-11 02:03:22 +0000 UTC" firstStartedPulling="2025-12-11 02:03:24.207068765 +0000 UTC m=+145.896106144" lastFinishedPulling="2025-12-11 02:04:20.553126858 +0000 UTC m=+202.242164247" observedRunningTime="2025-12-11 02:04:21.078657331 +0000 UTC m=+202.767694710" watchObservedRunningTime="2025-12-11 02:04:21.079391596 +0000 UTC m=+202.768428965" Dec 11 02:04:22 crc kubenswrapper[4824]: I1211 02:04:22.501205 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-nqjh5" Dec 11 02:04:22 crc kubenswrapper[4824]: I1211 02:04:22.501846 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-nqjh5" Dec 11 02:04:22 crc kubenswrapper[4824]: I1211 02:04:22.659002 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-86mf9" Dec 11 02:04:22 crc kubenswrapper[4824]: I1211 02:04:22.659069 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-86mf9" Dec 11 02:04:22 crc kubenswrapper[4824]: I1211 02:04:22.852340 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-ptkck" Dec 11 02:04:22 crc kubenswrapper[4824]: I1211 02:04:22.852448 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-ptkck" Dec 11 02:04:22 crc kubenswrapper[4824]: I1211 02:04:22.920800 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-ptkck" Dec 11 02:04:22 crc kubenswrapper[4824]: I1211 02:04:22.950301 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-lbp88" podStartSLOduration=4.868488599 podStartE2EDuration="58.950269626s" podCreationTimestamp="2025-12-11 02:03:24 +0000 UTC" firstStartedPulling="2025-12-11 02:03:26.282162305 +0000 UTC m=+147.971199684" lastFinishedPulling="2025-12-11 02:04:20.363943322 +0000 UTC m=+202.052980711" observedRunningTime="2025-12-11 02:04:21.101468275 +0000 UTC m=+202.790505654" watchObservedRunningTime="2025-12-11 02:04:22.950269626 +0000 UTC m=+204.639307045" Dec 11 02:04:22 crc kubenswrapper[4824]: I1211 02:04:22.988574 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ktsvl" event={"ID":"36403c25-41c3-4014-b64d-5eea981691a4","Type":"ContainerStarted","Data":"5fb31abfbd4dc8340810cda7ebd1729d84cada42e461d977c634deaec92edc27"} Dec 11 02:04:23 crc kubenswrapper[4824]: I1211 02:04:23.011274 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-ktsvl" podStartSLOduration=3.3297720699999998 podStartE2EDuration="58.011247045s" podCreationTimestamp="2025-12-11 02:03:25 +0000 UTC" firstStartedPulling="2025-12-11 02:03:27.411433164 +0000 UTC m=+149.100470543" lastFinishedPulling="2025-12-11 02:04:22.092908139 +0000 UTC m=+203.781945518" observedRunningTime="2025-12-11 02:04:23.010628494 +0000 UTC m=+204.699665903" watchObservedRunningTime="2025-12-11 02:04:23.011247045 +0000 UTC m=+204.700284444" Dec 11 02:04:23 crc kubenswrapper[4824]: I1211 02:04:23.047521 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zfgl8" Dec 11 02:04:23 crc kubenswrapper[4824]: I1211 02:04:23.047617 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zfgl8" Dec 11 02:04:23 crc kubenswrapper[4824]: I1211 02:04:23.119733 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zfgl8" Dec 11 02:04:23 crc kubenswrapper[4824]: I1211 02:04:23.539454 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-nqjh5" podUID="d9806a3c-cb3d-4b3e-8889-11f8cf49c030" containerName="registry-server" probeResult="failure" output=< Dec 11 02:04:23 crc kubenswrapper[4824]: timeout: failed to connect service ":50051" within 1s Dec 11 02:04:23 crc kubenswrapper[4824]: > Dec 11 02:04:23 crc kubenswrapper[4824]: I1211 02:04:23.730449 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-86mf9" podUID="b2b7c06f-4350-41be-b472-c75d21558911" containerName="registry-server" probeResult="failure" output=< Dec 11 02:04:23 crc kubenswrapper[4824]: timeout: failed to connect service ":50051" within 1s Dec 11 02:04:23 crc kubenswrapper[4824]: > Dec 11 02:04:24 crc kubenswrapper[4824]: I1211 02:04:24.286171 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-l9d4f"] Dec 11 02:04:24 crc kubenswrapper[4824]: I1211 02:04:24.443485 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-jc49g" Dec 11 02:04:24 crc kubenswrapper[4824]: I1211 02:04:24.443556 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-jc49g" Dec 11 02:04:24 crc kubenswrapper[4824]: I1211 02:04:24.509142 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-jc49g" Dec 11 02:04:24 crc kubenswrapper[4824]: I1211 02:04:24.837811 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-lbp88" Dec 11 02:04:24 crc kubenswrapper[4824]: I1211 02:04:24.838396 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-lbp88" Dec 11 02:04:24 crc kubenswrapper[4824]: I1211 02:04:24.910745 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-lbp88" Dec 11 02:04:25 crc kubenswrapper[4824]: I1211 02:04:25.062629 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-jc49g" Dec 11 02:04:25 crc kubenswrapper[4824]: I1211 02:04:25.071871 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-lbp88" Dec 11 02:04:26 crc kubenswrapper[4824]: I1211 02:04:26.074037 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/community-operators-zfgl8" podUID="a9bd6e85-2ed9-4525-9308-f6dd2f4eacda" containerName="registry-server" probeResult="failure" output=< Dec 11 02:04:26 crc kubenswrapper[4824]: timeout: failed to connect service ":50051" within 1s Dec 11 02:04:26 crc kubenswrapper[4824]: > Dec 11 02:04:26 crc kubenswrapper[4824]: I1211 02:04:26.272525 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-ktsvl" Dec 11 02:04:26 crc kubenswrapper[4824]: I1211 02:04:26.273301 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-ktsvl" Dec 11 02:04:27 crc kubenswrapper[4824]: I1211 02:04:27.341477 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-ktsvl" podUID="36403c25-41c3-4014-b64d-5eea981691a4" containerName="registry-server" probeResult="failure" output=< Dec 11 02:04:27 crc kubenswrapper[4824]: timeout: failed to connect service ":50051" within 1s Dec 11 02:04:27 crc kubenswrapper[4824]: > Dec 11 02:04:28 crc kubenswrapper[4824]: I1211 02:04:28.677326 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lbp88"] Dec 11 02:04:28 crc kubenswrapper[4824]: I1211 02:04:28.677864 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-lbp88" podUID="0a7ce62c-21db-428f-9b80-29ed781821e4" containerName="registry-server" containerID="cri-o://838ae2849505bbca80d522066bec9a30f20aeadc5bec9dc52e61d93579ae6688" gracePeriod=2 Dec 11 02:04:29 crc kubenswrapper[4824]: I1211 02:04:29.631675 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lbp88" Dec 11 02:04:29 crc kubenswrapper[4824]: I1211 02:04:29.763762 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j2t5k\" (UniqueName: \"kubernetes.io/projected/0a7ce62c-21db-428f-9b80-29ed781821e4-kube-api-access-j2t5k\") pod \"0a7ce62c-21db-428f-9b80-29ed781821e4\" (UID: \"0a7ce62c-21db-428f-9b80-29ed781821e4\") " Dec 11 02:04:29 crc kubenswrapper[4824]: I1211 02:04:29.763980 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a7ce62c-21db-428f-9b80-29ed781821e4-catalog-content\") pod \"0a7ce62c-21db-428f-9b80-29ed781821e4\" (UID: \"0a7ce62c-21db-428f-9b80-29ed781821e4\") " Dec 11 02:04:29 crc kubenswrapper[4824]: I1211 02:04:29.764071 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a7ce62c-21db-428f-9b80-29ed781821e4-utilities\") pod \"0a7ce62c-21db-428f-9b80-29ed781821e4\" (UID: \"0a7ce62c-21db-428f-9b80-29ed781821e4\") " Dec 11 02:04:29 crc kubenswrapper[4824]: I1211 02:04:29.765932 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a7ce62c-21db-428f-9b80-29ed781821e4-utilities" (OuterVolumeSpecName: "utilities") pod "0a7ce62c-21db-428f-9b80-29ed781821e4" (UID: "0a7ce62c-21db-428f-9b80-29ed781821e4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:04:29 crc kubenswrapper[4824]: I1211 02:04:29.778540 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a7ce62c-21db-428f-9b80-29ed781821e4-kube-api-access-j2t5k" (OuterVolumeSpecName: "kube-api-access-j2t5k") pod "0a7ce62c-21db-428f-9b80-29ed781821e4" (UID: "0a7ce62c-21db-428f-9b80-29ed781821e4"). InnerVolumeSpecName "kube-api-access-j2t5k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:04:29 crc kubenswrapper[4824]: I1211 02:04:29.814935 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a7ce62c-21db-428f-9b80-29ed781821e4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0a7ce62c-21db-428f-9b80-29ed781821e4" (UID: "0a7ce62c-21db-428f-9b80-29ed781821e4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:04:29 crc kubenswrapper[4824]: I1211 02:04:29.866536 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a7ce62c-21db-428f-9b80-29ed781821e4-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 02:04:29 crc kubenswrapper[4824]: I1211 02:04:29.866568 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j2t5k\" (UniqueName: \"kubernetes.io/projected/0a7ce62c-21db-428f-9b80-29ed781821e4-kube-api-access-j2t5k\") on node \"crc\" DevicePath \"\"" Dec 11 02:04:29 crc kubenswrapper[4824]: I1211 02:04:29.866578 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a7ce62c-21db-428f-9b80-29ed781821e4-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 02:04:30 crc kubenswrapper[4824]: I1211 02:04:30.035365 4824 generic.go:334] "Generic (PLEG): container finished" podID="0a7ce62c-21db-428f-9b80-29ed781821e4" containerID="838ae2849505bbca80d522066bec9a30f20aeadc5bec9dc52e61d93579ae6688" exitCode=0 Dec 11 02:04:30 crc kubenswrapper[4824]: I1211 02:04:30.035431 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lbp88" event={"ID":"0a7ce62c-21db-428f-9b80-29ed781821e4","Type":"ContainerDied","Data":"838ae2849505bbca80d522066bec9a30f20aeadc5bec9dc52e61d93579ae6688"} Dec 11 02:04:30 crc kubenswrapper[4824]: I1211 02:04:30.035460 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lbp88" Dec 11 02:04:30 crc kubenswrapper[4824]: I1211 02:04:30.035490 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lbp88" event={"ID":"0a7ce62c-21db-428f-9b80-29ed781821e4","Type":"ContainerDied","Data":"49ccf53fae791aec1bb7e69c206eb84414e85056ede6ef8701888f633879bfa4"} Dec 11 02:04:30 crc kubenswrapper[4824]: I1211 02:04:30.035513 4824 scope.go:117] "RemoveContainer" containerID="838ae2849505bbca80d522066bec9a30f20aeadc5bec9dc52e61d93579ae6688" Dec 11 02:04:30 crc kubenswrapper[4824]: I1211 02:04:30.071606 4824 scope.go:117] "RemoveContainer" containerID="d64886df414a3ee9d5be209f89962d72eaab7c62bdb9e3ce0f32efbab607e1e2" Dec 11 02:04:30 crc kubenswrapper[4824]: I1211 02:04:30.080585 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lbp88"] Dec 11 02:04:30 crc kubenswrapper[4824]: I1211 02:04:30.083538 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-lbp88"] Dec 11 02:04:30 crc kubenswrapper[4824]: I1211 02:04:30.101607 4824 scope.go:117] "RemoveContainer" containerID="d74812b216e037d8f413c4e0aba5661d8b148c305340c78600f17fb29a79386b" Dec 11 02:04:30 crc kubenswrapper[4824]: I1211 02:04:30.113247 4824 scope.go:117] "RemoveContainer" containerID="838ae2849505bbca80d522066bec9a30f20aeadc5bec9dc52e61d93579ae6688" Dec 11 02:04:30 crc kubenswrapper[4824]: E1211 02:04:30.113691 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"838ae2849505bbca80d522066bec9a30f20aeadc5bec9dc52e61d93579ae6688\": container with ID starting with 838ae2849505bbca80d522066bec9a30f20aeadc5bec9dc52e61d93579ae6688 not found: ID does not exist" containerID="838ae2849505bbca80d522066bec9a30f20aeadc5bec9dc52e61d93579ae6688" Dec 11 02:04:30 crc kubenswrapper[4824]: I1211 02:04:30.113746 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"838ae2849505bbca80d522066bec9a30f20aeadc5bec9dc52e61d93579ae6688"} err="failed to get container status \"838ae2849505bbca80d522066bec9a30f20aeadc5bec9dc52e61d93579ae6688\": rpc error: code = NotFound desc = could not find container \"838ae2849505bbca80d522066bec9a30f20aeadc5bec9dc52e61d93579ae6688\": container with ID starting with 838ae2849505bbca80d522066bec9a30f20aeadc5bec9dc52e61d93579ae6688 not found: ID does not exist" Dec 11 02:04:30 crc kubenswrapper[4824]: I1211 02:04:30.113827 4824 scope.go:117] "RemoveContainer" containerID="d64886df414a3ee9d5be209f89962d72eaab7c62bdb9e3ce0f32efbab607e1e2" Dec 11 02:04:30 crc kubenswrapper[4824]: E1211 02:04:30.114148 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d64886df414a3ee9d5be209f89962d72eaab7c62bdb9e3ce0f32efbab607e1e2\": container with ID starting with d64886df414a3ee9d5be209f89962d72eaab7c62bdb9e3ce0f32efbab607e1e2 not found: ID does not exist" containerID="d64886df414a3ee9d5be209f89962d72eaab7c62bdb9e3ce0f32efbab607e1e2" Dec 11 02:04:30 crc kubenswrapper[4824]: I1211 02:04:30.114173 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d64886df414a3ee9d5be209f89962d72eaab7c62bdb9e3ce0f32efbab607e1e2"} err="failed to get container status \"d64886df414a3ee9d5be209f89962d72eaab7c62bdb9e3ce0f32efbab607e1e2\": rpc error: code = NotFound desc = could not find container \"d64886df414a3ee9d5be209f89962d72eaab7c62bdb9e3ce0f32efbab607e1e2\": container with ID starting with d64886df414a3ee9d5be209f89962d72eaab7c62bdb9e3ce0f32efbab607e1e2 not found: ID does not exist" Dec 11 02:04:30 crc kubenswrapper[4824]: I1211 02:04:30.114186 4824 scope.go:117] "RemoveContainer" containerID="d74812b216e037d8f413c4e0aba5661d8b148c305340c78600f17fb29a79386b" Dec 11 02:04:30 crc kubenswrapper[4824]: E1211 02:04:30.114563 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d74812b216e037d8f413c4e0aba5661d8b148c305340c78600f17fb29a79386b\": container with ID starting with d74812b216e037d8f413c4e0aba5661d8b148c305340c78600f17fb29a79386b not found: ID does not exist" containerID="d74812b216e037d8f413c4e0aba5661d8b148c305340c78600f17fb29a79386b" Dec 11 02:04:30 crc kubenswrapper[4824]: I1211 02:04:30.114616 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d74812b216e037d8f413c4e0aba5661d8b148c305340c78600f17fb29a79386b"} err="failed to get container status \"d74812b216e037d8f413c4e0aba5661d8b148c305340c78600f17fb29a79386b\": rpc error: code = NotFound desc = could not find container \"d74812b216e037d8f413c4e0aba5661d8b148c305340c78600f17fb29a79386b\": container with ID starting with d74812b216e037d8f413c4e0aba5661d8b148c305340c78600f17fb29a79386b not found: ID does not exist" Dec 11 02:04:30 crc kubenswrapper[4824]: I1211 02:04:30.648093 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a7ce62c-21db-428f-9b80-29ed781821e4" path="/var/lib/kubelet/pods/0a7ce62c-21db-428f-9b80-29ed781821e4/volumes" Dec 11 02:04:32 crc kubenswrapper[4824]: I1211 02:04:32.563924 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-nqjh5" Dec 11 02:04:32 crc kubenswrapper[4824]: I1211 02:04:32.628568 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-nqjh5" Dec 11 02:04:32 crc kubenswrapper[4824]: I1211 02:04:32.731199 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-86mf9" Dec 11 02:04:32 crc kubenswrapper[4824]: I1211 02:04:32.773844 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-86mf9" Dec 11 02:04:32 crc kubenswrapper[4824]: I1211 02:04:32.909908 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-ptkck" Dec 11 02:04:33 crc kubenswrapper[4824]: I1211 02:04:33.109346 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zfgl8" Dec 11 02:04:35 crc kubenswrapper[4824]: I1211 02:04:35.082622 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zfgl8"] Dec 11 02:04:35 crc kubenswrapper[4824]: I1211 02:04:35.083257 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-zfgl8" podUID="a9bd6e85-2ed9-4525-9308-f6dd2f4eacda" containerName="registry-server" containerID="cri-o://b87a66e84175026b9e684782162b19fdd46a3c4b46e44e828dc4cf9e85d8d209" gracePeriod=2 Dec 11 02:04:35 crc kubenswrapper[4824]: I1211 02:04:35.273180 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ptkck"] Dec 11 02:04:35 crc kubenswrapper[4824]: I1211 02:04:35.273485 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-ptkck" podUID="9f40e0dc-3513-431e-95b3-af69c5cca852" containerName="registry-server" containerID="cri-o://1c10a0cb8f6d6f5e190de4653f01335ff8a29b49ee4d32ca6652cc9a23584f0f" gracePeriod=2 Dec 11 02:04:36 crc kubenswrapper[4824]: I1211 02:04:36.330761 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-ktsvl" Dec 11 02:04:36 crc kubenswrapper[4824]: I1211 02:04:36.417313 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-ktsvl" Dec 11 02:04:36 crc kubenswrapper[4824]: I1211 02:04:36.617595 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zfgl8" Dec 11 02:04:36 crc kubenswrapper[4824]: I1211 02:04:36.731034 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ptkck" Dec 11 02:04:36 crc kubenswrapper[4824]: I1211 02:04:36.784934 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lb5d7\" (UniqueName: \"kubernetes.io/projected/a9bd6e85-2ed9-4525-9308-f6dd2f4eacda-kube-api-access-lb5d7\") pod \"a9bd6e85-2ed9-4525-9308-f6dd2f4eacda\" (UID: \"a9bd6e85-2ed9-4525-9308-f6dd2f4eacda\") " Dec 11 02:04:36 crc kubenswrapper[4824]: I1211 02:04:36.784988 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9bd6e85-2ed9-4525-9308-f6dd2f4eacda-catalog-content\") pod \"a9bd6e85-2ed9-4525-9308-f6dd2f4eacda\" (UID: \"a9bd6e85-2ed9-4525-9308-f6dd2f4eacda\") " Dec 11 02:04:36 crc kubenswrapper[4824]: I1211 02:04:36.785104 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9bd6e85-2ed9-4525-9308-f6dd2f4eacda-utilities\") pod \"a9bd6e85-2ed9-4525-9308-f6dd2f4eacda\" (UID: \"a9bd6e85-2ed9-4525-9308-f6dd2f4eacda\") " Dec 11 02:04:36 crc kubenswrapper[4824]: I1211 02:04:36.786026 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9bd6e85-2ed9-4525-9308-f6dd2f4eacda-utilities" (OuterVolumeSpecName: "utilities") pod "a9bd6e85-2ed9-4525-9308-f6dd2f4eacda" (UID: "a9bd6e85-2ed9-4525-9308-f6dd2f4eacda"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:04:36 crc kubenswrapper[4824]: I1211 02:04:36.790824 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9bd6e85-2ed9-4525-9308-f6dd2f4eacda-kube-api-access-lb5d7" (OuterVolumeSpecName: "kube-api-access-lb5d7") pod "a9bd6e85-2ed9-4525-9308-f6dd2f4eacda" (UID: "a9bd6e85-2ed9-4525-9308-f6dd2f4eacda"). InnerVolumeSpecName "kube-api-access-lb5d7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:04:36 crc kubenswrapper[4824]: I1211 02:04:36.835196 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9bd6e85-2ed9-4525-9308-f6dd2f4eacda-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a9bd6e85-2ed9-4525-9308-f6dd2f4eacda" (UID: "a9bd6e85-2ed9-4525-9308-f6dd2f4eacda"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:04:36 crc kubenswrapper[4824]: I1211 02:04:36.887832 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f40e0dc-3513-431e-95b3-af69c5cca852-utilities\") pod \"9f40e0dc-3513-431e-95b3-af69c5cca852\" (UID: \"9f40e0dc-3513-431e-95b3-af69c5cca852\") " Dec 11 02:04:36 crc kubenswrapper[4824]: I1211 02:04:36.887907 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f40e0dc-3513-431e-95b3-af69c5cca852-catalog-content\") pod \"9f40e0dc-3513-431e-95b3-af69c5cca852\" (UID: \"9f40e0dc-3513-431e-95b3-af69c5cca852\") " Dec 11 02:04:36 crc kubenswrapper[4824]: I1211 02:04:36.887991 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t6z6w\" (UniqueName: \"kubernetes.io/projected/9f40e0dc-3513-431e-95b3-af69c5cca852-kube-api-access-t6z6w\") pod \"9f40e0dc-3513-431e-95b3-af69c5cca852\" (UID: \"9f40e0dc-3513-431e-95b3-af69c5cca852\") " Dec 11 02:04:36 crc kubenswrapper[4824]: I1211 02:04:36.888368 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9bd6e85-2ed9-4525-9308-f6dd2f4eacda-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 02:04:36 crc kubenswrapper[4824]: I1211 02:04:36.888395 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lb5d7\" (UniqueName: \"kubernetes.io/projected/a9bd6e85-2ed9-4525-9308-f6dd2f4eacda-kube-api-access-lb5d7\") on node \"crc\" DevicePath \"\"" Dec 11 02:04:36 crc kubenswrapper[4824]: I1211 02:04:36.888409 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9bd6e85-2ed9-4525-9308-f6dd2f4eacda-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 02:04:36 crc kubenswrapper[4824]: I1211 02:04:36.889299 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9f40e0dc-3513-431e-95b3-af69c5cca852-utilities" (OuterVolumeSpecName: "utilities") pod "9f40e0dc-3513-431e-95b3-af69c5cca852" (UID: "9f40e0dc-3513-431e-95b3-af69c5cca852"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:04:36 crc kubenswrapper[4824]: I1211 02:04:36.892652 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f40e0dc-3513-431e-95b3-af69c5cca852-kube-api-access-t6z6w" (OuterVolumeSpecName: "kube-api-access-t6z6w") pod "9f40e0dc-3513-431e-95b3-af69c5cca852" (UID: "9f40e0dc-3513-431e-95b3-af69c5cca852"). InnerVolumeSpecName "kube-api-access-t6z6w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:04:36 crc kubenswrapper[4824]: I1211 02:04:36.957539 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9f40e0dc-3513-431e-95b3-af69c5cca852-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9f40e0dc-3513-431e-95b3-af69c5cca852" (UID: "9f40e0dc-3513-431e-95b3-af69c5cca852"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:04:36 crc kubenswrapper[4824]: I1211 02:04:36.989798 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f40e0dc-3513-431e-95b3-af69c5cca852-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 02:04:36 crc kubenswrapper[4824]: I1211 02:04:36.989836 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f40e0dc-3513-431e-95b3-af69c5cca852-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 02:04:36 crc kubenswrapper[4824]: I1211 02:04:36.989848 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t6z6w\" (UniqueName: \"kubernetes.io/projected/9f40e0dc-3513-431e-95b3-af69c5cca852-kube-api-access-t6z6w\") on node \"crc\" DevicePath \"\"" Dec 11 02:04:37 crc kubenswrapper[4824]: I1211 02:04:37.086436 4824 generic.go:334] "Generic (PLEG): container finished" podID="9f40e0dc-3513-431e-95b3-af69c5cca852" containerID="1c10a0cb8f6d6f5e190de4653f01335ff8a29b49ee4d32ca6652cc9a23584f0f" exitCode=0 Dec 11 02:04:37 crc kubenswrapper[4824]: I1211 02:04:37.086553 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ptkck" event={"ID":"9f40e0dc-3513-431e-95b3-af69c5cca852","Type":"ContainerDied","Data":"1c10a0cb8f6d6f5e190de4653f01335ff8a29b49ee4d32ca6652cc9a23584f0f"} Dec 11 02:04:37 crc kubenswrapper[4824]: I1211 02:04:37.086590 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ptkck" event={"ID":"9f40e0dc-3513-431e-95b3-af69c5cca852","Type":"ContainerDied","Data":"01a2f59c747f884738fabe124af1ffc22217be02033720fc6a5f7ff385165b96"} Dec 11 02:04:37 crc kubenswrapper[4824]: I1211 02:04:37.086612 4824 scope.go:117] "RemoveContainer" containerID="1c10a0cb8f6d6f5e190de4653f01335ff8a29b49ee4d32ca6652cc9a23584f0f" Dec 11 02:04:37 crc kubenswrapper[4824]: I1211 02:04:37.086736 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ptkck" Dec 11 02:04:37 crc kubenswrapper[4824]: I1211 02:04:37.091437 4824 generic.go:334] "Generic (PLEG): container finished" podID="a9bd6e85-2ed9-4525-9308-f6dd2f4eacda" containerID="b87a66e84175026b9e684782162b19fdd46a3c4b46e44e828dc4cf9e85d8d209" exitCode=0 Dec 11 02:04:37 crc kubenswrapper[4824]: I1211 02:04:37.091974 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zfgl8" Dec 11 02:04:37 crc kubenswrapper[4824]: I1211 02:04:37.092735 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zfgl8" event={"ID":"a9bd6e85-2ed9-4525-9308-f6dd2f4eacda","Type":"ContainerDied","Data":"b87a66e84175026b9e684782162b19fdd46a3c4b46e44e828dc4cf9e85d8d209"} Dec 11 02:04:37 crc kubenswrapper[4824]: I1211 02:04:37.092797 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zfgl8" event={"ID":"a9bd6e85-2ed9-4525-9308-f6dd2f4eacda","Type":"ContainerDied","Data":"67802c46fd3d9710d00d5a844f1f2f945cfa31af7c9b0a31c56e462c01d34d40"} Dec 11 02:04:37 crc kubenswrapper[4824]: I1211 02:04:37.113418 4824 scope.go:117] "RemoveContainer" containerID="a86925c4f82414c9d36757ccd7e92a47e9da8ea449470b638454774d0a02240e" Dec 11 02:04:37 crc kubenswrapper[4824]: I1211 02:04:37.134767 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ptkck"] Dec 11 02:04:37 crc kubenswrapper[4824]: I1211 02:04:37.145396 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-ptkck"] Dec 11 02:04:37 crc kubenswrapper[4824]: I1211 02:04:37.149100 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zfgl8"] Dec 11 02:04:37 crc kubenswrapper[4824]: I1211 02:04:37.151346 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-zfgl8"] Dec 11 02:04:37 crc kubenswrapper[4824]: I1211 02:04:37.170186 4824 scope.go:117] "RemoveContainer" containerID="18ec7c455c85f9cc7aad983ed5ae6aff2d9a0109928101d53120cb25c21cce9b" Dec 11 02:04:37 crc kubenswrapper[4824]: I1211 02:04:37.185203 4824 scope.go:117] "RemoveContainer" containerID="1c10a0cb8f6d6f5e190de4653f01335ff8a29b49ee4d32ca6652cc9a23584f0f" Dec 11 02:04:37 crc kubenswrapper[4824]: E1211 02:04:37.185600 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c10a0cb8f6d6f5e190de4653f01335ff8a29b49ee4d32ca6652cc9a23584f0f\": container with ID starting with 1c10a0cb8f6d6f5e190de4653f01335ff8a29b49ee4d32ca6652cc9a23584f0f not found: ID does not exist" containerID="1c10a0cb8f6d6f5e190de4653f01335ff8a29b49ee4d32ca6652cc9a23584f0f" Dec 11 02:04:37 crc kubenswrapper[4824]: I1211 02:04:37.185639 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c10a0cb8f6d6f5e190de4653f01335ff8a29b49ee4d32ca6652cc9a23584f0f"} err="failed to get container status \"1c10a0cb8f6d6f5e190de4653f01335ff8a29b49ee4d32ca6652cc9a23584f0f\": rpc error: code = NotFound desc = could not find container \"1c10a0cb8f6d6f5e190de4653f01335ff8a29b49ee4d32ca6652cc9a23584f0f\": container with ID starting with 1c10a0cb8f6d6f5e190de4653f01335ff8a29b49ee4d32ca6652cc9a23584f0f not found: ID does not exist" Dec 11 02:04:37 crc kubenswrapper[4824]: I1211 02:04:37.185673 4824 scope.go:117] "RemoveContainer" containerID="a86925c4f82414c9d36757ccd7e92a47e9da8ea449470b638454774d0a02240e" Dec 11 02:04:37 crc kubenswrapper[4824]: E1211 02:04:37.186257 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a86925c4f82414c9d36757ccd7e92a47e9da8ea449470b638454774d0a02240e\": container with ID starting with a86925c4f82414c9d36757ccd7e92a47e9da8ea449470b638454774d0a02240e not found: ID does not exist" containerID="a86925c4f82414c9d36757ccd7e92a47e9da8ea449470b638454774d0a02240e" Dec 11 02:04:37 crc kubenswrapper[4824]: I1211 02:04:37.186365 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a86925c4f82414c9d36757ccd7e92a47e9da8ea449470b638454774d0a02240e"} err="failed to get container status \"a86925c4f82414c9d36757ccd7e92a47e9da8ea449470b638454774d0a02240e\": rpc error: code = NotFound desc = could not find container \"a86925c4f82414c9d36757ccd7e92a47e9da8ea449470b638454774d0a02240e\": container with ID starting with a86925c4f82414c9d36757ccd7e92a47e9da8ea449470b638454774d0a02240e not found: ID does not exist" Dec 11 02:04:37 crc kubenswrapper[4824]: I1211 02:04:37.186490 4824 scope.go:117] "RemoveContainer" containerID="18ec7c455c85f9cc7aad983ed5ae6aff2d9a0109928101d53120cb25c21cce9b" Dec 11 02:04:37 crc kubenswrapper[4824]: E1211 02:04:37.186854 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"18ec7c455c85f9cc7aad983ed5ae6aff2d9a0109928101d53120cb25c21cce9b\": container with ID starting with 18ec7c455c85f9cc7aad983ed5ae6aff2d9a0109928101d53120cb25c21cce9b not found: ID does not exist" containerID="18ec7c455c85f9cc7aad983ed5ae6aff2d9a0109928101d53120cb25c21cce9b" Dec 11 02:04:37 crc kubenswrapper[4824]: I1211 02:04:37.186883 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"18ec7c455c85f9cc7aad983ed5ae6aff2d9a0109928101d53120cb25c21cce9b"} err="failed to get container status \"18ec7c455c85f9cc7aad983ed5ae6aff2d9a0109928101d53120cb25c21cce9b\": rpc error: code = NotFound desc = could not find container \"18ec7c455c85f9cc7aad983ed5ae6aff2d9a0109928101d53120cb25c21cce9b\": container with ID starting with 18ec7c455c85f9cc7aad983ed5ae6aff2d9a0109928101d53120cb25c21cce9b not found: ID does not exist" Dec 11 02:04:37 crc kubenswrapper[4824]: I1211 02:04:37.186902 4824 scope.go:117] "RemoveContainer" containerID="b87a66e84175026b9e684782162b19fdd46a3c4b46e44e828dc4cf9e85d8d209" Dec 11 02:04:37 crc kubenswrapper[4824]: I1211 02:04:37.203352 4824 scope.go:117] "RemoveContainer" containerID="e817d7850b869ef56a3c1e5bfddd7e9e1fcb4618675c57e8d7b7ef9f30fe668c" Dec 11 02:04:37 crc kubenswrapper[4824]: I1211 02:04:37.220416 4824 scope.go:117] "RemoveContainer" containerID="e1aad0c20b9c4d8ce6d5efd2b9be643373b60da8cce4695bf039ea40a61a66b8" Dec 11 02:04:37 crc kubenswrapper[4824]: I1211 02:04:37.239344 4824 scope.go:117] "RemoveContainer" containerID="b87a66e84175026b9e684782162b19fdd46a3c4b46e44e828dc4cf9e85d8d209" Dec 11 02:04:37 crc kubenswrapper[4824]: E1211 02:04:37.239899 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b87a66e84175026b9e684782162b19fdd46a3c4b46e44e828dc4cf9e85d8d209\": container with ID starting with b87a66e84175026b9e684782162b19fdd46a3c4b46e44e828dc4cf9e85d8d209 not found: ID does not exist" containerID="b87a66e84175026b9e684782162b19fdd46a3c4b46e44e828dc4cf9e85d8d209" Dec 11 02:04:37 crc kubenswrapper[4824]: I1211 02:04:37.239955 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b87a66e84175026b9e684782162b19fdd46a3c4b46e44e828dc4cf9e85d8d209"} err="failed to get container status \"b87a66e84175026b9e684782162b19fdd46a3c4b46e44e828dc4cf9e85d8d209\": rpc error: code = NotFound desc = could not find container \"b87a66e84175026b9e684782162b19fdd46a3c4b46e44e828dc4cf9e85d8d209\": container with ID starting with b87a66e84175026b9e684782162b19fdd46a3c4b46e44e828dc4cf9e85d8d209 not found: ID does not exist" Dec 11 02:04:37 crc kubenswrapper[4824]: I1211 02:04:37.240002 4824 scope.go:117] "RemoveContainer" containerID="e817d7850b869ef56a3c1e5bfddd7e9e1fcb4618675c57e8d7b7ef9f30fe668c" Dec 11 02:04:37 crc kubenswrapper[4824]: E1211 02:04:37.240372 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e817d7850b869ef56a3c1e5bfddd7e9e1fcb4618675c57e8d7b7ef9f30fe668c\": container with ID starting with e817d7850b869ef56a3c1e5bfddd7e9e1fcb4618675c57e8d7b7ef9f30fe668c not found: ID does not exist" containerID="e817d7850b869ef56a3c1e5bfddd7e9e1fcb4618675c57e8d7b7ef9f30fe668c" Dec 11 02:04:37 crc kubenswrapper[4824]: I1211 02:04:37.240403 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e817d7850b869ef56a3c1e5bfddd7e9e1fcb4618675c57e8d7b7ef9f30fe668c"} err="failed to get container status \"e817d7850b869ef56a3c1e5bfddd7e9e1fcb4618675c57e8d7b7ef9f30fe668c\": rpc error: code = NotFound desc = could not find container \"e817d7850b869ef56a3c1e5bfddd7e9e1fcb4618675c57e8d7b7ef9f30fe668c\": container with ID starting with e817d7850b869ef56a3c1e5bfddd7e9e1fcb4618675c57e8d7b7ef9f30fe668c not found: ID does not exist" Dec 11 02:04:37 crc kubenswrapper[4824]: I1211 02:04:37.240422 4824 scope.go:117] "RemoveContainer" containerID="e1aad0c20b9c4d8ce6d5efd2b9be643373b60da8cce4695bf039ea40a61a66b8" Dec 11 02:04:37 crc kubenswrapper[4824]: E1211 02:04:37.240846 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e1aad0c20b9c4d8ce6d5efd2b9be643373b60da8cce4695bf039ea40a61a66b8\": container with ID starting with e1aad0c20b9c4d8ce6d5efd2b9be643373b60da8cce4695bf039ea40a61a66b8 not found: ID does not exist" containerID="e1aad0c20b9c4d8ce6d5efd2b9be643373b60da8cce4695bf039ea40a61a66b8" Dec 11 02:04:37 crc kubenswrapper[4824]: I1211 02:04:37.240890 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1aad0c20b9c4d8ce6d5efd2b9be643373b60da8cce4695bf039ea40a61a66b8"} err="failed to get container status \"e1aad0c20b9c4d8ce6d5efd2b9be643373b60da8cce4695bf039ea40a61a66b8\": rpc error: code = NotFound desc = could not find container \"e1aad0c20b9c4d8ce6d5efd2b9be643373b60da8cce4695bf039ea40a61a66b8\": container with ID starting with e1aad0c20b9c4d8ce6d5efd2b9be643373b60da8cce4695bf039ea40a61a66b8 not found: ID does not exist" Dec 11 02:04:37 crc kubenswrapper[4824]: I1211 02:04:37.251700 4824 patch_prober.go:28] interesting pod/machine-config-daemon-gx6xt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 02:04:37 crc kubenswrapper[4824]: I1211 02:04:37.251775 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 02:04:37 crc kubenswrapper[4824]: I1211 02:04:37.251834 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" Dec 11 02:04:37 crc kubenswrapper[4824]: I1211 02:04:37.252793 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"dcf289ac27dbda4f24aabbde3fb5119329e7b4fc7d6e018168da9a69661a4c57"} pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 11 02:04:37 crc kubenswrapper[4824]: I1211 02:04:37.252888 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" containerName="machine-config-daemon" containerID="cri-o://dcf289ac27dbda4f24aabbde3fb5119329e7b4fc7d6e018168da9a69661a4c57" gracePeriod=600 Dec 11 02:04:38 crc kubenswrapper[4824]: I1211 02:04:38.104221 4824 generic.go:334] "Generic (PLEG): container finished" podID="44b156e4-64a4-4d45-aa5e-9b10a862faed" containerID="dcf289ac27dbda4f24aabbde3fb5119329e7b4fc7d6e018168da9a69661a4c57" exitCode=0 Dec 11 02:04:38 crc kubenswrapper[4824]: I1211 02:04:38.104321 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" event={"ID":"44b156e4-64a4-4d45-aa5e-9b10a862faed","Type":"ContainerDied","Data":"dcf289ac27dbda4f24aabbde3fb5119329e7b4fc7d6e018168da9a69661a4c57"} Dec 11 02:04:38 crc kubenswrapper[4824]: I1211 02:04:38.104390 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" event={"ID":"44b156e4-64a4-4d45-aa5e-9b10a862faed","Type":"ContainerStarted","Data":"e5bb671b5b5757c6365db5e523a726c313eb514acdb624a11ae5ddd89c2d5112"} Dec 11 02:04:38 crc kubenswrapper[4824]: I1211 02:04:38.645196 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f40e0dc-3513-431e-95b3-af69c5cca852" path="/var/lib/kubelet/pods/9f40e0dc-3513-431e-95b3-af69c5cca852/volumes" Dec 11 02:04:38 crc kubenswrapper[4824]: I1211 02:04:38.646768 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a9bd6e85-2ed9-4525-9308-f6dd2f4eacda" path="/var/lib/kubelet/pods/a9bd6e85-2ed9-4525-9308-f6dd2f4eacda/volumes" Dec 11 02:04:39 crc kubenswrapper[4824]: I1211 02:04:39.678186 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ktsvl"] Dec 11 02:04:39 crc kubenswrapper[4824]: I1211 02:04:39.679169 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-ktsvl" podUID="36403c25-41c3-4014-b64d-5eea981691a4" containerName="registry-server" containerID="cri-o://5fb31abfbd4dc8340810cda7ebd1729d84cada42e461d977c634deaec92edc27" gracePeriod=2 Dec 11 02:04:40 crc kubenswrapper[4824]: I1211 02:04:40.026633 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ktsvl" Dec 11 02:04:40 crc kubenswrapper[4824]: I1211 02:04:40.118752 4824 generic.go:334] "Generic (PLEG): container finished" podID="36403c25-41c3-4014-b64d-5eea981691a4" containerID="5fb31abfbd4dc8340810cda7ebd1729d84cada42e461d977c634deaec92edc27" exitCode=0 Dec 11 02:04:40 crc kubenswrapper[4824]: I1211 02:04:40.118797 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ktsvl" event={"ID":"36403c25-41c3-4014-b64d-5eea981691a4","Type":"ContainerDied","Data":"5fb31abfbd4dc8340810cda7ebd1729d84cada42e461d977c634deaec92edc27"} Dec 11 02:04:40 crc kubenswrapper[4824]: I1211 02:04:40.118830 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ktsvl" event={"ID":"36403c25-41c3-4014-b64d-5eea981691a4","Type":"ContainerDied","Data":"6137d9e1d216ff1ffb575a5aa83ea1f1c9c8cf6cdc6df5a7e3d222a01d5a97d7"} Dec 11 02:04:40 crc kubenswrapper[4824]: I1211 02:04:40.118842 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ktsvl" Dec 11 02:04:40 crc kubenswrapper[4824]: I1211 02:04:40.118850 4824 scope.go:117] "RemoveContainer" containerID="5fb31abfbd4dc8340810cda7ebd1729d84cada42e461d977c634deaec92edc27" Dec 11 02:04:40 crc kubenswrapper[4824]: I1211 02:04:40.136510 4824 scope.go:117] "RemoveContainer" containerID="b7aa8fbd046cac1564f79857c8e0309f8b729f197e5be40f5baf20a10e5f0b2b" Dec 11 02:04:40 crc kubenswrapper[4824]: I1211 02:04:40.136579 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/36403c25-41c3-4014-b64d-5eea981691a4-utilities\") pod \"36403c25-41c3-4014-b64d-5eea981691a4\" (UID: \"36403c25-41c3-4014-b64d-5eea981691a4\") " Dec 11 02:04:40 crc kubenswrapper[4824]: I1211 02:04:40.136665 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-968xz\" (UniqueName: \"kubernetes.io/projected/36403c25-41c3-4014-b64d-5eea981691a4-kube-api-access-968xz\") pod \"36403c25-41c3-4014-b64d-5eea981691a4\" (UID: \"36403c25-41c3-4014-b64d-5eea981691a4\") " Dec 11 02:04:40 crc kubenswrapper[4824]: I1211 02:04:40.136739 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/36403c25-41c3-4014-b64d-5eea981691a4-catalog-content\") pod \"36403c25-41c3-4014-b64d-5eea981691a4\" (UID: \"36403c25-41c3-4014-b64d-5eea981691a4\") " Dec 11 02:04:40 crc kubenswrapper[4824]: I1211 02:04:40.137416 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/36403c25-41c3-4014-b64d-5eea981691a4-utilities" (OuterVolumeSpecName: "utilities") pod "36403c25-41c3-4014-b64d-5eea981691a4" (UID: "36403c25-41c3-4014-b64d-5eea981691a4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:04:40 crc kubenswrapper[4824]: I1211 02:04:40.150320 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36403c25-41c3-4014-b64d-5eea981691a4-kube-api-access-968xz" (OuterVolumeSpecName: "kube-api-access-968xz") pod "36403c25-41c3-4014-b64d-5eea981691a4" (UID: "36403c25-41c3-4014-b64d-5eea981691a4"). InnerVolumeSpecName "kube-api-access-968xz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:04:40 crc kubenswrapper[4824]: I1211 02:04:40.158459 4824 scope.go:117] "RemoveContainer" containerID="d9ab1d347c29c4ec85f3a7faef64e75bf9d9e4d8b1b8422f530ca91921e902c0" Dec 11 02:04:40 crc kubenswrapper[4824]: I1211 02:04:40.181030 4824 scope.go:117] "RemoveContainer" containerID="5fb31abfbd4dc8340810cda7ebd1729d84cada42e461d977c634deaec92edc27" Dec 11 02:04:40 crc kubenswrapper[4824]: E1211 02:04:40.181555 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5fb31abfbd4dc8340810cda7ebd1729d84cada42e461d977c634deaec92edc27\": container with ID starting with 5fb31abfbd4dc8340810cda7ebd1729d84cada42e461d977c634deaec92edc27 not found: ID does not exist" containerID="5fb31abfbd4dc8340810cda7ebd1729d84cada42e461d977c634deaec92edc27" Dec 11 02:04:40 crc kubenswrapper[4824]: I1211 02:04:40.181588 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5fb31abfbd4dc8340810cda7ebd1729d84cada42e461d977c634deaec92edc27"} err="failed to get container status \"5fb31abfbd4dc8340810cda7ebd1729d84cada42e461d977c634deaec92edc27\": rpc error: code = NotFound desc = could not find container \"5fb31abfbd4dc8340810cda7ebd1729d84cada42e461d977c634deaec92edc27\": container with ID starting with 5fb31abfbd4dc8340810cda7ebd1729d84cada42e461d977c634deaec92edc27 not found: ID does not exist" Dec 11 02:04:40 crc kubenswrapper[4824]: I1211 02:04:40.181611 4824 scope.go:117] "RemoveContainer" containerID="b7aa8fbd046cac1564f79857c8e0309f8b729f197e5be40f5baf20a10e5f0b2b" Dec 11 02:04:40 crc kubenswrapper[4824]: E1211 02:04:40.182093 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b7aa8fbd046cac1564f79857c8e0309f8b729f197e5be40f5baf20a10e5f0b2b\": container with ID starting with b7aa8fbd046cac1564f79857c8e0309f8b729f197e5be40f5baf20a10e5f0b2b not found: ID does not exist" containerID="b7aa8fbd046cac1564f79857c8e0309f8b729f197e5be40f5baf20a10e5f0b2b" Dec 11 02:04:40 crc kubenswrapper[4824]: I1211 02:04:40.182190 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b7aa8fbd046cac1564f79857c8e0309f8b729f197e5be40f5baf20a10e5f0b2b"} err="failed to get container status \"b7aa8fbd046cac1564f79857c8e0309f8b729f197e5be40f5baf20a10e5f0b2b\": rpc error: code = NotFound desc = could not find container \"b7aa8fbd046cac1564f79857c8e0309f8b729f197e5be40f5baf20a10e5f0b2b\": container with ID starting with b7aa8fbd046cac1564f79857c8e0309f8b729f197e5be40f5baf20a10e5f0b2b not found: ID does not exist" Dec 11 02:04:40 crc kubenswrapper[4824]: I1211 02:04:40.182268 4824 scope.go:117] "RemoveContainer" containerID="d9ab1d347c29c4ec85f3a7faef64e75bf9d9e4d8b1b8422f530ca91921e902c0" Dec 11 02:04:40 crc kubenswrapper[4824]: E1211 02:04:40.182624 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d9ab1d347c29c4ec85f3a7faef64e75bf9d9e4d8b1b8422f530ca91921e902c0\": container with ID starting with d9ab1d347c29c4ec85f3a7faef64e75bf9d9e4d8b1b8422f530ca91921e902c0 not found: ID does not exist" containerID="d9ab1d347c29c4ec85f3a7faef64e75bf9d9e4d8b1b8422f530ca91921e902c0" Dec 11 02:04:40 crc kubenswrapper[4824]: I1211 02:04:40.182650 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9ab1d347c29c4ec85f3a7faef64e75bf9d9e4d8b1b8422f530ca91921e902c0"} err="failed to get container status \"d9ab1d347c29c4ec85f3a7faef64e75bf9d9e4d8b1b8422f530ca91921e902c0\": rpc error: code = NotFound desc = could not find container \"d9ab1d347c29c4ec85f3a7faef64e75bf9d9e4d8b1b8422f530ca91921e902c0\": container with ID starting with d9ab1d347c29c4ec85f3a7faef64e75bf9d9e4d8b1b8422f530ca91921e902c0 not found: ID does not exist" Dec 11 02:04:40 crc kubenswrapper[4824]: I1211 02:04:40.238231 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-968xz\" (UniqueName: \"kubernetes.io/projected/36403c25-41c3-4014-b64d-5eea981691a4-kube-api-access-968xz\") on node \"crc\" DevicePath \"\"" Dec 11 02:04:40 crc kubenswrapper[4824]: I1211 02:04:40.238516 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/36403c25-41c3-4014-b64d-5eea981691a4-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 02:04:40 crc kubenswrapper[4824]: I1211 02:04:40.246270 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/36403c25-41c3-4014-b64d-5eea981691a4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "36403c25-41c3-4014-b64d-5eea981691a4" (UID: "36403c25-41c3-4014-b64d-5eea981691a4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:04:40 crc kubenswrapper[4824]: I1211 02:04:40.340022 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/36403c25-41c3-4014-b64d-5eea981691a4-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 02:04:40 crc kubenswrapper[4824]: I1211 02:04:40.446213 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ktsvl"] Dec 11 02:04:40 crc kubenswrapper[4824]: I1211 02:04:40.449449 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-ktsvl"] Dec 11 02:04:40 crc kubenswrapper[4824]: I1211 02:04:40.640502 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="36403c25-41c3-4014-b64d-5eea981691a4" path="/var/lib/kubelet/pods/36403c25-41c3-4014-b64d-5eea981691a4/volumes" Dec 11 02:04:47 crc kubenswrapper[4824]: I1211 02:04:47.163392 4824 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 11 02:04:47 crc kubenswrapper[4824]: E1211 02:04:47.164680 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9bd6e85-2ed9-4525-9308-f6dd2f4eacda" containerName="extract-utilities" Dec 11 02:04:47 crc kubenswrapper[4824]: I1211 02:04:47.164713 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9bd6e85-2ed9-4525-9308-f6dd2f4eacda" containerName="extract-utilities" Dec 11 02:04:47 crc kubenswrapper[4824]: E1211 02:04:47.164737 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f40e0dc-3513-431e-95b3-af69c5cca852" containerName="extract-utilities" Dec 11 02:04:47 crc kubenswrapper[4824]: I1211 02:04:47.164753 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f40e0dc-3513-431e-95b3-af69c5cca852" containerName="extract-utilities" Dec 11 02:04:47 crc kubenswrapper[4824]: E1211 02:04:47.164778 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f40e0dc-3513-431e-95b3-af69c5cca852" containerName="registry-server" Dec 11 02:04:47 crc kubenswrapper[4824]: I1211 02:04:47.164791 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f40e0dc-3513-431e-95b3-af69c5cca852" containerName="registry-server" Dec 11 02:04:47 crc kubenswrapper[4824]: E1211 02:04:47.164817 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36403c25-41c3-4014-b64d-5eea981691a4" containerName="extract-content" Dec 11 02:04:47 crc kubenswrapper[4824]: I1211 02:04:47.164831 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="36403c25-41c3-4014-b64d-5eea981691a4" containerName="extract-content" Dec 11 02:04:47 crc kubenswrapper[4824]: E1211 02:04:47.164844 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9bd6e85-2ed9-4525-9308-f6dd2f4eacda" containerName="extract-content" Dec 11 02:04:47 crc kubenswrapper[4824]: I1211 02:04:47.164859 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9bd6e85-2ed9-4525-9308-f6dd2f4eacda" containerName="extract-content" Dec 11 02:04:47 crc kubenswrapper[4824]: E1211 02:04:47.164878 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a7ce62c-21db-428f-9b80-29ed781821e4" containerName="registry-server" Dec 11 02:04:47 crc kubenswrapper[4824]: I1211 02:04:47.164930 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a7ce62c-21db-428f-9b80-29ed781821e4" containerName="registry-server" Dec 11 02:04:47 crc kubenswrapper[4824]: E1211 02:04:47.164950 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36403c25-41c3-4014-b64d-5eea981691a4" containerName="registry-server" Dec 11 02:04:47 crc kubenswrapper[4824]: I1211 02:04:47.164962 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="36403c25-41c3-4014-b64d-5eea981691a4" containerName="registry-server" Dec 11 02:04:47 crc kubenswrapper[4824]: E1211 02:04:47.164982 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a7ce62c-21db-428f-9b80-29ed781821e4" containerName="extract-utilities" Dec 11 02:04:47 crc kubenswrapper[4824]: I1211 02:04:47.164994 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a7ce62c-21db-428f-9b80-29ed781821e4" containerName="extract-utilities" Dec 11 02:04:47 crc kubenswrapper[4824]: E1211 02:04:47.165013 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9bd6e85-2ed9-4525-9308-f6dd2f4eacda" containerName="registry-server" Dec 11 02:04:47 crc kubenswrapper[4824]: I1211 02:04:47.165028 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9bd6e85-2ed9-4525-9308-f6dd2f4eacda" containerName="registry-server" Dec 11 02:04:47 crc kubenswrapper[4824]: E1211 02:04:47.165046 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36403c25-41c3-4014-b64d-5eea981691a4" containerName="extract-utilities" Dec 11 02:04:47 crc kubenswrapper[4824]: I1211 02:04:47.165058 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="36403c25-41c3-4014-b64d-5eea981691a4" containerName="extract-utilities" Dec 11 02:04:47 crc kubenswrapper[4824]: E1211 02:04:47.165080 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a7ce62c-21db-428f-9b80-29ed781821e4" containerName="extract-content" Dec 11 02:04:47 crc kubenswrapper[4824]: I1211 02:04:47.165093 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a7ce62c-21db-428f-9b80-29ed781821e4" containerName="extract-content" Dec 11 02:04:47 crc kubenswrapper[4824]: E1211 02:04:47.165146 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f40e0dc-3513-431e-95b3-af69c5cca852" containerName="extract-content" Dec 11 02:04:47 crc kubenswrapper[4824]: I1211 02:04:47.165164 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f40e0dc-3513-431e-95b3-af69c5cca852" containerName="extract-content" Dec 11 02:04:47 crc kubenswrapper[4824]: I1211 02:04:47.165336 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="36403c25-41c3-4014-b64d-5eea981691a4" containerName="registry-server" Dec 11 02:04:47 crc kubenswrapper[4824]: I1211 02:04:47.165361 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f40e0dc-3513-431e-95b3-af69c5cca852" containerName="registry-server" Dec 11 02:04:47 crc kubenswrapper[4824]: I1211 02:04:47.165377 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9bd6e85-2ed9-4525-9308-f6dd2f4eacda" containerName="registry-server" Dec 11 02:04:47 crc kubenswrapper[4824]: I1211 02:04:47.165397 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a7ce62c-21db-428f-9b80-29ed781821e4" containerName="registry-server" Dec 11 02:04:47 crc kubenswrapper[4824]: I1211 02:04:47.166001 4824 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 11 02:04:47 crc kubenswrapper[4824]: I1211 02:04:47.166230 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 02:04:47 crc kubenswrapper[4824]: I1211 02:04:47.166587 4824 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 11 02:04:47 crc kubenswrapper[4824]: I1211 02:04:47.166652 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://c26f06d2ed5e6e98b61a51f073130675f2c49acda106b8aedcbb2a6d4f7580c1" gracePeriod=15 Dec 11 02:04:47 crc kubenswrapper[4824]: I1211 02:04:47.166707 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://20af9b10a67c8b21a7abcd23e42bc412798d159c8422e7c8ff58f5b997ca19a9" gracePeriod=15 Dec 11 02:04:47 crc kubenswrapper[4824]: I1211 02:04:47.166812 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://b7c766f8de73ca762ffde0c18b2d1c46bb152b2b37535a611c5a0e454a21b7d0" gracePeriod=15 Dec 11 02:04:47 crc kubenswrapper[4824]: I1211 02:04:47.166702 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://f14a940a72b2404a6cbd52a69e150b03266bcd7f5c1d3f4d07b51c0c69e51786" gracePeriod=15 Dec 11 02:04:47 crc kubenswrapper[4824]: I1211 02:04:47.166894 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://e9bbfc6edfe050b3fa9dbeee5f98d416bcb2a1533a38608a881b75396fd0b059" gracePeriod=15 Dec 11 02:04:47 crc kubenswrapper[4824]: E1211 02:04:47.168699 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 11 02:04:47 crc kubenswrapper[4824]: I1211 02:04:47.168748 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 11 02:04:47 crc kubenswrapper[4824]: E1211 02:04:47.168782 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 11 02:04:47 crc kubenswrapper[4824]: I1211 02:04:47.168798 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 11 02:04:47 crc kubenswrapper[4824]: E1211 02:04:47.168821 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 11 02:04:47 crc kubenswrapper[4824]: I1211 02:04:47.168837 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 11 02:04:47 crc kubenswrapper[4824]: E1211 02:04:47.168857 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 11 02:04:47 crc kubenswrapper[4824]: I1211 02:04:47.168873 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 11 02:04:47 crc kubenswrapper[4824]: E1211 02:04:47.168926 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 11 02:04:47 crc kubenswrapper[4824]: I1211 02:04:47.168943 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 11 02:04:47 crc kubenswrapper[4824]: E1211 02:04:47.168969 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 11 02:04:47 crc kubenswrapper[4824]: I1211 02:04:47.168985 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 11 02:04:47 crc kubenswrapper[4824]: E1211 02:04:47.169007 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 11 02:04:47 crc kubenswrapper[4824]: I1211 02:04:47.169022 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 11 02:04:47 crc kubenswrapper[4824]: I1211 02:04:47.169289 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 11 02:04:47 crc kubenswrapper[4824]: I1211 02:04:47.169323 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 11 02:04:47 crc kubenswrapper[4824]: I1211 02:04:47.169346 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 11 02:04:47 crc kubenswrapper[4824]: I1211 02:04:47.169365 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 11 02:04:47 crc kubenswrapper[4824]: I1211 02:04:47.169381 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 11 02:04:47 crc kubenswrapper[4824]: I1211 02:04:47.169399 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 11 02:04:47 crc kubenswrapper[4824]: I1211 02:04:47.171672 4824 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="f4b27818a5e8e43d0dc095d08835c792" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" Dec 11 02:04:47 crc kubenswrapper[4824]: I1211 02:04:47.350915 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 02:04:47 crc kubenswrapper[4824]: I1211 02:04:47.350973 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 02:04:47 crc kubenswrapper[4824]: I1211 02:04:47.350999 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 02:04:47 crc kubenswrapper[4824]: I1211 02:04:47.351073 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 02:04:47 crc kubenswrapper[4824]: I1211 02:04:47.351125 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 02:04:47 crc kubenswrapper[4824]: I1211 02:04:47.351153 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 02:04:47 crc kubenswrapper[4824]: I1211 02:04:47.351176 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 02:04:47 crc kubenswrapper[4824]: I1211 02:04:47.351237 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 02:04:47 crc kubenswrapper[4824]: I1211 02:04:47.452986 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 02:04:47 crc kubenswrapper[4824]: I1211 02:04:47.453176 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 02:04:47 crc kubenswrapper[4824]: I1211 02:04:47.453200 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 02:04:47 crc kubenswrapper[4824]: I1211 02:04:47.453272 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 02:04:47 crc kubenswrapper[4824]: I1211 02:04:47.453334 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 02:04:47 crc kubenswrapper[4824]: I1211 02:04:47.453394 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 02:04:47 crc kubenswrapper[4824]: I1211 02:04:47.453452 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 02:04:47 crc kubenswrapper[4824]: I1211 02:04:47.454569 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 02:04:47 crc kubenswrapper[4824]: I1211 02:04:47.453467 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 02:04:47 crc kubenswrapper[4824]: I1211 02:04:47.453517 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 02:04:47 crc kubenswrapper[4824]: I1211 02:04:47.453335 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 02:04:47 crc kubenswrapper[4824]: I1211 02:04:47.454686 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 02:04:47 crc kubenswrapper[4824]: I1211 02:04:47.454766 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 02:04:47 crc kubenswrapper[4824]: I1211 02:04:47.454788 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 02:04:47 crc kubenswrapper[4824]: I1211 02:04:47.454853 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 02:04:47 crc kubenswrapper[4824]: I1211 02:04:47.454906 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 02:04:48 crc kubenswrapper[4824]: I1211 02:04:48.215691 4824 generic.go:334] "Generic (PLEG): container finished" podID="2b27ed82-f069-42db-a3ae-907b1283d49e" containerID="be38bc8fafc68eb0e4533c1738ee94821a0bd5ccad6b1145759cd08a07865e4c" exitCode=0 Dec 11 02:04:48 crc kubenswrapper[4824]: I1211 02:04:48.215780 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"2b27ed82-f069-42db-a3ae-907b1283d49e","Type":"ContainerDied","Data":"be38bc8fafc68eb0e4533c1738ee94821a0bd5ccad6b1145759cd08a07865e4c"} Dec 11 02:04:48 crc kubenswrapper[4824]: I1211 02:04:48.217015 4824 status_manager.go:851] "Failed to get status for pod" podUID="2b27ed82-f069-42db-a3ae-907b1283d49e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 11 02:04:48 crc kubenswrapper[4824]: I1211 02:04:48.218212 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 11 02:04:48 crc kubenswrapper[4824]: I1211 02:04:48.219741 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 11 02:04:48 crc kubenswrapper[4824]: I1211 02:04:48.220536 4824 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="b7c766f8de73ca762ffde0c18b2d1c46bb152b2b37535a611c5a0e454a21b7d0" exitCode=0 Dec 11 02:04:48 crc kubenswrapper[4824]: I1211 02:04:48.220560 4824 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="f14a940a72b2404a6cbd52a69e150b03266bcd7f5c1d3f4d07b51c0c69e51786" exitCode=0 Dec 11 02:04:48 crc kubenswrapper[4824]: I1211 02:04:48.220571 4824 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="20af9b10a67c8b21a7abcd23e42bc412798d159c8422e7c8ff58f5b997ca19a9" exitCode=0 Dec 11 02:04:48 crc kubenswrapper[4824]: I1211 02:04:48.220583 4824 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="e9bbfc6edfe050b3fa9dbeee5f98d416bcb2a1533a38608a881b75396fd0b059" exitCode=2 Dec 11 02:04:48 crc kubenswrapper[4824]: I1211 02:04:48.220636 4824 scope.go:117] "RemoveContainer" containerID="8d45819f0611c91b20c83eeee1092031a3441fb69f3015241632ad419880227d" Dec 11 02:04:48 crc kubenswrapper[4824]: I1211 02:04:48.637311 4824 status_manager.go:851] "Failed to get status for pod" podUID="2b27ed82-f069-42db-a3ae-907b1283d49e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 11 02:04:49 crc kubenswrapper[4824]: I1211 02:04:49.233888 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 11 02:04:49 crc kubenswrapper[4824]: I1211 02:04:49.318503 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-l9d4f" podUID="1bdf5eb2-fe05-4e24-9e00-43c72afc3583" containerName="oauth-openshift" containerID="cri-o://0121a6485cec61a8b44c16765f66c94217718710b924d2ae72ff108600ccfa3e" gracePeriod=15 Dec 11 02:04:49 crc kubenswrapper[4824]: I1211 02:04:49.632076 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 11 02:04:49 crc kubenswrapper[4824]: I1211 02:04:49.633244 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 02:04:49 crc kubenswrapper[4824]: I1211 02:04:49.633918 4824 status_manager.go:851] "Failed to get status for pod" podUID="2b27ed82-f069-42db-a3ae-907b1283d49e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 11 02:04:49 crc kubenswrapper[4824]: I1211 02:04:49.634638 4824 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 11 02:04:49 crc kubenswrapper[4824]: I1211 02:04:49.635599 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 11 02:04:49 crc kubenswrapper[4824]: I1211 02:04:49.636047 4824 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 11 02:04:49 crc kubenswrapper[4824]: I1211 02:04:49.636384 4824 status_manager.go:851] "Failed to get status for pod" podUID="2b27ed82-f069-42db-a3ae-907b1283d49e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 11 02:04:49 crc kubenswrapper[4824]: I1211 02:04:49.803609 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2b27ed82-f069-42db-a3ae-907b1283d49e-kubelet-dir\") pod \"2b27ed82-f069-42db-a3ae-907b1283d49e\" (UID: \"2b27ed82-f069-42db-a3ae-907b1283d49e\") " Dec 11 02:04:49 crc kubenswrapper[4824]: I1211 02:04:49.803799 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 11 02:04:49 crc kubenswrapper[4824]: I1211 02:04:49.803821 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2b27ed82-f069-42db-a3ae-907b1283d49e-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "2b27ed82-f069-42db-a3ae-907b1283d49e" (UID: "2b27ed82-f069-42db-a3ae-907b1283d49e"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 02:04:49 crc kubenswrapper[4824]: I1211 02:04:49.803963 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/2b27ed82-f069-42db-a3ae-907b1283d49e-var-lock\") pod \"2b27ed82-f069-42db-a3ae-907b1283d49e\" (UID: \"2b27ed82-f069-42db-a3ae-907b1283d49e\") " Dec 11 02:04:49 crc kubenswrapper[4824]: I1211 02:04:49.803985 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 02:04:49 crc kubenswrapper[4824]: I1211 02:04:49.804063 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 11 02:04:49 crc kubenswrapper[4824]: I1211 02:04:49.804099 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2b27ed82-f069-42db-a3ae-907b1283d49e-var-lock" (OuterVolumeSpecName: "var-lock") pod "2b27ed82-f069-42db-a3ae-907b1283d49e" (UID: "2b27ed82-f069-42db-a3ae-907b1283d49e"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 02:04:49 crc kubenswrapper[4824]: I1211 02:04:49.804224 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 02:04:49 crc kubenswrapper[4824]: I1211 02:04:49.804248 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 11 02:04:49 crc kubenswrapper[4824]: I1211 02:04:49.804338 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 02:04:49 crc kubenswrapper[4824]: I1211 02:04:49.804370 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2b27ed82-f069-42db-a3ae-907b1283d49e-kube-api-access\") pod \"2b27ed82-f069-42db-a3ae-907b1283d49e\" (UID: \"2b27ed82-f069-42db-a3ae-907b1283d49e\") " Dec 11 02:04:49 crc kubenswrapper[4824]: I1211 02:04:49.805180 4824 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 11 02:04:49 crc kubenswrapper[4824]: I1211 02:04:49.805222 4824 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/2b27ed82-f069-42db-a3ae-907b1283d49e-var-lock\") on node \"crc\" DevicePath \"\"" Dec 11 02:04:49 crc kubenswrapper[4824]: I1211 02:04:49.805245 4824 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Dec 11 02:04:49 crc kubenswrapper[4824]: I1211 02:04:49.805269 4824 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 11 02:04:49 crc kubenswrapper[4824]: I1211 02:04:49.805287 4824 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2b27ed82-f069-42db-a3ae-907b1283d49e-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 11 02:04:49 crc kubenswrapper[4824]: I1211 02:04:49.812749 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b27ed82-f069-42db-a3ae-907b1283d49e-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "2b27ed82-f069-42db-a3ae-907b1283d49e" (UID: "2b27ed82-f069-42db-a3ae-907b1283d49e"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:04:49 crc kubenswrapper[4824]: I1211 02:04:49.906947 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2b27ed82-f069-42db-a3ae-907b1283d49e-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 11 02:04:50 crc kubenswrapper[4824]: I1211 02:04:50.251521 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 11 02:04:50 crc kubenswrapper[4824]: I1211 02:04:50.253481 4824 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="c26f06d2ed5e6e98b61a51f073130675f2c49acda106b8aedcbb2a6d4f7580c1" exitCode=0 Dec 11 02:04:50 crc kubenswrapper[4824]: I1211 02:04:50.253602 4824 scope.go:117] "RemoveContainer" containerID="b7c766f8de73ca762ffde0c18b2d1c46bb152b2b37535a611c5a0e454a21b7d0" Dec 11 02:04:50 crc kubenswrapper[4824]: I1211 02:04:50.253824 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 02:04:50 crc kubenswrapper[4824]: I1211 02:04:50.256730 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"2b27ed82-f069-42db-a3ae-907b1283d49e","Type":"ContainerDied","Data":"43bac705fc828f2b56abaacd88464c66825a19d5f559a14a0d2f005c5bd0b7c3"} Dec 11 02:04:50 crc kubenswrapper[4824]: I1211 02:04:50.256827 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="43bac705fc828f2b56abaacd88464c66825a19d5f559a14a0d2f005c5bd0b7c3" Dec 11 02:04:50 crc kubenswrapper[4824]: I1211 02:04:50.256957 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 11 02:04:50 crc kubenswrapper[4824]: I1211 02:04:50.268827 4824 generic.go:334] "Generic (PLEG): container finished" podID="1bdf5eb2-fe05-4e24-9e00-43c72afc3583" containerID="0121a6485cec61a8b44c16765f66c94217718710b924d2ae72ff108600ccfa3e" exitCode=0 Dec 11 02:04:50 crc kubenswrapper[4824]: I1211 02:04:50.268881 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-l9d4f" event={"ID":"1bdf5eb2-fe05-4e24-9e00-43c72afc3583","Type":"ContainerDied","Data":"0121a6485cec61a8b44c16765f66c94217718710b924d2ae72ff108600ccfa3e"} Dec 11 02:04:50 crc kubenswrapper[4824]: I1211 02:04:50.283591 4824 status_manager.go:851] "Failed to get status for pod" podUID="2b27ed82-f069-42db-a3ae-907b1283d49e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 11 02:04:50 crc kubenswrapper[4824]: I1211 02:04:50.283795 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-l9d4f" Dec 11 02:04:50 crc kubenswrapper[4824]: I1211 02:04:50.284007 4824 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 11 02:04:50 crc kubenswrapper[4824]: I1211 02:04:50.284495 4824 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 11 02:04:50 crc kubenswrapper[4824]: I1211 02:04:50.284874 4824 status_manager.go:851] "Failed to get status for pod" podUID="2b27ed82-f069-42db-a3ae-907b1283d49e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 11 02:04:50 crc kubenswrapper[4824]: I1211 02:04:50.285099 4824 status_manager.go:851] "Failed to get status for pod" podUID="1bdf5eb2-fe05-4e24-9e00-43c72afc3583" pod="openshift-authentication/oauth-openshift-558db77b4-l9d4f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-l9d4f\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 11 02:04:50 crc kubenswrapper[4824]: I1211 02:04:50.293106 4824 status_manager.go:851] "Failed to get status for pod" podUID="1bdf5eb2-fe05-4e24-9e00-43c72afc3583" pod="openshift-authentication/oauth-openshift-558db77b4-l9d4f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-l9d4f\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 11 02:04:50 crc kubenswrapper[4824]: I1211 02:04:50.293493 4824 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 11 02:04:50 crc kubenswrapper[4824]: I1211 02:04:50.294009 4824 status_manager.go:851] "Failed to get status for pod" podUID="2b27ed82-f069-42db-a3ae-907b1283d49e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 11 02:04:50 crc kubenswrapper[4824]: I1211 02:04:50.294858 4824 scope.go:117] "RemoveContainer" containerID="f14a940a72b2404a6cbd52a69e150b03266bcd7f5c1d3f4d07b51c0c69e51786" Dec 11 02:04:50 crc kubenswrapper[4824]: I1211 02:04:50.315190 4824 scope.go:117] "RemoveContainer" containerID="20af9b10a67c8b21a7abcd23e42bc412798d159c8422e7c8ff58f5b997ca19a9" Dec 11 02:04:50 crc kubenswrapper[4824]: I1211 02:04:50.332439 4824 scope.go:117] "RemoveContainer" containerID="e9bbfc6edfe050b3fa9dbeee5f98d416bcb2a1533a38608a881b75396fd0b059" Dec 11 02:04:50 crc kubenswrapper[4824]: I1211 02:04:50.349003 4824 scope.go:117] "RemoveContainer" containerID="c26f06d2ed5e6e98b61a51f073130675f2c49acda106b8aedcbb2a6d4f7580c1" Dec 11 02:04:50 crc kubenswrapper[4824]: I1211 02:04:50.372237 4824 scope.go:117] "RemoveContainer" containerID="bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62" Dec 11 02:04:50 crc kubenswrapper[4824]: I1211 02:04:50.401697 4824 scope.go:117] "RemoveContainer" containerID="b7c766f8de73ca762ffde0c18b2d1c46bb152b2b37535a611c5a0e454a21b7d0" Dec 11 02:04:50 crc kubenswrapper[4824]: E1211 02:04:50.402670 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b7c766f8de73ca762ffde0c18b2d1c46bb152b2b37535a611c5a0e454a21b7d0\": container with ID starting with b7c766f8de73ca762ffde0c18b2d1c46bb152b2b37535a611c5a0e454a21b7d0 not found: ID does not exist" containerID="b7c766f8de73ca762ffde0c18b2d1c46bb152b2b37535a611c5a0e454a21b7d0" Dec 11 02:04:50 crc kubenswrapper[4824]: I1211 02:04:50.402755 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b7c766f8de73ca762ffde0c18b2d1c46bb152b2b37535a611c5a0e454a21b7d0"} err="failed to get container status \"b7c766f8de73ca762ffde0c18b2d1c46bb152b2b37535a611c5a0e454a21b7d0\": rpc error: code = NotFound desc = could not find container \"b7c766f8de73ca762ffde0c18b2d1c46bb152b2b37535a611c5a0e454a21b7d0\": container with ID starting with b7c766f8de73ca762ffde0c18b2d1c46bb152b2b37535a611c5a0e454a21b7d0 not found: ID does not exist" Dec 11 02:04:50 crc kubenswrapper[4824]: I1211 02:04:50.402796 4824 scope.go:117] "RemoveContainer" containerID="f14a940a72b2404a6cbd52a69e150b03266bcd7f5c1d3f4d07b51c0c69e51786" Dec 11 02:04:50 crc kubenswrapper[4824]: E1211 02:04:50.403192 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f14a940a72b2404a6cbd52a69e150b03266bcd7f5c1d3f4d07b51c0c69e51786\": container with ID starting with f14a940a72b2404a6cbd52a69e150b03266bcd7f5c1d3f4d07b51c0c69e51786 not found: ID does not exist" containerID="f14a940a72b2404a6cbd52a69e150b03266bcd7f5c1d3f4d07b51c0c69e51786" Dec 11 02:04:50 crc kubenswrapper[4824]: I1211 02:04:50.403233 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f14a940a72b2404a6cbd52a69e150b03266bcd7f5c1d3f4d07b51c0c69e51786"} err="failed to get container status \"f14a940a72b2404a6cbd52a69e150b03266bcd7f5c1d3f4d07b51c0c69e51786\": rpc error: code = NotFound desc = could not find container \"f14a940a72b2404a6cbd52a69e150b03266bcd7f5c1d3f4d07b51c0c69e51786\": container with ID starting with f14a940a72b2404a6cbd52a69e150b03266bcd7f5c1d3f4d07b51c0c69e51786 not found: ID does not exist" Dec 11 02:04:50 crc kubenswrapper[4824]: I1211 02:04:50.403264 4824 scope.go:117] "RemoveContainer" containerID="20af9b10a67c8b21a7abcd23e42bc412798d159c8422e7c8ff58f5b997ca19a9" Dec 11 02:04:50 crc kubenswrapper[4824]: E1211 02:04:50.403613 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"20af9b10a67c8b21a7abcd23e42bc412798d159c8422e7c8ff58f5b997ca19a9\": container with ID starting with 20af9b10a67c8b21a7abcd23e42bc412798d159c8422e7c8ff58f5b997ca19a9 not found: ID does not exist" containerID="20af9b10a67c8b21a7abcd23e42bc412798d159c8422e7c8ff58f5b997ca19a9" Dec 11 02:04:50 crc kubenswrapper[4824]: I1211 02:04:50.403644 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20af9b10a67c8b21a7abcd23e42bc412798d159c8422e7c8ff58f5b997ca19a9"} err="failed to get container status \"20af9b10a67c8b21a7abcd23e42bc412798d159c8422e7c8ff58f5b997ca19a9\": rpc error: code = NotFound desc = could not find container \"20af9b10a67c8b21a7abcd23e42bc412798d159c8422e7c8ff58f5b997ca19a9\": container with ID starting with 20af9b10a67c8b21a7abcd23e42bc412798d159c8422e7c8ff58f5b997ca19a9 not found: ID does not exist" Dec 11 02:04:50 crc kubenswrapper[4824]: I1211 02:04:50.403664 4824 scope.go:117] "RemoveContainer" containerID="e9bbfc6edfe050b3fa9dbeee5f98d416bcb2a1533a38608a881b75396fd0b059" Dec 11 02:04:50 crc kubenswrapper[4824]: E1211 02:04:50.403977 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e9bbfc6edfe050b3fa9dbeee5f98d416bcb2a1533a38608a881b75396fd0b059\": container with ID starting with e9bbfc6edfe050b3fa9dbeee5f98d416bcb2a1533a38608a881b75396fd0b059 not found: ID does not exist" containerID="e9bbfc6edfe050b3fa9dbeee5f98d416bcb2a1533a38608a881b75396fd0b059" Dec 11 02:04:50 crc kubenswrapper[4824]: I1211 02:04:50.404017 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9bbfc6edfe050b3fa9dbeee5f98d416bcb2a1533a38608a881b75396fd0b059"} err="failed to get container status \"e9bbfc6edfe050b3fa9dbeee5f98d416bcb2a1533a38608a881b75396fd0b059\": rpc error: code = NotFound desc = could not find container \"e9bbfc6edfe050b3fa9dbeee5f98d416bcb2a1533a38608a881b75396fd0b059\": container with ID starting with e9bbfc6edfe050b3fa9dbeee5f98d416bcb2a1533a38608a881b75396fd0b059 not found: ID does not exist" Dec 11 02:04:50 crc kubenswrapper[4824]: I1211 02:04:50.404045 4824 scope.go:117] "RemoveContainer" containerID="c26f06d2ed5e6e98b61a51f073130675f2c49acda106b8aedcbb2a6d4f7580c1" Dec 11 02:04:50 crc kubenswrapper[4824]: E1211 02:04:50.404693 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c26f06d2ed5e6e98b61a51f073130675f2c49acda106b8aedcbb2a6d4f7580c1\": container with ID starting with c26f06d2ed5e6e98b61a51f073130675f2c49acda106b8aedcbb2a6d4f7580c1 not found: ID does not exist" containerID="c26f06d2ed5e6e98b61a51f073130675f2c49acda106b8aedcbb2a6d4f7580c1" Dec 11 02:04:50 crc kubenswrapper[4824]: I1211 02:04:50.404744 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c26f06d2ed5e6e98b61a51f073130675f2c49acda106b8aedcbb2a6d4f7580c1"} err="failed to get container status \"c26f06d2ed5e6e98b61a51f073130675f2c49acda106b8aedcbb2a6d4f7580c1\": rpc error: code = NotFound desc = could not find container \"c26f06d2ed5e6e98b61a51f073130675f2c49acda106b8aedcbb2a6d4f7580c1\": container with ID starting with c26f06d2ed5e6e98b61a51f073130675f2c49acda106b8aedcbb2a6d4f7580c1 not found: ID does not exist" Dec 11 02:04:50 crc kubenswrapper[4824]: I1211 02:04:50.404784 4824 scope.go:117] "RemoveContainer" containerID="bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62" Dec 11 02:04:50 crc kubenswrapper[4824]: E1211 02:04:50.405234 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62\": container with ID starting with bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62 not found: ID does not exist" containerID="bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62" Dec 11 02:04:50 crc kubenswrapper[4824]: I1211 02:04:50.405280 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62"} err="failed to get container status \"bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62\": rpc error: code = NotFound desc = could not find container \"bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62\": container with ID starting with bdef14c7d164a8b5697d0ec0e44647429389f370bccd81a01e00a6c46b110c62 not found: ID does not exist" Dec 11 02:04:50 crc kubenswrapper[4824]: I1211 02:04:50.414952 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/1bdf5eb2-fe05-4e24-9e00-43c72afc3583-v4-0-config-system-router-certs\") pod \"1bdf5eb2-fe05-4e24-9e00-43c72afc3583\" (UID: \"1bdf5eb2-fe05-4e24-9e00-43c72afc3583\") " Dec 11 02:04:50 crc kubenswrapper[4824]: I1211 02:04:50.415012 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/1bdf5eb2-fe05-4e24-9e00-43c72afc3583-v4-0-config-system-serving-cert\") pod \"1bdf5eb2-fe05-4e24-9e00-43c72afc3583\" (UID: \"1bdf5eb2-fe05-4e24-9e00-43c72afc3583\") " Dec 11 02:04:50 crc kubenswrapper[4824]: I1211 02:04:50.415048 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/1bdf5eb2-fe05-4e24-9e00-43c72afc3583-v4-0-config-user-template-error\") pod \"1bdf5eb2-fe05-4e24-9e00-43c72afc3583\" (UID: \"1bdf5eb2-fe05-4e24-9e00-43c72afc3583\") " Dec 11 02:04:50 crc kubenswrapper[4824]: I1211 02:04:50.415084 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/1bdf5eb2-fe05-4e24-9e00-43c72afc3583-v4-0-config-user-idp-0-file-data\") pod \"1bdf5eb2-fe05-4e24-9e00-43c72afc3583\" (UID: \"1bdf5eb2-fe05-4e24-9e00-43c72afc3583\") " Dec 11 02:04:50 crc kubenswrapper[4824]: I1211 02:04:50.415171 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/1bdf5eb2-fe05-4e24-9e00-43c72afc3583-v4-0-config-system-cliconfig\") pod \"1bdf5eb2-fe05-4e24-9e00-43c72afc3583\" (UID: \"1bdf5eb2-fe05-4e24-9e00-43c72afc3583\") " Dec 11 02:04:50 crc kubenswrapper[4824]: I1211 02:04:50.415203 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/1bdf5eb2-fe05-4e24-9e00-43c72afc3583-v4-0-config-user-template-login\") pod \"1bdf5eb2-fe05-4e24-9e00-43c72afc3583\" (UID: \"1bdf5eb2-fe05-4e24-9e00-43c72afc3583\") " Dec 11 02:04:50 crc kubenswrapper[4824]: I1211 02:04:50.415227 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/1bdf5eb2-fe05-4e24-9e00-43c72afc3583-v4-0-config-system-ocp-branding-template\") pod \"1bdf5eb2-fe05-4e24-9e00-43c72afc3583\" (UID: \"1bdf5eb2-fe05-4e24-9e00-43c72afc3583\") " Dec 11 02:04:50 crc kubenswrapper[4824]: I1211 02:04:50.415263 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/1bdf5eb2-fe05-4e24-9e00-43c72afc3583-v4-0-config-system-service-ca\") pod \"1bdf5eb2-fe05-4e24-9e00-43c72afc3583\" (UID: \"1bdf5eb2-fe05-4e24-9e00-43c72afc3583\") " Dec 11 02:04:50 crc kubenswrapper[4824]: I1211 02:04:50.415286 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n28fr\" (UniqueName: \"kubernetes.io/projected/1bdf5eb2-fe05-4e24-9e00-43c72afc3583-kube-api-access-n28fr\") pod \"1bdf5eb2-fe05-4e24-9e00-43c72afc3583\" (UID: \"1bdf5eb2-fe05-4e24-9e00-43c72afc3583\") " Dec 11 02:04:50 crc kubenswrapper[4824]: I1211 02:04:50.415332 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1bdf5eb2-fe05-4e24-9e00-43c72afc3583-audit-dir\") pod \"1bdf5eb2-fe05-4e24-9e00-43c72afc3583\" (UID: \"1bdf5eb2-fe05-4e24-9e00-43c72afc3583\") " Dec 11 02:04:50 crc kubenswrapper[4824]: I1211 02:04:50.415356 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/1bdf5eb2-fe05-4e24-9e00-43c72afc3583-audit-policies\") pod \"1bdf5eb2-fe05-4e24-9e00-43c72afc3583\" (UID: \"1bdf5eb2-fe05-4e24-9e00-43c72afc3583\") " Dec 11 02:04:50 crc kubenswrapper[4824]: I1211 02:04:50.415375 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/1bdf5eb2-fe05-4e24-9e00-43c72afc3583-v4-0-config-system-session\") pod \"1bdf5eb2-fe05-4e24-9e00-43c72afc3583\" (UID: \"1bdf5eb2-fe05-4e24-9e00-43c72afc3583\") " Dec 11 02:04:50 crc kubenswrapper[4824]: I1211 02:04:50.415413 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/1bdf5eb2-fe05-4e24-9e00-43c72afc3583-v4-0-config-user-template-provider-selection\") pod \"1bdf5eb2-fe05-4e24-9e00-43c72afc3583\" (UID: \"1bdf5eb2-fe05-4e24-9e00-43c72afc3583\") " Dec 11 02:04:50 crc kubenswrapper[4824]: I1211 02:04:50.415437 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bdf5eb2-fe05-4e24-9e00-43c72afc3583-v4-0-config-system-trusted-ca-bundle\") pod \"1bdf5eb2-fe05-4e24-9e00-43c72afc3583\" (UID: \"1bdf5eb2-fe05-4e24-9e00-43c72afc3583\") " Dec 11 02:04:50 crc kubenswrapper[4824]: I1211 02:04:50.416053 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bdf5eb2-fe05-4e24-9e00-43c72afc3583-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "1bdf5eb2-fe05-4e24-9e00-43c72afc3583" (UID: "1bdf5eb2-fe05-4e24-9e00-43c72afc3583"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:04:50 crc kubenswrapper[4824]: I1211 02:04:50.416122 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bdf5eb2-fe05-4e24-9e00-43c72afc3583-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "1bdf5eb2-fe05-4e24-9e00-43c72afc3583" (UID: "1bdf5eb2-fe05-4e24-9e00-43c72afc3583"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:04:50 crc kubenswrapper[4824]: I1211 02:04:50.416530 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1bdf5eb2-fe05-4e24-9e00-43c72afc3583-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "1bdf5eb2-fe05-4e24-9e00-43c72afc3583" (UID: "1bdf5eb2-fe05-4e24-9e00-43c72afc3583"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 02:04:50 crc kubenswrapper[4824]: I1211 02:04:50.416936 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bdf5eb2-fe05-4e24-9e00-43c72afc3583-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "1bdf5eb2-fe05-4e24-9e00-43c72afc3583" (UID: "1bdf5eb2-fe05-4e24-9e00-43c72afc3583"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:04:50 crc kubenswrapper[4824]: I1211 02:04:50.417561 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bdf5eb2-fe05-4e24-9e00-43c72afc3583-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "1bdf5eb2-fe05-4e24-9e00-43c72afc3583" (UID: "1bdf5eb2-fe05-4e24-9e00-43c72afc3583"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:04:50 crc kubenswrapper[4824]: I1211 02:04:50.420879 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bdf5eb2-fe05-4e24-9e00-43c72afc3583-kube-api-access-n28fr" (OuterVolumeSpecName: "kube-api-access-n28fr") pod "1bdf5eb2-fe05-4e24-9e00-43c72afc3583" (UID: "1bdf5eb2-fe05-4e24-9e00-43c72afc3583"). InnerVolumeSpecName "kube-api-access-n28fr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:04:50 crc kubenswrapper[4824]: I1211 02:04:50.421285 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bdf5eb2-fe05-4e24-9e00-43c72afc3583-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "1bdf5eb2-fe05-4e24-9e00-43c72afc3583" (UID: "1bdf5eb2-fe05-4e24-9e00-43c72afc3583"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:04:50 crc kubenswrapper[4824]: I1211 02:04:50.421606 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bdf5eb2-fe05-4e24-9e00-43c72afc3583-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "1bdf5eb2-fe05-4e24-9e00-43c72afc3583" (UID: "1bdf5eb2-fe05-4e24-9e00-43c72afc3583"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:04:50 crc kubenswrapper[4824]: I1211 02:04:50.421778 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bdf5eb2-fe05-4e24-9e00-43c72afc3583-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "1bdf5eb2-fe05-4e24-9e00-43c72afc3583" (UID: "1bdf5eb2-fe05-4e24-9e00-43c72afc3583"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:04:50 crc kubenswrapper[4824]: I1211 02:04:50.421999 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bdf5eb2-fe05-4e24-9e00-43c72afc3583-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "1bdf5eb2-fe05-4e24-9e00-43c72afc3583" (UID: "1bdf5eb2-fe05-4e24-9e00-43c72afc3583"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:04:50 crc kubenswrapper[4824]: I1211 02:04:50.422313 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bdf5eb2-fe05-4e24-9e00-43c72afc3583-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "1bdf5eb2-fe05-4e24-9e00-43c72afc3583" (UID: "1bdf5eb2-fe05-4e24-9e00-43c72afc3583"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:04:50 crc kubenswrapper[4824]: I1211 02:04:50.422701 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bdf5eb2-fe05-4e24-9e00-43c72afc3583-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "1bdf5eb2-fe05-4e24-9e00-43c72afc3583" (UID: "1bdf5eb2-fe05-4e24-9e00-43c72afc3583"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:04:50 crc kubenswrapper[4824]: I1211 02:04:50.422879 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bdf5eb2-fe05-4e24-9e00-43c72afc3583-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "1bdf5eb2-fe05-4e24-9e00-43c72afc3583" (UID: "1bdf5eb2-fe05-4e24-9e00-43c72afc3583"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:04:50 crc kubenswrapper[4824]: I1211 02:04:50.425585 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bdf5eb2-fe05-4e24-9e00-43c72afc3583-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "1bdf5eb2-fe05-4e24-9e00-43c72afc3583" (UID: "1bdf5eb2-fe05-4e24-9e00-43c72afc3583"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:04:50 crc kubenswrapper[4824]: I1211 02:04:50.516691 4824 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1bdf5eb2-fe05-4e24-9e00-43c72afc3583-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 11 02:04:50 crc kubenswrapper[4824]: I1211 02:04:50.516745 4824 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/1bdf5eb2-fe05-4e24-9e00-43c72afc3583-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 11 02:04:50 crc kubenswrapper[4824]: I1211 02:04:50.516766 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/1bdf5eb2-fe05-4e24-9e00-43c72afc3583-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 11 02:04:50 crc kubenswrapper[4824]: I1211 02:04:50.516782 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/1bdf5eb2-fe05-4e24-9e00-43c72afc3583-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 11 02:04:50 crc kubenswrapper[4824]: I1211 02:04:50.516796 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bdf5eb2-fe05-4e24-9e00-43c72afc3583-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 02:04:50 crc kubenswrapper[4824]: I1211 02:04:50.516810 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/1bdf5eb2-fe05-4e24-9e00-43c72afc3583-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 11 02:04:50 crc kubenswrapper[4824]: I1211 02:04:50.516822 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/1bdf5eb2-fe05-4e24-9e00-43c72afc3583-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 02:04:50 crc kubenswrapper[4824]: I1211 02:04:50.516834 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/1bdf5eb2-fe05-4e24-9e00-43c72afc3583-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 11 02:04:50 crc kubenswrapper[4824]: I1211 02:04:50.516846 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/1bdf5eb2-fe05-4e24-9e00-43c72afc3583-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 11 02:04:50 crc kubenswrapper[4824]: I1211 02:04:50.516858 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/1bdf5eb2-fe05-4e24-9e00-43c72afc3583-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 11 02:04:50 crc kubenswrapper[4824]: I1211 02:04:50.516869 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/1bdf5eb2-fe05-4e24-9e00-43c72afc3583-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 11 02:04:50 crc kubenswrapper[4824]: I1211 02:04:50.516882 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/1bdf5eb2-fe05-4e24-9e00-43c72afc3583-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 11 02:04:50 crc kubenswrapper[4824]: I1211 02:04:50.516894 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/1bdf5eb2-fe05-4e24-9e00-43c72afc3583-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 11 02:04:50 crc kubenswrapper[4824]: I1211 02:04:50.516909 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n28fr\" (UniqueName: \"kubernetes.io/projected/1bdf5eb2-fe05-4e24-9e00-43c72afc3583-kube-api-access-n28fr\") on node \"crc\" DevicePath \"\"" Dec 11 02:04:50 crc kubenswrapper[4824]: I1211 02:04:50.644492 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Dec 11 02:04:50 crc kubenswrapper[4824]: E1211 02:04:50.856976 4824 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 11 02:04:50 crc kubenswrapper[4824]: E1211 02:04:50.857532 4824 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 11 02:04:50 crc kubenswrapper[4824]: E1211 02:04:50.857960 4824 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 11 02:04:50 crc kubenswrapper[4824]: E1211 02:04:50.858933 4824 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 11 02:04:50 crc kubenswrapper[4824]: E1211 02:04:50.859542 4824 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 11 02:04:50 crc kubenswrapper[4824]: I1211 02:04:50.859649 4824 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Dec 11 02:04:50 crc kubenswrapper[4824]: E1211 02:04:50.860051 4824 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.50:6443: connect: connection refused" interval="200ms" Dec 11 02:04:51 crc kubenswrapper[4824]: E1211 02:04:51.060781 4824 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.50:6443: connect: connection refused" interval="400ms" Dec 11 02:04:51 crc kubenswrapper[4824]: I1211 02:04:51.280599 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-l9d4f" event={"ID":"1bdf5eb2-fe05-4e24-9e00-43c72afc3583","Type":"ContainerDied","Data":"71533e4d94be5bde493c3aad37cdc1a4a3ef65e95bc5078dfd4145af73c5a227"} Dec 11 02:04:51 crc kubenswrapper[4824]: I1211 02:04:51.280667 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-l9d4f" Dec 11 02:04:51 crc kubenswrapper[4824]: I1211 02:04:51.281010 4824 scope.go:117] "RemoveContainer" containerID="0121a6485cec61a8b44c16765f66c94217718710b924d2ae72ff108600ccfa3e" Dec 11 02:04:51 crc kubenswrapper[4824]: I1211 02:04:51.282056 4824 status_manager.go:851] "Failed to get status for pod" podUID="1bdf5eb2-fe05-4e24-9e00-43c72afc3583" pod="openshift-authentication/oauth-openshift-558db77b4-l9d4f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-l9d4f\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 11 02:04:51 crc kubenswrapper[4824]: I1211 02:04:51.282949 4824 status_manager.go:851] "Failed to get status for pod" podUID="2b27ed82-f069-42db-a3ae-907b1283d49e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 11 02:04:51 crc kubenswrapper[4824]: I1211 02:04:51.289609 4824 status_manager.go:851] "Failed to get status for pod" podUID="1bdf5eb2-fe05-4e24-9e00-43c72afc3583" pod="openshift-authentication/oauth-openshift-558db77b4-l9d4f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-l9d4f\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 11 02:04:51 crc kubenswrapper[4824]: I1211 02:04:51.290502 4824 status_manager.go:851] "Failed to get status for pod" podUID="2b27ed82-f069-42db-a3ae-907b1283d49e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 11 02:04:51 crc kubenswrapper[4824]: E1211 02:04:51.462170 4824 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.50:6443: connect: connection refused" interval="800ms" Dec 11 02:04:52 crc kubenswrapper[4824]: E1211 02:04:52.226606 4824 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.50:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 02:04:52 crc kubenswrapper[4824]: I1211 02:04:52.227375 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 02:04:52 crc kubenswrapper[4824]: W1211 02:04:52.258813 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-aef07959276c06d97b42b787dadb33518a95a371cef4e95588b5cac0818a0133 WatchSource:0}: Error finding container aef07959276c06d97b42b787dadb33518a95a371cef4e95588b5cac0818a0133: Status 404 returned error can't find the container with id aef07959276c06d97b42b787dadb33518a95a371cef4e95588b5cac0818a0133 Dec 11 02:04:52 crc kubenswrapper[4824]: E1211 02:04:52.262876 4824 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.50:6443: connect: connection refused" interval="1.6s" Dec 11 02:04:52 crc kubenswrapper[4824]: E1211 02:04:52.263572 4824 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.50:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.188006f27025279a openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 02:04:52.262569882 +0000 UTC m=+233.951607291,LastTimestamp:2025-12-11 02:04:52.262569882 +0000 UTC m=+233.951607291,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 02:04:52 crc kubenswrapper[4824]: I1211 02:04:52.301320 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"aef07959276c06d97b42b787dadb33518a95a371cef4e95588b5cac0818a0133"} Dec 11 02:04:53 crc kubenswrapper[4824]: I1211 02:04:53.313724 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"05d0fb77af8d72804d94abf2e96c18aa970ee327ca6a5a6fc8fe725eeb741dba"} Dec 11 02:04:53 crc kubenswrapper[4824]: E1211 02:04:53.315405 4824 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.50:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 02:04:53 crc kubenswrapper[4824]: I1211 02:04:53.315758 4824 status_manager.go:851] "Failed to get status for pod" podUID="1bdf5eb2-fe05-4e24-9e00-43c72afc3583" pod="openshift-authentication/oauth-openshift-558db77b4-l9d4f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-l9d4f\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 11 02:04:53 crc kubenswrapper[4824]: I1211 02:04:53.316519 4824 status_manager.go:851] "Failed to get status for pod" podUID="2b27ed82-f069-42db-a3ae-907b1283d49e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 11 02:04:53 crc kubenswrapper[4824]: E1211 02:04:53.864000 4824 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.50:6443: connect: connection refused" interval="3.2s" Dec 11 02:04:54 crc kubenswrapper[4824]: E1211 02:04:54.321894 4824 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.50:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 02:04:57 crc kubenswrapper[4824]: E1211 02:04:57.065626 4824 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.50:6443: connect: connection refused" interval="6.4s" Dec 11 02:04:58 crc kubenswrapper[4824]: E1211 02:04:58.246977 4824 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.50:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.188006f27025279a openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 02:04:52.262569882 +0000 UTC m=+233.951607291,LastTimestamp:2025-12-11 02:04:52.262569882 +0000 UTC m=+233.951607291,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 02:04:58 crc kubenswrapper[4824]: I1211 02:04:58.639904 4824 status_manager.go:851] "Failed to get status for pod" podUID="1bdf5eb2-fe05-4e24-9e00-43c72afc3583" pod="openshift-authentication/oauth-openshift-558db77b4-l9d4f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-l9d4f\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 11 02:04:58 crc kubenswrapper[4824]: I1211 02:04:58.640865 4824 status_manager.go:851] "Failed to get status for pod" podUID="2b27ed82-f069-42db-a3ae-907b1283d49e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 11 02:04:59 crc kubenswrapper[4824]: I1211 02:04:59.632500 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 02:04:59 crc kubenswrapper[4824]: I1211 02:04:59.634756 4824 status_manager.go:851] "Failed to get status for pod" podUID="1bdf5eb2-fe05-4e24-9e00-43c72afc3583" pod="openshift-authentication/oauth-openshift-558db77b4-l9d4f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-l9d4f\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 11 02:04:59 crc kubenswrapper[4824]: I1211 02:04:59.635210 4824 status_manager.go:851] "Failed to get status for pod" podUID="2b27ed82-f069-42db-a3ae-907b1283d49e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 11 02:04:59 crc kubenswrapper[4824]: I1211 02:04:59.649974 4824 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="beab386b-6c65-4bc3-9ef0-dc3f68b4403f" Dec 11 02:04:59 crc kubenswrapper[4824]: I1211 02:04:59.650011 4824 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="beab386b-6c65-4bc3-9ef0-dc3f68b4403f" Dec 11 02:04:59 crc kubenswrapper[4824]: E1211 02:04:59.650459 4824 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 02:04:59 crc kubenswrapper[4824]: I1211 02:04:59.650894 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 02:04:59 crc kubenswrapper[4824]: W1211 02:04:59.674618 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-bdb488b95da10bbd6157b370c9f0b8c06a24c6e27acaab53e5a18df1234ce907 WatchSource:0}: Error finding container bdb488b95da10bbd6157b370c9f0b8c06a24c6e27acaab53e5a18df1234ce907: Status 404 returned error can't find the container with id bdb488b95da10bbd6157b370c9f0b8c06a24c6e27acaab53e5a18df1234ce907 Dec 11 02:05:00 crc kubenswrapper[4824]: I1211 02:05:00.366602 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 11 02:05:00 crc kubenswrapper[4824]: I1211 02:05:00.367095 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"922af7f67b99a0ce08776f3f75c0c377e6ea69af1720b6178a6d1e2f9b4b8ba3"} Dec 11 02:05:00 crc kubenswrapper[4824]: I1211 02:05:00.367893 4824 scope.go:117] "RemoveContainer" containerID="922af7f67b99a0ce08776f3f75c0c377e6ea69af1720b6178a6d1e2f9b4b8ba3" Dec 11 02:05:00 crc kubenswrapper[4824]: I1211 02:05:00.368407 4824 status_manager.go:851] "Failed to get status for pod" podUID="1bdf5eb2-fe05-4e24-9e00-43c72afc3583" pod="openshift-authentication/oauth-openshift-558db77b4-l9d4f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-l9d4f\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 11 02:05:00 crc kubenswrapper[4824]: I1211 02:05:00.367029 4824 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="922af7f67b99a0ce08776f3f75c0c377e6ea69af1720b6178a6d1e2f9b4b8ba3" exitCode=1 Dec 11 02:05:00 crc kubenswrapper[4824]: I1211 02:05:00.370065 4824 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 11 02:05:00 crc kubenswrapper[4824]: I1211 02:05:00.370566 4824 status_manager.go:851] "Failed to get status for pod" podUID="2b27ed82-f069-42db-a3ae-907b1283d49e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 11 02:05:00 crc kubenswrapper[4824]: I1211 02:05:00.372489 4824 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="f911c1c00e10a173524d9801ed0d510deabda2b33f3c8e142cf738d02396edca" exitCode=0 Dec 11 02:05:00 crc kubenswrapper[4824]: I1211 02:05:00.372563 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"f911c1c00e10a173524d9801ed0d510deabda2b33f3c8e142cf738d02396edca"} Dec 11 02:05:00 crc kubenswrapper[4824]: I1211 02:05:00.372609 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"bdb488b95da10bbd6157b370c9f0b8c06a24c6e27acaab53e5a18df1234ce907"} Dec 11 02:05:00 crc kubenswrapper[4824]: I1211 02:05:00.373036 4824 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="beab386b-6c65-4bc3-9ef0-dc3f68b4403f" Dec 11 02:05:00 crc kubenswrapper[4824]: I1211 02:05:00.373071 4824 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="beab386b-6c65-4bc3-9ef0-dc3f68b4403f" Dec 11 02:05:00 crc kubenswrapper[4824]: E1211 02:05:00.373528 4824 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 02:05:00 crc kubenswrapper[4824]: I1211 02:05:00.374253 4824 status_manager.go:851] "Failed to get status for pod" podUID="2b27ed82-f069-42db-a3ae-907b1283d49e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 11 02:05:00 crc kubenswrapper[4824]: I1211 02:05:00.374781 4824 status_manager.go:851] "Failed to get status for pod" podUID="1bdf5eb2-fe05-4e24-9e00-43c72afc3583" pod="openshift-authentication/oauth-openshift-558db77b4-l9d4f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-l9d4f\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 11 02:05:00 crc kubenswrapper[4824]: I1211 02:05:00.375312 4824 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 11 02:05:01 crc kubenswrapper[4824]: I1211 02:05:01.388456 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 11 02:05:01 crc kubenswrapper[4824]: I1211 02:05:01.388887 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"905c456d05b56778448af338a340744c24962427d762e5ff4a2844213f67c402"} Dec 11 02:05:01 crc kubenswrapper[4824]: I1211 02:05:01.391710 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"5cbdf1c043397ae6c8a6d77570bf8e23f48c96b026e4abb704e5b7f47dab3853"} Dec 11 02:05:01 crc kubenswrapper[4824]: I1211 02:05:01.391735 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"59620049e7528c21b602e8cc871d4edd8f988691852dcc1b072598107f7440f1"} Dec 11 02:05:01 crc kubenswrapper[4824]: I1211 02:05:01.391746 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"c7457afb400e82e87beb461ff650b5acd50b2ac401fc9f062fc8d8bbf4244288"} Dec 11 02:05:02 crc kubenswrapper[4824]: I1211 02:05:02.400540 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"8172b1bee3a0536d5cac459b105431e4c4b6d370cf04ee9eef35997d5b0a5eea"} Dec 11 02:05:02 crc kubenswrapper[4824]: I1211 02:05:02.400588 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"67614914a77556180339d930a7b5be86d1254da0ff0a54717591a6b773cd63f8"} Dec 11 02:05:02 crc kubenswrapper[4824]: I1211 02:05:02.400829 4824 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="beab386b-6c65-4bc3-9ef0-dc3f68b4403f" Dec 11 02:05:02 crc kubenswrapper[4824]: I1211 02:05:02.400847 4824 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="beab386b-6c65-4bc3-9ef0-dc3f68b4403f" Dec 11 02:05:02 crc kubenswrapper[4824]: I1211 02:05:02.401097 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 02:05:04 crc kubenswrapper[4824]: I1211 02:05:04.652016 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 02:05:04 crc kubenswrapper[4824]: I1211 02:05:04.652850 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 02:05:04 crc kubenswrapper[4824]: I1211 02:05:04.659838 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 02:05:07 crc kubenswrapper[4824]: I1211 02:05:07.070796 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 02:05:07 crc kubenswrapper[4824]: I1211 02:05:07.231015 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 02:05:07 crc kubenswrapper[4824]: I1211 02:05:07.231659 4824 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Dec 11 02:05:07 crc kubenswrapper[4824]: I1211 02:05:07.231737 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Dec 11 02:05:07 crc kubenswrapper[4824]: I1211 02:05:07.418922 4824 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 02:05:08 crc kubenswrapper[4824]: I1211 02:05:08.440719 4824 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="beab386b-6c65-4bc3-9ef0-dc3f68b4403f" Dec 11 02:05:08 crc kubenswrapper[4824]: I1211 02:05:08.441182 4824 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="beab386b-6c65-4bc3-9ef0-dc3f68b4403f" Dec 11 02:05:08 crc kubenswrapper[4824]: I1211 02:05:08.451612 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 02:05:08 crc kubenswrapper[4824]: I1211 02:05:08.652902 4824 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="acd9695e-f0a8-430d-b0f0-e9d3a4cdb526" Dec 11 02:05:09 crc kubenswrapper[4824]: I1211 02:05:09.446846 4824 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="beab386b-6c65-4bc3-9ef0-dc3f68b4403f" Dec 11 02:05:09 crc kubenswrapper[4824]: I1211 02:05:09.446898 4824 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="beab386b-6c65-4bc3-9ef0-dc3f68b4403f" Dec 11 02:05:09 crc kubenswrapper[4824]: I1211 02:05:09.450998 4824 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="acd9695e-f0a8-430d-b0f0-e9d3a4cdb526" Dec 11 02:05:17 crc kubenswrapper[4824]: I1211 02:05:17.231039 4824 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Dec 11 02:05:17 crc kubenswrapper[4824]: I1211 02:05:17.231796 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Dec 11 02:05:17 crc kubenswrapper[4824]: I1211 02:05:17.486966 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 11 02:05:17 crc kubenswrapper[4824]: I1211 02:05:17.725030 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 11 02:05:18 crc kubenswrapper[4824]: I1211 02:05:18.188531 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 11 02:05:18 crc kubenswrapper[4824]: I1211 02:05:18.359049 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 11 02:05:18 crc kubenswrapper[4824]: I1211 02:05:18.543811 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 11 02:05:18 crc kubenswrapper[4824]: I1211 02:05:18.574581 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 11 02:05:18 crc kubenswrapper[4824]: I1211 02:05:18.704048 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 11 02:05:18 crc kubenswrapper[4824]: I1211 02:05:18.979652 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 11 02:05:19 crc kubenswrapper[4824]: I1211 02:05:19.144161 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 11 02:05:19 crc kubenswrapper[4824]: I1211 02:05:19.162481 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 11 02:05:19 crc kubenswrapper[4824]: I1211 02:05:19.174773 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 11 02:05:19 crc kubenswrapper[4824]: I1211 02:05:19.287634 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 11 02:05:19 crc kubenswrapper[4824]: I1211 02:05:19.377326 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 11 02:05:19 crc kubenswrapper[4824]: I1211 02:05:19.541790 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 11 02:05:19 crc kubenswrapper[4824]: I1211 02:05:19.641968 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 11 02:05:19 crc kubenswrapper[4824]: I1211 02:05:19.938276 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 11 02:05:20 crc kubenswrapper[4824]: I1211 02:05:20.324462 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 11 02:05:20 crc kubenswrapper[4824]: I1211 02:05:20.340863 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 11 02:05:20 crc kubenswrapper[4824]: I1211 02:05:20.437790 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 11 02:05:20 crc kubenswrapper[4824]: I1211 02:05:20.568107 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 11 02:05:20 crc kubenswrapper[4824]: I1211 02:05:20.636944 4824 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 11 02:05:21 crc kubenswrapper[4824]: I1211 02:05:21.052974 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 11 02:05:21 crc kubenswrapper[4824]: I1211 02:05:21.352473 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 11 02:05:21 crc kubenswrapper[4824]: I1211 02:05:21.472047 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 11 02:05:21 crc kubenswrapper[4824]: I1211 02:05:21.836228 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 11 02:05:21 crc kubenswrapper[4824]: I1211 02:05:21.855795 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 11 02:05:21 crc kubenswrapper[4824]: I1211 02:05:21.872179 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 11 02:05:21 crc kubenswrapper[4824]: I1211 02:05:21.875085 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 11 02:05:21 crc kubenswrapper[4824]: I1211 02:05:21.907411 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 11 02:05:22 crc kubenswrapper[4824]: I1211 02:05:21.999933 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 11 02:05:22 crc kubenswrapper[4824]: I1211 02:05:22.150706 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 11 02:05:22 crc kubenswrapper[4824]: I1211 02:05:22.218163 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 11 02:05:22 crc kubenswrapper[4824]: I1211 02:05:22.222620 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 11 02:05:22 crc kubenswrapper[4824]: I1211 02:05:22.271197 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 11 02:05:22 crc kubenswrapper[4824]: I1211 02:05:22.272161 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 11 02:05:22 crc kubenswrapper[4824]: I1211 02:05:22.336459 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 11 02:05:22 crc kubenswrapper[4824]: I1211 02:05:22.394969 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 11 02:05:22 crc kubenswrapper[4824]: I1211 02:05:22.404649 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 11 02:05:22 crc kubenswrapper[4824]: I1211 02:05:22.448505 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 11 02:05:22 crc kubenswrapper[4824]: I1211 02:05:22.546878 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 11 02:05:22 crc kubenswrapper[4824]: I1211 02:05:22.577470 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 11 02:05:22 crc kubenswrapper[4824]: I1211 02:05:22.608951 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 11 02:05:22 crc kubenswrapper[4824]: I1211 02:05:22.631306 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 11 02:05:22 crc kubenswrapper[4824]: I1211 02:05:22.851355 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 11 02:05:23 crc kubenswrapper[4824]: I1211 02:05:23.022487 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 11 02:05:23 crc kubenswrapper[4824]: I1211 02:05:23.131813 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 11 02:05:23 crc kubenswrapper[4824]: I1211 02:05:23.164845 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 11 02:05:23 crc kubenswrapper[4824]: I1211 02:05:23.211735 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 11 02:05:23 crc kubenswrapper[4824]: I1211 02:05:23.390178 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 11 02:05:23 crc kubenswrapper[4824]: I1211 02:05:23.449460 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 11 02:05:23 crc kubenswrapper[4824]: I1211 02:05:23.458617 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 11 02:05:23 crc kubenswrapper[4824]: I1211 02:05:23.546652 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 11 02:05:23 crc kubenswrapper[4824]: I1211 02:05:23.553965 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 11 02:05:23 crc kubenswrapper[4824]: I1211 02:05:23.580763 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 11 02:05:23 crc kubenswrapper[4824]: I1211 02:05:23.627656 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 11 02:05:23 crc kubenswrapper[4824]: I1211 02:05:23.673398 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 11 02:05:23 crc kubenswrapper[4824]: I1211 02:05:23.693517 4824 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 11 02:05:23 crc kubenswrapper[4824]: I1211 02:05:23.720554 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 11 02:05:23 crc kubenswrapper[4824]: I1211 02:05:23.730353 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 11 02:05:23 crc kubenswrapper[4824]: I1211 02:05:23.776213 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 11 02:05:23 crc kubenswrapper[4824]: I1211 02:05:23.827695 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 11 02:05:23 crc kubenswrapper[4824]: I1211 02:05:23.908019 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 11 02:05:23 crc kubenswrapper[4824]: I1211 02:05:23.985173 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 11 02:05:23 crc kubenswrapper[4824]: I1211 02:05:23.994497 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 11 02:05:24 crc kubenswrapper[4824]: I1211 02:05:24.031487 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 11 02:05:24 crc kubenswrapper[4824]: I1211 02:05:24.069934 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 11 02:05:24 crc kubenswrapper[4824]: I1211 02:05:24.239024 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 11 02:05:24 crc kubenswrapper[4824]: I1211 02:05:24.320687 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 11 02:05:24 crc kubenswrapper[4824]: I1211 02:05:24.372502 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 11 02:05:24 crc kubenswrapper[4824]: I1211 02:05:24.381040 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 11 02:05:24 crc kubenswrapper[4824]: I1211 02:05:24.403146 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 11 02:05:24 crc kubenswrapper[4824]: I1211 02:05:24.485262 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 11 02:05:24 crc kubenswrapper[4824]: I1211 02:05:24.507199 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 11 02:05:24 crc kubenswrapper[4824]: I1211 02:05:24.554824 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 11 02:05:24 crc kubenswrapper[4824]: I1211 02:05:24.626826 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 11 02:05:24 crc kubenswrapper[4824]: I1211 02:05:24.635246 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 11 02:05:24 crc kubenswrapper[4824]: I1211 02:05:24.673389 4824 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 11 02:05:24 crc kubenswrapper[4824]: I1211 02:05:24.679781 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-authentication/oauth-openshift-558db77b4-l9d4f"] Dec 11 02:05:24 crc kubenswrapper[4824]: I1211 02:05:24.679860 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 11 02:05:24 crc kubenswrapper[4824]: I1211 02:05:24.683387 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 02:05:24 crc kubenswrapper[4824]: I1211 02:05:24.695616 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=17.695596117 podStartE2EDuration="17.695596117s" podCreationTimestamp="2025-12-11 02:05:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:05:24.692921722 +0000 UTC m=+266.381959101" watchObservedRunningTime="2025-12-11 02:05:24.695596117 +0000 UTC m=+266.384633506" Dec 11 02:05:24 crc kubenswrapper[4824]: I1211 02:05:24.721024 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 11 02:05:25 crc kubenswrapper[4824]: I1211 02:05:25.020589 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 11 02:05:25 crc kubenswrapper[4824]: I1211 02:05:25.080305 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 11 02:05:25 crc kubenswrapper[4824]: I1211 02:05:25.139900 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 11 02:05:25 crc kubenswrapper[4824]: I1211 02:05:25.260457 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 11 02:05:25 crc kubenswrapper[4824]: I1211 02:05:25.298923 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 11 02:05:25 crc kubenswrapper[4824]: I1211 02:05:25.323566 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 11 02:05:25 crc kubenswrapper[4824]: I1211 02:05:25.341482 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 11 02:05:25 crc kubenswrapper[4824]: I1211 02:05:25.371692 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 11 02:05:25 crc kubenswrapper[4824]: I1211 02:05:25.404484 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 11 02:05:25 crc kubenswrapper[4824]: I1211 02:05:25.740208 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 11 02:05:25 crc kubenswrapper[4824]: I1211 02:05:25.821599 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 11 02:05:25 crc kubenswrapper[4824]: I1211 02:05:25.863611 4824 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 11 02:05:25 crc kubenswrapper[4824]: I1211 02:05:25.875207 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 11 02:05:25 crc kubenswrapper[4824]: I1211 02:05:25.912996 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 11 02:05:25 crc kubenswrapper[4824]: I1211 02:05:25.941003 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 11 02:05:25 crc kubenswrapper[4824]: I1211 02:05:25.986366 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 11 02:05:26 crc kubenswrapper[4824]: I1211 02:05:26.008055 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 11 02:05:26 crc kubenswrapper[4824]: I1211 02:05:26.011425 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 11 02:05:26 crc kubenswrapper[4824]: I1211 02:05:26.014337 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 11 02:05:26 crc kubenswrapper[4824]: I1211 02:05:26.089941 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 11 02:05:26 crc kubenswrapper[4824]: I1211 02:05:26.134154 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 11 02:05:26 crc kubenswrapper[4824]: I1211 02:05:26.140302 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 11 02:05:26 crc kubenswrapper[4824]: I1211 02:05:26.161364 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 11 02:05:26 crc kubenswrapper[4824]: I1211 02:05:26.200598 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 11 02:05:26 crc kubenswrapper[4824]: I1211 02:05:26.283776 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 11 02:05:26 crc kubenswrapper[4824]: I1211 02:05:26.298795 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 11 02:05:26 crc kubenswrapper[4824]: I1211 02:05:26.443815 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 11 02:05:26 crc kubenswrapper[4824]: I1211 02:05:26.513811 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 11 02:05:26 crc kubenswrapper[4824]: I1211 02:05:26.525909 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 11 02:05:26 crc kubenswrapper[4824]: I1211 02:05:26.644539 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bdf5eb2-fe05-4e24-9e00-43c72afc3583" path="/var/lib/kubelet/pods/1bdf5eb2-fe05-4e24-9e00-43c72afc3583/volumes" Dec 11 02:05:26 crc kubenswrapper[4824]: I1211 02:05:26.838891 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 11 02:05:26 crc kubenswrapper[4824]: I1211 02:05:26.847722 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 11 02:05:26 crc kubenswrapper[4824]: I1211 02:05:26.949979 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 11 02:05:26 crc kubenswrapper[4824]: I1211 02:05:26.995108 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 11 02:05:27 crc kubenswrapper[4824]: I1211 02:05:27.124772 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 11 02:05:27 crc kubenswrapper[4824]: I1211 02:05:27.139672 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 11 02:05:27 crc kubenswrapper[4824]: I1211 02:05:27.216085 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 11 02:05:27 crc kubenswrapper[4824]: I1211 02:05:27.230543 4824 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Dec 11 02:05:27 crc kubenswrapper[4824]: I1211 02:05:27.230605 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Dec 11 02:05:27 crc kubenswrapper[4824]: I1211 02:05:27.230667 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 02:05:27 crc kubenswrapper[4824]: I1211 02:05:27.231408 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 11 02:05:27 crc kubenswrapper[4824]: I1211 02:05:27.231526 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="kube-controller-manager" containerStatusID={"Type":"cri-o","ID":"905c456d05b56778448af338a340744c24962427d762e5ff4a2844213f67c402"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container kube-controller-manager failed startup probe, will be restarted" Dec 11 02:05:27 crc kubenswrapper[4824]: I1211 02:05:27.231719 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" containerID="cri-o://905c456d05b56778448af338a340744c24962427d762e5ff4a2844213f67c402" gracePeriod=30 Dec 11 02:05:27 crc kubenswrapper[4824]: I1211 02:05:27.264370 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 11 02:05:27 crc kubenswrapper[4824]: I1211 02:05:27.298863 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 11 02:05:27 crc kubenswrapper[4824]: I1211 02:05:27.298855 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 11 02:05:27 crc kubenswrapper[4824]: I1211 02:05:27.463785 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 11 02:05:27 crc kubenswrapper[4824]: I1211 02:05:27.514851 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 11 02:05:27 crc kubenswrapper[4824]: I1211 02:05:27.599638 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 11 02:05:27 crc kubenswrapper[4824]: I1211 02:05:27.608247 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 11 02:05:27 crc kubenswrapper[4824]: I1211 02:05:27.635745 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 11 02:05:27 crc kubenswrapper[4824]: I1211 02:05:27.642438 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 11 02:05:27 crc kubenswrapper[4824]: I1211 02:05:27.730860 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 11 02:05:27 crc kubenswrapper[4824]: I1211 02:05:27.874347 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 11 02:05:27 crc kubenswrapper[4824]: I1211 02:05:27.875296 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 11 02:05:27 crc kubenswrapper[4824]: I1211 02:05:27.932584 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 11 02:05:27 crc kubenswrapper[4824]: I1211 02:05:27.989344 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 11 02:05:27 crc kubenswrapper[4824]: I1211 02:05:27.995247 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 11 02:05:28 crc kubenswrapper[4824]: I1211 02:05:28.153271 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 11 02:05:28 crc kubenswrapper[4824]: I1211 02:05:28.177105 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 11 02:05:28 crc kubenswrapper[4824]: I1211 02:05:28.252431 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 11 02:05:28 crc kubenswrapper[4824]: I1211 02:05:28.264156 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 11 02:05:28 crc kubenswrapper[4824]: I1211 02:05:28.337605 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 11 02:05:28 crc kubenswrapper[4824]: I1211 02:05:28.351355 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 11 02:05:28 crc kubenswrapper[4824]: I1211 02:05:28.357639 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 11 02:05:28 crc kubenswrapper[4824]: I1211 02:05:28.372590 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 11 02:05:28 crc kubenswrapper[4824]: I1211 02:05:28.380959 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 11 02:05:28 crc kubenswrapper[4824]: I1211 02:05:28.381467 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 11 02:05:28 crc kubenswrapper[4824]: I1211 02:05:28.407094 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 11 02:05:28 crc kubenswrapper[4824]: I1211 02:05:28.467490 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 11 02:05:28 crc kubenswrapper[4824]: I1211 02:05:28.507947 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 11 02:05:28 crc kubenswrapper[4824]: I1211 02:05:28.660343 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 11 02:05:28 crc kubenswrapper[4824]: I1211 02:05:28.676981 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 11 02:05:28 crc kubenswrapper[4824]: I1211 02:05:28.700745 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 11 02:05:28 crc kubenswrapper[4824]: I1211 02:05:28.751685 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 11 02:05:28 crc kubenswrapper[4824]: I1211 02:05:28.804856 4824 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 11 02:05:28 crc kubenswrapper[4824]: I1211 02:05:28.875507 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 11 02:05:28 crc kubenswrapper[4824]: I1211 02:05:28.977882 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 11 02:05:28 crc kubenswrapper[4824]: I1211 02:05:28.987523 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 11 02:05:29 crc kubenswrapper[4824]: I1211 02:05:29.163430 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 11 02:05:29 crc kubenswrapper[4824]: I1211 02:05:29.226633 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 11 02:05:29 crc kubenswrapper[4824]: I1211 02:05:29.320149 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 11 02:05:29 crc kubenswrapper[4824]: I1211 02:05:29.336098 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 11 02:05:29 crc kubenswrapper[4824]: I1211 02:05:29.343818 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 11 02:05:29 crc kubenswrapper[4824]: I1211 02:05:29.356464 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 11 02:05:29 crc kubenswrapper[4824]: I1211 02:05:29.637767 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 11 02:05:29 crc kubenswrapper[4824]: I1211 02:05:29.639310 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 11 02:05:29 crc kubenswrapper[4824]: I1211 02:05:29.643059 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 11 02:05:29 crc kubenswrapper[4824]: I1211 02:05:29.707430 4824 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 11 02:05:29 crc kubenswrapper[4824]: I1211 02:05:29.723507 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 11 02:05:29 crc kubenswrapper[4824]: I1211 02:05:29.792407 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 11 02:05:29 crc kubenswrapper[4824]: I1211 02:05:29.817236 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 11 02:05:29 crc kubenswrapper[4824]: I1211 02:05:29.870853 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 11 02:05:29 crc kubenswrapper[4824]: I1211 02:05:29.898213 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 11 02:05:29 crc kubenswrapper[4824]: I1211 02:05:29.945449 4824 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 11 02:05:29 crc kubenswrapper[4824]: I1211 02:05:29.945728 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://05d0fb77af8d72804d94abf2e96c18aa970ee327ca6a5a6fc8fe725eeb741dba" gracePeriod=5 Dec 11 02:05:29 crc kubenswrapper[4824]: I1211 02:05:29.953075 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 11 02:05:30 crc kubenswrapper[4824]: I1211 02:05:30.035517 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 11 02:05:30 crc kubenswrapper[4824]: I1211 02:05:30.037997 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 11 02:05:30 crc kubenswrapper[4824]: I1211 02:05:30.077985 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 11 02:05:30 crc kubenswrapper[4824]: I1211 02:05:30.117865 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 11 02:05:30 crc kubenswrapper[4824]: I1211 02:05:30.137949 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 11 02:05:30 crc kubenswrapper[4824]: I1211 02:05:30.207759 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 11 02:05:30 crc kubenswrapper[4824]: I1211 02:05:30.221281 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 11 02:05:30 crc kubenswrapper[4824]: I1211 02:05:30.241453 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 11 02:05:30 crc kubenswrapper[4824]: I1211 02:05:30.293227 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 11 02:05:30 crc kubenswrapper[4824]: I1211 02:05:30.303723 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 11 02:05:30 crc kubenswrapper[4824]: I1211 02:05:30.352863 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 11 02:05:30 crc kubenswrapper[4824]: I1211 02:05:30.355847 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 11 02:05:30 crc kubenswrapper[4824]: I1211 02:05:30.356817 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 11 02:05:30 crc kubenswrapper[4824]: I1211 02:05:30.388823 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 11 02:05:30 crc kubenswrapper[4824]: I1211 02:05:30.434289 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 11 02:05:30 crc kubenswrapper[4824]: I1211 02:05:30.438685 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 11 02:05:30 crc kubenswrapper[4824]: I1211 02:05:30.468947 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 11 02:05:30 crc kubenswrapper[4824]: I1211 02:05:30.556234 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 11 02:05:30 crc kubenswrapper[4824]: I1211 02:05:30.602260 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 11 02:05:30 crc kubenswrapper[4824]: I1211 02:05:30.673764 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 11 02:05:30 crc kubenswrapper[4824]: I1211 02:05:30.675639 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 11 02:05:30 crc kubenswrapper[4824]: I1211 02:05:30.702884 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 11 02:05:30 crc kubenswrapper[4824]: I1211 02:05:30.718582 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 11 02:05:30 crc kubenswrapper[4824]: I1211 02:05:30.725431 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 11 02:05:30 crc kubenswrapper[4824]: I1211 02:05:30.727430 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 11 02:05:30 crc kubenswrapper[4824]: I1211 02:05:30.781138 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 11 02:05:30 crc kubenswrapper[4824]: I1211 02:05:30.807900 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 11 02:05:30 crc kubenswrapper[4824]: I1211 02:05:30.842099 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 11 02:05:30 crc kubenswrapper[4824]: I1211 02:05:30.845991 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 11 02:05:30 crc kubenswrapper[4824]: I1211 02:05:30.882366 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 11 02:05:31 crc kubenswrapper[4824]: I1211 02:05:31.257976 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 11 02:05:31 crc kubenswrapper[4824]: I1211 02:05:31.335010 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 11 02:05:31 crc kubenswrapper[4824]: I1211 02:05:31.364554 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 11 02:05:31 crc kubenswrapper[4824]: I1211 02:05:31.406776 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 11 02:05:31 crc kubenswrapper[4824]: I1211 02:05:31.475485 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 11 02:05:31 crc kubenswrapper[4824]: I1211 02:05:31.485960 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 11 02:05:31 crc kubenswrapper[4824]: I1211 02:05:31.508788 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 11 02:05:31 crc kubenswrapper[4824]: I1211 02:05:31.565520 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 11 02:05:31 crc kubenswrapper[4824]: I1211 02:05:31.705607 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 11 02:05:31 crc kubenswrapper[4824]: I1211 02:05:31.743655 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 11 02:05:31 crc kubenswrapper[4824]: I1211 02:05:31.879557 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 11 02:05:31 crc kubenswrapper[4824]: I1211 02:05:31.883464 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 11 02:05:32 crc kubenswrapper[4824]: I1211 02:05:32.036828 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 11 02:05:32 crc kubenswrapper[4824]: I1211 02:05:32.090516 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 11 02:05:32 crc kubenswrapper[4824]: I1211 02:05:32.236551 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 11 02:05:32 crc kubenswrapper[4824]: I1211 02:05:32.242581 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 11 02:05:32 crc kubenswrapper[4824]: I1211 02:05:32.322986 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 11 02:05:32 crc kubenswrapper[4824]: I1211 02:05:32.406012 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 11 02:05:32 crc kubenswrapper[4824]: I1211 02:05:32.508856 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 11 02:05:32 crc kubenswrapper[4824]: I1211 02:05:32.553630 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 11 02:05:32 crc kubenswrapper[4824]: I1211 02:05:32.560682 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 11 02:05:32 crc kubenswrapper[4824]: I1211 02:05:32.578482 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 11 02:05:32 crc kubenswrapper[4824]: I1211 02:05:32.620872 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 11 02:05:32 crc kubenswrapper[4824]: I1211 02:05:32.743347 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 11 02:05:32 crc kubenswrapper[4824]: I1211 02:05:32.817998 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 11 02:05:32 crc kubenswrapper[4824]: I1211 02:05:32.869545 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 11 02:05:32 crc kubenswrapper[4824]: I1211 02:05:32.911350 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 11 02:05:33 crc kubenswrapper[4824]: I1211 02:05:33.000628 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 11 02:05:33 crc kubenswrapper[4824]: I1211 02:05:33.033206 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 11 02:05:33 crc kubenswrapper[4824]: I1211 02:05:33.094393 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 11 02:05:33 crc kubenswrapper[4824]: I1211 02:05:33.318501 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 11 02:05:33 crc kubenswrapper[4824]: I1211 02:05:33.339769 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 11 02:05:33 crc kubenswrapper[4824]: I1211 02:05:33.464475 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 11 02:05:33 crc kubenswrapper[4824]: I1211 02:05:33.753902 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 11 02:05:34 crc kubenswrapper[4824]: I1211 02:05:34.022668 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 11 02:05:34 crc kubenswrapper[4824]: I1211 02:05:34.079826 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 11 02:05:34 crc kubenswrapper[4824]: I1211 02:05:34.437949 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 11 02:05:34 crc kubenswrapper[4824]: I1211 02:05:34.489461 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 11 02:05:34 crc kubenswrapper[4824]: I1211 02:05:34.833831 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 11 02:05:35 crc kubenswrapper[4824]: I1211 02:05:35.351025 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 11 02:05:35 crc kubenswrapper[4824]: I1211 02:05:35.616290 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 11 02:05:35 crc kubenswrapper[4824]: I1211 02:05:35.616395 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 02:05:35 crc kubenswrapper[4824]: I1211 02:05:35.646350 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 11 02:05:35 crc kubenswrapper[4824]: I1211 02:05:35.646421 4824 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="05d0fb77af8d72804d94abf2e96c18aa970ee327ca6a5a6fc8fe725eeb741dba" exitCode=137 Dec 11 02:05:35 crc kubenswrapper[4824]: I1211 02:05:35.646490 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 02:05:35 crc kubenswrapper[4824]: I1211 02:05:35.646512 4824 scope.go:117] "RemoveContainer" containerID="05d0fb77af8d72804d94abf2e96c18aa970ee327ca6a5a6fc8fe725eeb741dba" Dec 11 02:05:35 crc kubenswrapper[4824]: I1211 02:05:35.674001 4824 scope.go:117] "RemoveContainer" containerID="05d0fb77af8d72804d94abf2e96c18aa970ee327ca6a5a6fc8fe725eeb741dba" Dec 11 02:05:35 crc kubenswrapper[4824]: E1211 02:05:35.675003 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"05d0fb77af8d72804d94abf2e96c18aa970ee327ca6a5a6fc8fe725eeb741dba\": container with ID starting with 05d0fb77af8d72804d94abf2e96c18aa970ee327ca6a5a6fc8fe725eeb741dba not found: ID does not exist" containerID="05d0fb77af8d72804d94abf2e96c18aa970ee327ca6a5a6fc8fe725eeb741dba" Dec 11 02:05:35 crc kubenswrapper[4824]: I1211 02:05:35.675057 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"05d0fb77af8d72804d94abf2e96c18aa970ee327ca6a5a6fc8fe725eeb741dba"} err="failed to get container status \"05d0fb77af8d72804d94abf2e96c18aa970ee327ca6a5a6fc8fe725eeb741dba\": rpc error: code = NotFound desc = could not find container \"05d0fb77af8d72804d94abf2e96c18aa970ee327ca6a5a6fc8fe725eeb741dba\": container with ID starting with 05d0fb77af8d72804d94abf2e96c18aa970ee327ca6a5a6fc8fe725eeb741dba not found: ID does not exist" Dec 11 02:05:35 crc kubenswrapper[4824]: I1211 02:05:35.773344 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 11 02:05:35 crc kubenswrapper[4824]: I1211 02:05:35.773399 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 11 02:05:35 crc kubenswrapper[4824]: I1211 02:05:35.773463 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 11 02:05:35 crc kubenswrapper[4824]: I1211 02:05:35.773493 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 11 02:05:35 crc kubenswrapper[4824]: I1211 02:05:35.773542 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 11 02:05:35 crc kubenswrapper[4824]: I1211 02:05:35.773580 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 02:05:35 crc kubenswrapper[4824]: I1211 02:05:35.773611 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 02:05:35 crc kubenswrapper[4824]: I1211 02:05:35.773666 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 02:05:35 crc kubenswrapper[4824]: I1211 02:05:35.773678 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 02:05:35 crc kubenswrapper[4824]: I1211 02:05:35.773909 4824 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 11 02:05:35 crc kubenswrapper[4824]: I1211 02:05:35.773934 4824 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Dec 11 02:05:35 crc kubenswrapper[4824]: I1211 02:05:35.773953 4824 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Dec 11 02:05:35 crc kubenswrapper[4824]: I1211 02:05:35.773972 4824 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Dec 11 02:05:35 crc kubenswrapper[4824]: I1211 02:05:35.785368 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 02:05:35 crc kubenswrapper[4824]: I1211 02:05:35.874995 4824 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 11 02:05:36 crc kubenswrapper[4824]: I1211 02:05:36.644196 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Dec 11 02:05:39 crc kubenswrapper[4824]: I1211 02:05:39.068318 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-5697b9fbd5-xqk8j"] Dec 11 02:05:39 crc kubenswrapper[4824]: E1211 02:05:39.068983 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 11 02:05:39 crc kubenswrapper[4824]: I1211 02:05:39.068999 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 11 02:05:39 crc kubenswrapper[4824]: E1211 02:05:39.069050 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bdf5eb2-fe05-4e24-9e00-43c72afc3583" containerName="oauth-openshift" Dec 11 02:05:39 crc kubenswrapper[4824]: I1211 02:05:39.069056 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bdf5eb2-fe05-4e24-9e00-43c72afc3583" containerName="oauth-openshift" Dec 11 02:05:39 crc kubenswrapper[4824]: E1211 02:05:39.069069 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b27ed82-f069-42db-a3ae-907b1283d49e" containerName="installer" Dec 11 02:05:39 crc kubenswrapper[4824]: I1211 02:05:39.069074 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b27ed82-f069-42db-a3ae-907b1283d49e" containerName="installer" Dec 11 02:05:39 crc kubenswrapper[4824]: I1211 02:05:39.069224 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 11 02:05:39 crc kubenswrapper[4824]: I1211 02:05:39.069234 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bdf5eb2-fe05-4e24-9e00-43c72afc3583" containerName="oauth-openshift" Dec 11 02:05:39 crc kubenswrapper[4824]: I1211 02:05:39.069243 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b27ed82-f069-42db-a3ae-907b1283d49e" containerName="installer" Dec 11 02:05:39 crc kubenswrapper[4824]: I1211 02:05:39.069583 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-5697b9fbd5-xqk8j" Dec 11 02:05:39 crc kubenswrapper[4824]: I1211 02:05:39.076727 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 11 02:05:39 crc kubenswrapper[4824]: I1211 02:05:39.077002 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 11 02:05:39 crc kubenswrapper[4824]: I1211 02:05:39.077282 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 11 02:05:39 crc kubenswrapper[4824]: I1211 02:05:39.077389 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 11 02:05:39 crc kubenswrapper[4824]: I1211 02:05:39.077551 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 11 02:05:39 crc kubenswrapper[4824]: I1211 02:05:39.077688 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 11 02:05:39 crc kubenswrapper[4824]: I1211 02:05:39.077872 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 11 02:05:39 crc kubenswrapper[4824]: I1211 02:05:39.077905 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 11 02:05:39 crc kubenswrapper[4824]: I1211 02:05:39.077996 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 11 02:05:39 crc kubenswrapper[4824]: I1211 02:05:39.078236 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 11 02:05:39 crc kubenswrapper[4824]: I1211 02:05:39.078486 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 11 02:05:39 crc kubenswrapper[4824]: I1211 02:05:39.079107 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 11 02:05:39 crc kubenswrapper[4824]: I1211 02:05:39.085068 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 11 02:05:39 crc kubenswrapper[4824]: I1211 02:05:39.087463 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 11 02:05:39 crc kubenswrapper[4824]: I1211 02:05:39.088712 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-5697b9fbd5-xqk8j"] Dec 11 02:05:39 crc kubenswrapper[4824]: I1211 02:05:39.092760 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 11 02:05:39 crc kubenswrapper[4824]: I1211 02:05:39.215676 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bbb57596-8d31-40a2-996c-9cacd765512c-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5697b9fbd5-xqk8j\" (UID: \"bbb57596-8d31-40a2-996c-9cacd765512c\") " pod="openshift-authentication/oauth-openshift-5697b9fbd5-xqk8j" Dec 11 02:05:39 crc kubenswrapper[4824]: I1211 02:05:39.215748 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/bbb57596-8d31-40a2-996c-9cacd765512c-audit-policies\") pod \"oauth-openshift-5697b9fbd5-xqk8j\" (UID: \"bbb57596-8d31-40a2-996c-9cacd765512c\") " pod="openshift-authentication/oauth-openshift-5697b9fbd5-xqk8j" Dec 11 02:05:39 crc kubenswrapper[4824]: I1211 02:05:39.215776 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-slkcl\" (UniqueName: \"kubernetes.io/projected/bbb57596-8d31-40a2-996c-9cacd765512c-kube-api-access-slkcl\") pod \"oauth-openshift-5697b9fbd5-xqk8j\" (UID: \"bbb57596-8d31-40a2-996c-9cacd765512c\") " pod="openshift-authentication/oauth-openshift-5697b9fbd5-xqk8j" Dec 11 02:05:39 crc kubenswrapper[4824]: I1211 02:05:39.215800 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/bbb57596-8d31-40a2-996c-9cacd765512c-audit-dir\") pod \"oauth-openshift-5697b9fbd5-xqk8j\" (UID: \"bbb57596-8d31-40a2-996c-9cacd765512c\") " pod="openshift-authentication/oauth-openshift-5697b9fbd5-xqk8j" Dec 11 02:05:39 crc kubenswrapper[4824]: I1211 02:05:39.215821 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/bbb57596-8d31-40a2-996c-9cacd765512c-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5697b9fbd5-xqk8j\" (UID: \"bbb57596-8d31-40a2-996c-9cacd765512c\") " pod="openshift-authentication/oauth-openshift-5697b9fbd5-xqk8j" Dec 11 02:05:39 crc kubenswrapper[4824]: I1211 02:05:39.215846 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/bbb57596-8d31-40a2-996c-9cacd765512c-v4-0-config-system-router-certs\") pod \"oauth-openshift-5697b9fbd5-xqk8j\" (UID: \"bbb57596-8d31-40a2-996c-9cacd765512c\") " pod="openshift-authentication/oauth-openshift-5697b9fbd5-xqk8j" Dec 11 02:05:39 crc kubenswrapper[4824]: I1211 02:05:39.215870 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/bbb57596-8d31-40a2-996c-9cacd765512c-v4-0-config-user-template-error\") pod \"oauth-openshift-5697b9fbd5-xqk8j\" (UID: \"bbb57596-8d31-40a2-996c-9cacd765512c\") " pod="openshift-authentication/oauth-openshift-5697b9fbd5-xqk8j" Dec 11 02:05:39 crc kubenswrapper[4824]: I1211 02:05:39.215900 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/bbb57596-8d31-40a2-996c-9cacd765512c-v4-0-config-system-session\") pod \"oauth-openshift-5697b9fbd5-xqk8j\" (UID: \"bbb57596-8d31-40a2-996c-9cacd765512c\") " pod="openshift-authentication/oauth-openshift-5697b9fbd5-xqk8j" Dec 11 02:05:39 crc kubenswrapper[4824]: I1211 02:05:39.215930 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/bbb57596-8d31-40a2-996c-9cacd765512c-v4-0-config-user-template-login\") pod \"oauth-openshift-5697b9fbd5-xqk8j\" (UID: \"bbb57596-8d31-40a2-996c-9cacd765512c\") " pod="openshift-authentication/oauth-openshift-5697b9fbd5-xqk8j" Dec 11 02:05:39 crc kubenswrapper[4824]: I1211 02:05:39.215962 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/bbb57596-8d31-40a2-996c-9cacd765512c-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5697b9fbd5-xqk8j\" (UID: \"bbb57596-8d31-40a2-996c-9cacd765512c\") " pod="openshift-authentication/oauth-openshift-5697b9fbd5-xqk8j" Dec 11 02:05:39 crc kubenswrapper[4824]: I1211 02:05:39.215997 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/bbb57596-8d31-40a2-996c-9cacd765512c-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5697b9fbd5-xqk8j\" (UID: \"bbb57596-8d31-40a2-996c-9cacd765512c\") " pod="openshift-authentication/oauth-openshift-5697b9fbd5-xqk8j" Dec 11 02:05:39 crc kubenswrapper[4824]: I1211 02:05:39.216036 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/bbb57596-8d31-40a2-996c-9cacd765512c-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5697b9fbd5-xqk8j\" (UID: \"bbb57596-8d31-40a2-996c-9cacd765512c\") " pod="openshift-authentication/oauth-openshift-5697b9fbd5-xqk8j" Dec 11 02:05:39 crc kubenswrapper[4824]: I1211 02:05:39.216058 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/bbb57596-8d31-40a2-996c-9cacd765512c-v4-0-config-system-service-ca\") pod \"oauth-openshift-5697b9fbd5-xqk8j\" (UID: \"bbb57596-8d31-40a2-996c-9cacd765512c\") " pod="openshift-authentication/oauth-openshift-5697b9fbd5-xqk8j" Dec 11 02:05:39 crc kubenswrapper[4824]: I1211 02:05:39.216083 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/bbb57596-8d31-40a2-996c-9cacd765512c-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5697b9fbd5-xqk8j\" (UID: \"bbb57596-8d31-40a2-996c-9cacd765512c\") " pod="openshift-authentication/oauth-openshift-5697b9fbd5-xqk8j" Dec 11 02:05:39 crc kubenswrapper[4824]: I1211 02:05:39.317104 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/bbb57596-8d31-40a2-996c-9cacd765512c-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5697b9fbd5-xqk8j\" (UID: \"bbb57596-8d31-40a2-996c-9cacd765512c\") " pod="openshift-authentication/oauth-openshift-5697b9fbd5-xqk8j" Dec 11 02:05:39 crc kubenswrapper[4824]: I1211 02:05:39.317269 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/bbb57596-8d31-40a2-996c-9cacd765512c-v4-0-config-system-service-ca\") pod \"oauth-openshift-5697b9fbd5-xqk8j\" (UID: \"bbb57596-8d31-40a2-996c-9cacd765512c\") " pod="openshift-authentication/oauth-openshift-5697b9fbd5-xqk8j" Dec 11 02:05:39 crc kubenswrapper[4824]: I1211 02:05:39.317310 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/bbb57596-8d31-40a2-996c-9cacd765512c-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5697b9fbd5-xqk8j\" (UID: \"bbb57596-8d31-40a2-996c-9cacd765512c\") " pod="openshift-authentication/oauth-openshift-5697b9fbd5-xqk8j" Dec 11 02:05:39 crc kubenswrapper[4824]: I1211 02:05:39.317346 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/bbb57596-8d31-40a2-996c-9cacd765512c-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5697b9fbd5-xqk8j\" (UID: \"bbb57596-8d31-40a2-996c-9cacd765512c\") " pod="openshift-authentication/oauth-openshift-5697b9fbd5-xqk8j" Dec 11 02:05:39 crc kubenswrapper[4824]: I1211 02:05:39.317389 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bbb57596-8d31-40a2-996c-9cacd765512c-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5697b9fbd5-xqk8j\" (UID: \"bbb57596-8d31-40a2-996c-9cacd765512c\") " pod="openshift-authentication/oauth-openshift-5697b9fbd5-xqk8j" Dec 11 02:05:39 crc kubenswrapper[4824]: I1211 02:05:39.317438 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/bbb57596-8d31-40a2-996c-9cacd765512c-audit-policies\") pod \"oauth-openshift-5697b9fbd5-xqk8j\" (UID: \"bbb57596-8d31-40a2-996c-9cacd765512c\") " pod="openshift-authentication/oauth-openshift-5697b9fbd5-xqk8j" Dec 11 02:05:39 crc kubenswrapper[4824]: I1211 02:05:39.317485 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-slkcl\" (UniqueName: \"kubernetes.io/projected/bbb57596-8d31-40a2-996c-9cacd765512c-kube-api-access-slkcl\") pod \"oauth-openshift-5697b9fbd5-xqk8j\" (UID: \"bbb57596-8d31-40a2-996c-9cacd765512c\") " pod="openshift-authentication/oauth-openshift-5697b9fbd5-xqk8j" Dec 11 02:05:39 crc kubenswrapper[4824]: I1211 02:05:39.317539 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/bbb57596-8d31-40a2-996c-9cacd765512c-audit-dir\") pod \"oauth-openshift-5697b9fbd5-xqk8j\" (UID: \"bbb57596-8d31-40a2-996c-9cacd765512c\") " pod="openshift-authentication/oauth-openshift-5697b9fbd5-xqk8j" Dec 11 02:05:39 crc kubenswrapper[4824]: I1211 02:05:39.317588 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/bbb57596-8d31-40a2-996c-9cacd765512c-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5697b9fbd5-xqk8j\" (UID: \"bbb57596-8d31-40a2-996c-9cacd765512c\") " pod="openshift-authentication/oauth-openshift-5697b9fbd5-xqk8j" Dec 11 02:05:39 crc kubenswrapper[4824]: I1211 02:05:39.317646 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/bbb57596-8d31-40a2-996c-9cacd765512c-v4-0-config-system-router-certs\") pod \"oauth-openshift-5697b9fbd5-xqk8j\" (UID: \"bbb57596-8d31-40a2-996c-9cacd765512c\") " pod="openshift-authentication/oauth-openshift-5697b9fbd5-xqk8j" Dec 11 02:05:39 crc kubenswrapper[4824]: I1211 02:05:39.317698 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/bbb57596-8d31-40a2-996c-9cacd765512c-v4-0-config-user-template-error\") pod \"oauth-openshift-5697b9fbd5-xqk8j\" (UID: \"bbb57596-8d31-40a2-996c-9cacd765512c\") " pod="openshift-authentication/oauth-openshift-5697b9fbd5-xqk8j" Dec 11 02:05:39 crc kubenswrapper[4824]: I1211 02:05:39.317765 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/bbb57596-8d31-40a2-996c-9cacd765512c-v4-0-config-system-session\") pod \"oauth-openshift-5697b9fbd5-xqk8j\" (UID: \"bbb57596-8d31-40a2-996c-9cacd765512c\") " pod="openshift-authentication/oauth-openshift-5697b9fbd5-xqk8j" Dec 11 02:05:39 crc kubenswrapper[4824]: I1211 02:05:39.317815 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/bbb57596-8d31-40a2-996c-9cacd765512c-v4-0-config-user-template-login\") pod \"oauth-openshift-5697b9fbd5-xqk8j\" (UID: \"bbb57596-8d31-40a2-996c-9cacd765512c\") " pod="openshift-authentication/oauth-openshift-5697b9fbd5-xqk8j" Dec 11 02:05:39 crc kubenswrapper[4824]: I1211 02:05:39.317882 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/bbb57596-8d31-40a2-996c-9cacd765512c-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5697b9fbd5-xqk8j\" (UID: \"bbb57596-8d31-40a2-996c-9cacd765512c\") " pod="openshift-authentication/oauth-openshift-5697b9fbd5-xqk8j" Dec 11 02:05:39 crc kubenswrapper[4824]: I1211 02:05:39.317947 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/bbb57596-8d31-40a2-996c-9cacd765512c-audit-dir\") pod \"oauth-openshift-5697b9fbd5-xqk8j\" (UID: \"bbb57596-8d31-40a2-996c-9cacd765512c\") " pod="openshift-authentication/oauth-openshift-5697b9fbd5-xqk8j" Dec 11 02:05:39 crc kubenswrapper[4824]: I1211 02:05:39.319017 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/bbb57596-8d31-40a2-996c-9cacd765512c-audit-policies\") pod \"oauth-openshift-5697b9fbd5-xqk8j\" (UID: \"bbb57596-8d31-40a2-996c-9cacd765512c\") " pod="openshift-authentication/oauth-openshift-5697b9fbd5-xqk8j" Dec 11 02:05:39 crc kubenswrapper[4824]: I1211 02:05:39.319099 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/bbb57596-8d31-40a2-996c-9cacd765512c-v4-0-config-system-service-ca\") pod \"oauth-openshift-5697b9fbd5-xqk8j\" (UID: \"bbb57596-8d31-40a2-996c-9cacd765512c\") " pod="openshift-authentication/oauth-openshift-5697b9fbd5-xqk8j" Dec 11 02:05:39 crc kubenswrapper[4824]: I1211 02:05:39.319805 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bbb57596-8d31-40a2-996c-9cacd765512c-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5697b9fbd5-xqk8j\" (UID: \"bbb57596-8d31-40a2-996c-9cacd765512c\") " pod="openshift-authentication/oauth-openshift-5697b9fbd5-xqk8j" Dec 11 02:05:39 crc kubenswrapper[4824]: I1211 02:05:39.321043 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/bbb57596-8d31-40a2-996c-9cacd765512c-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5697b9fbd5-xqk8j\" (UID: \"bbb57596-8d31-40a2-996c-9cacd765512c\") " pod="openshift-authentication/oauth-openshift-5697b9fbd5-xqk8j" Dec 11 02:05:39 crc kubenswrapper[4824]: I1211 02:05:39.326865 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/bbb57596-8d31-40a2-996c-9cacd765512c-v4-0-config-system-router-certs\") pod \"oauth-openshift-5697b9fbd5-xqk8j\" (UID: \"bbb57596-8d31-40a2-996c-9cacd765512c\") " pod="openshift-authentication/oauth-openshift-5697b9fbd5-xqk8j" Dec 11 02:05:39 crc kubenswrapper[4824]: I1211 02:05:39.327642 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/bbb57596-8d31-40a2-996c-9cacd765512c-v4-0-config-user-template-login\") pod \"oauth-openshift-5697b9fbd5-xqk8j\" (UID: \"bbb57596-8d31-40a2-996c-9cacd765512c\") " pod="openshift-authentication/oauth-openshift-5697b9fbd5-xqk8j" Dec 11 02:05:39 crc kubenswrapper[4824]: I1211 02:05:39.327768 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/bbb57596-8d31-40a2-996c-9cacd765512c-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5697b9fbd5-xqk8j\" (UID: \"bbb57596-8d31-40a2-996c-9cacd765512c\") " pod="openshift-authentication/oauth-openshift-5697b9fbd5-xqk8j" Dec 11 02:05:39 crc kubenswrapper[4824]: I1211 02:05:39.328163 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/bbb57596-8d31-40a2-996c-9cacd765512c-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5697b9fbd5-xqk8j\" (UID: \"bbb57596-8d31-40a2-996c-9cacd765512c\") " pod="openshift-authentication/oauth-openshift-5697b9fbd5-xqk8j" Dec 11 02:05:39 crc kubenswrapper[4824]: I1211 02:05:39.328637 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/bbb57596-8d31-40a2-996c-9cacd765512c-v4-0-config-system-session\") pod \"oauth-openshift-5697b9fbd5-xqk8j\" (UID: \"bbb57596-8d31-40a2-996c-9cacd765512c\") " pod="openshift-authentication/oauth-openshift-5697b9fbd5-xqk8j" Dec 11 02:05:39 crc kubenswrapper[4824]: I1211 02:05:39.329547 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/bbb57596-8d31-40a2-996c-9cacd765512c-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5697b9fbd5-xqk8j\" (UID: \"bbb57596-8d31-40a2-996c-9cacd765512c\") " pod="openshift-authentication/oauth-openshift-5697b9fbd5-xqk8j" Dec 11 02:05:39 crc kubenswrapper[4824]: I1211 02:05:39.329620 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/bbb57596-8d31-40a2-996c-9cacd765512c-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5697b9fbd5-xqk8j\" (UID: \"bbb57596-8d31-40a2-996c-9cacd765512c\") " pod="openshift-authentication/oauth-openshift-5697b9fbd5-xqk8j" Dec 11 02:05:39 crc kubenswrapper[4824]: I1211 02:05:39.330229 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/bbb57596-8d31-40a2-996c-9cacd765512c-v4-0-config-user-template-error\") pod \"oauth-openshift-5697b9fbd5-xqk8j\" (UID: \"bbb57596-8d31-40a2-996c-9cacd765512c\") " pod="openshift-authentication/oauth-openshift-5697b9fbd5-xqk8j" Dec 11 02:05:39 crc kubenswrapper[4824]: I1211 02:05:39.354962 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-slkcl\" (UniqueName: \"kubernetes.io/projected/bbb57596-8d31-40a2-996c-9cacd765512c-kube-api-access-slkcl\") pod \"oauth-openshift-5697b9fbd5-xqk8j\" (UID: \"bbb57596-8d31-40a2-996c-9cacd765512c\") " pod="openshift-authentication/oauth-openshift-5697b9fbd5-xqk8j" Dec 11 02:05:39 crc kubenswrapper[4824]: I1211 02:05:39.394650 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-5697b9fbd5-xqk8j" Dec 11 02:05:39 crc kubenswrapper[4824]: I1211 02:05:39.660128 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-5697b9fbd5-xqk8j"] Dec 11 02:05:40 crc kubenswrapper[4824]: I1211 02:05:40.681726 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-5697b9fbd5-xqk8j" event={"ID":"bbb57596-8d31-40a2-996c-9cacd765512c","Type":"ContainerStarted","Data":"7973564afa555ce281cd5979139fe8c187388730e7603878afcab0e93a50058e"} Dec 11 02:05:40 crc kubenswrapper[4824]: I1211 02:05:40.685979 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-5697b9fbd5-xqk8j" event={"ID":"bbb57596-8d31-40a2-996c-9cacd765512c","Type":"ContainerStarted","Data":"c24d8bd833197937a41f5c7582e0df0235ed8db2e4edf4f36c633a9b67451ae6"} Dec 11 02:05:40 crc kubenswrapper[4824]: I1211 02:05:40.686281 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-5697b9fbd5-xqk8j" Dec 11 02:05:40 crc kubenswrapper[4824]: I1211 02:05:40.690637 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-5697b9fbd5-xqk8j" Dec 11 02:05:40 crc kubenswrapper[4824]: I1211 02:05:40.725037 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-5697b9fbd5-xqk8j" podStartSLOduration=76.725015537 podStartE2EDuration="1m16.725015537s" podCreationTimestamp="2025-12-11 02:04:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:05:40.718588636 +0000 UTC m=+282.407626055" watchObservedRunningTime="2025-12-11 02:05:40.725015537 +0000 UTC m=+282.414052946" Dec 11 02:05:57 crc kubenswrapper[4824]: I1211 02:05:57.813652 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Dec 11 02:05:57 crc kubenswrapper[4824]: I1211 02:05:57.816636 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 11 02:05:57 crc kubenswrapper[4824]: I1211 02:05:57.816726 4824 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="905c456d05b56778448af338a340744c24962427d762e5ff4a2844213f67c402" exitCode=137 Dec 11 02:05:57 crc kubenswrapper[4824]: I1211 02:05:57.816771 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"905c456d05b56778448af338a340744c24962427d762e5ff4a2844213f67c402"} Dec 11 02:05:57 crc kubenswrapper[4824]: I1211 02:05:57.816815 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"8f737f82b582ebaf75576a8f1c910c8ebe99fdbdb56b49873860da110ae20d5f"} Dec 11 02:05:57 crc kubenswrapper[4824]: I1211 02:05:57.816843 4824 scope.go:117] "RemoveContainer" containerID="922af7f67b99a0ce08776f3f75c0c377e6ea69af1720b6178a6d1e2f9b4b8ba3" Dec 11 02:05:58 crc kubenswrapper[4824]: I1211 02:05:58.478581 4824 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials Dec 11 02:05:58 crc kubenswrapper[4824]: I1211 02:05:58.826411 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Dec 11 02:06:07 crc kubenswrapper[4824]: I1211 02:06:07.071639 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 02:06:07 crc kubenswrapper[4824]: I1211 02:06:07.230101 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 02:06:07 crc kubenswrapper[4824]: I1211 02:06:07.237267 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 02:06:07 crc kubenswrapper[4824]: I1211 02:06:07.901147 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 02:06:16 crc kubenswrapper[4824]: I1211 02:06:16.876196 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-msljv"] Dec 11 02:06:16 crc kubenswrapper[4824]: I1211 02:06:16.877105 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-msljv" podUID="5407364a-e2b6-43b4-9eb0-deb12a022289" containerName="route-controller-manager" containerID="cri-o://dc7036b2d9bdde3b494cb3e26180f2b717abfa0d317411a55d23e90f4b10e57d" gracePeriod=30 Dec 11 02:06:16 crc kubenswrapper[4824]: I1211 02:06:16.983796 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-qr5xz"] Dec 11 02:06:16 crc kubenswrapper[4824]: I1211 02:06:16.983998 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-qr5xz" podUID="9e099899-aaf7-4bd8-b37e-8616b5060e33" containerName="controller-manager" containerID="cri-o://5a3bbef81765195bec5e85dc7ab9b24eec8b0ff2accaa3e59b2312b6d539413d" gracePeriod=30 Dec 11 02:06:17 crc kubenswrapper[4824]: I1211 02:06:17.318147 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-qr5xz" Dec 11 02:06:17 crc kubenswrapper[4824]: I1211 02:06:17.448024 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e099899-aaf7-4bd8-b37e-8616b5060e33-config\") pod \"9e099899-aaf7-4bd8-b37e-8616b5060e33\" (UID: \"9e099899-aaf7-4bd8-b37e-8616b5060e33\") " Dec 11 02:06:17 crc kubenswrapper[4824]: I1211 02:06:17.448165 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9e099899-aaf7-4bd8-b37e-8616b5060e33-serving-cert\") pod \"9e099899-aaf7-4bd8-b37e-8616b5060e33\" (UID: \"9e099899-aaf7-4bd8-b37e-8616b5060e33\") " Dec 11 02:06:17 crc kubenswrapper[4824]: I1211 02:06:17.448207 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6r6md\" (UniqueName: \"kubernetes.io/projected/9e099899-aaf7-4bd8-b37e-8616b5060e33-kube-api-access-6r6md\") pod \"9e099899-aaf7-4bd8-b37e-8616b5060e33\" (UID: \"9e099899-aaf7-4bd8-b37e-8616b5060e33\") " Dec 11 02:06:17 crc kubenswrapper[4824]: I1211 02:06:17.448256 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9e099899-aaf7-4bd8-b37e-8616b5060e33-client-ca\") pod \"9e099899-aaf7-4bd8-b37e-8616b5060e33\" (UID: \"9e099899-aaf7-4bd8-b37e-8616b5060e33\") " Dec 11 02:06:17 crc kubenswrapper[4824]: I1211 02:06:17.448294 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9e099899-aaf7-4bd8-b37e-8616b5060e33-proxy-ca-bundles\") pod \"9e099899-aaf7-4bd8-b37e-8616b5060e33\" (UID: \"9e099899-aaf7-4bd8-b37e-8616b5060e33\") " Dec 11 02:06:17 crc kubenswrapper[4824]: I1211 02:06:17.449469 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e099899-aaf7-4bd8-b37e-8616b5060e33-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "9e099899-aaf7-4bd8-b37e-8616b5060e33" (UID: "9e099899-aaf7-4bd8-b37e-8616b5060e33"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:06:17 crc kubenswrapper[4824]: I1211 02:06:17.449828 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e099899-aaf7-4bd8-b37e-8616b5060e33-client-ca" (OuterVolumeSpecName: "client-ca") pod "9e099899-aaf7-4bd8-b37e-8616b5060e33" (UID: "9e099899-aaf7-4bd8-b37e-8616b5060e33"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:06:17 crc kubenswrapper[4824]: I1211 02:06:17.450406 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e099899-aaf7-4bd8-b37e-8616b5060e33-config" (OuterVolumeSpecName: "config") pod "9e099899-aaf7-4bd8-b37e-8616b5060e33" (UID: "9e099899-aaf7-4bd8-b37e-8616b5060e33"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:06:17 crc kubenswrapper[4824]: I1211 02:06:17.459048 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e099899-aaf7-4bd8-b37e-8616b5060e33-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9e099899-aaf7-4bd8-b37e-8616b5060e33" (UID: "9e099899-aaf7-4bd8-b37e-8616b5060e33"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:06:17 crc kubenswrapper[4824]: I1211 02:06:17.459660 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e099899-aaf7-4bd8-b37e-8616b5060e33-kube-api-access-6r6md" (OuterVolumeSpecName: "kube-api-access-6r6md") pod "9e099899-aaf7-4bd8-b37e-8616b5060e33" (UID: "9e099899-aaf7-4bd8-b37e-8616b5060e33"). InnerVolumeSpecName "kube-api-access-6r6md". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:06:17 crc kubenswrapper[4824]: I1211 02:06:17.550280 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e099899-aaf7-4bd8-b37e-8616b5060e33-config\") on node \"crc\" DevicePath \"\"" Dec 11 02:06:17 crc kubenswrapper[4824]: I1211 02:06:17.550326 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9e099899-aaf7-4bd8-b37e-8616b5060e33-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 02:06:17 crc kubenswrapper[4824]: I1211 02:06:17.550348 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6r6md\" (UniqueName: \"kubernetes.io/projected/9e099899-aaf7-4bd8-b37e-8616b5060e33-kube-api-access-6r6md\") on node \"crc\" DevicePath \"\"" Dec 11 02:06:17 crc kubenswrapper[4824]: I1211 02:06:17.550367 4824 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9e099899-aaf7-4bd8-b37e-8616b5060e33-client-ca\") on node \"crc\" DevicePath \"\"" Dec 11 02:06:17 crc kubenswrapper[4824]: I1211 02:06:17.550384 4824 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9e099899-aaf7-4bd8-b37e-8616b5060e33-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 11 02:06:17 crc kubenswrapper[4824]: I1211 02:06:17.854898 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-msljv" Dec 11 02:06:17 crc kubenswrapper[4824]: I1211 02:06:17.952322 4824 generic.go:334] "Generic (PLEG): container finished" podID="5407364a-e2b6-43b4-9eb0-deb12a022289" containerID="dc7036b2d9bdde3b494cb3e26180f2b717abfa0d317411a55d23e90f4b10e57d" exitCode=0 Dec 11 02:06:17 crc kubenswrapper[4824]: I1211 02:06:17.952377 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-msljv" Dec 11 02:06:17 crc kubenswrapper[4824]: I1211 02:06:17.952377 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-msljv" event={"ID":"5407364a-e2b6-43b4-9eb0-deb12a022289","Type":"ContainerDied","Data":"dc7036b2d9bdde3b494cb3e26180f2b717abfa0d317411a55d23e90f4b10e57d"} Dec 11 02:06:17 crc kubenswrapper[4824]: I1211 02:06:17.952430 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-msljv" event={"ID":"5407364a-e2b6-43b4-9eb0-deb12a022289","Type":"ContainerDied","Data":"1311461ca6b530fd2d5299f8f6c5a8f9f353e64c24afe69528bcac3772261b60"} Dec 11 02:06:17 crc kubenswrapper[4824]: I1211 02:06:17.952454 4824 scope.go:117] "RemoveContainer" containerID="dc7036b2d9bdde3b494cb3e26180f2b717abfa0d317411a55d23e90f4b10e57d" Dec 11 02:06:17 crc kubenswrapper[4824]: I1211 02:06:17.955246 4824 generic.go:334] "Generic (PLEG): container finished" podID="9e099899-aaf7-4bd8-b37e-8616b5060e33" containerID="5a3bbef81765195bec5e85dc7ab9b24eec8b0ff2accaa3e59b2312b6d539413d" exitCode=0 Dec 11 02:06:17 crc kubenswrapper[4824]: I1211 02:06:17.955280 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-qr5xz" event={"ID":"9e099899-aaf7-4bd8-b37e-8616b5060e33","Type":"ContainerDied","Data":"5a3bbef81765195bec5e85dc7ab9b24eec8b0ff2accaa3e59b2312b6d539413d"} Dec 11 02:06:17 crc kubenswrapper[4824]: I1211 02:06:17.955337 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-qr5xz" Dec 11 02:06:17 crc kubenswrapper[4824]: I1211 02:06:17.955553 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-qr5xz" event={"ID":"9e099899-aaf7-4bd8-b37e-8616b5060e33","Type":"ContainerDied","Data":"c3b9fdc4b65c7a93573ec7c1b416fa6a98dd8adb41cd979ae09ab5dded0ce475"} Dec 11 02:06:17 crc kubenswrapper[4824]: I1211 02:06:17.965604 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5407364a-e2b6-43b4-9eb0-deb12a022289-config\") pod \"5407364a-e2b6-43b4-9eb0-deb12a022289\" (UID: \"5407364a-e2b6-43b4-9eb0-deb12a022289\") " Dec 11 02:06:17 crc kubenswrapper[4824]: I1211 02:06:17.965650 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5407364a-e2b6-43b4-9eb0-deb12a022289-client-ca\") pod \"5407364a-e2b6-43b4-9eb0-deb12a022289\" (UID: \"5407364a-e2b6-43b4-9eb0-deb12a022289\") " Dec 11 02:06:17 crc kubenswrapper[4824]: I1211 02:06:17.965710 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4x4bd\" (UniqueName: \"kubernetes.io/projected/5407364a-e2b6-43b4-9eb0-deb12a022289-kube-api-access-4x4bd\") pod \"5407364a-e2b6-43b4-9eb0-deb12a022289\" (UID: \"5407364a-e2b6-43b4-9eb0-deb12a022289\") " Dec 11 02:06:17 crc kubenswrapper[4824]: I1211 02:06:17.965768 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5407364a-e2b6-43b4-9eb0-deb12a022289-serving-cert\") pod \"5407364a-e2b6-43b4-9eb0-deb12a022289\" (UID: \"5407364a-e2b6-43b4-9eb0-deb12a022289\") " Dec 11 02:06:17 crc kubenswrapper[4824]: I1211 02:06:17.966387 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5407364a-e2b6-43b4-9eb0-deb12a022289-client-ca" (OuterVolumeSpecName: "client-ca") pod "5407364a-e2b6-43b4-9eb0-deb12a022289" (UID: "5407364a-e2b6-43b4-9eb0-deb12a022289"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:06:17 crc kubenswrapper[4824]: I1211 02:06:17.967392 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5407364a-e2b6-43b4-9eb0-deb12a022289-config" (OuterVolumeSpecName: "config") pod "5407364a-e2b6-43b4-9eb0-deb12a022289" (UID: "5407364a-e2b6-43b4-9eb0-deb12a022289"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:06:17 crc kubenswrapper[4824]: I1211 02:06:17.969577 4824 scope.go:117] "RemoveContainer" containerID="dc7036b2d9bdde3b494cb3e26180f2b717abfa0d317411a55d23e90f4b10e57d" Dec 11 02:06:17 crc kubenswrapper[4824]: E1211 02:06:17.970238 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc7036b2d9bdde3b494cb3e26180f2b717abfa0d317411a55d23e90f4b10e57d\": container with ID starting with dc7036b2d9bdde3b494cb3e26180f2b717abfa0d317411a55d23e90f4b10e57d not found: ID does not exist" containerID="dc7036b2d9bdde3b494cb3e26180f2b717abfa0d317411a55d23e90f4b10e57d" Dec 11 02:06:17 crc kubenswrapper[4824]: I1211 02:06:17.970273 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc7036b2d9bdde3b494cb3e26180f2b717abfa0d317411a55d23e90f4b10e57d"} err="failed to get container status \"dc7036b2d9bdde3b494cb3e26180f2b717abfa0d317411a55d23e90f4b10e57d\": rpc error: code = NotFound desc = could not find container \"dc7036b2d9bdde3b494cb3e26180f2b717abfa0d317411a55d23e90f4b10e57d\": container with ID starting with dc7036b2d9bdde3b494cb3e26180f2b717abfa0d317411a55d23e90f4b10e57d not found: ID does not exist" Dec 11 02:06:17 crc kubenswrapper[4824]: I1211 02:06:17.970294 4824 scope.go:117] "RemoveContainer" containerID="5a3bbef81765195bec5e85dc7ab9b24eec8b0ff2accaa3e59b2312b6d539413d" Dec 11 02:06:17 crc kubenswrapper[4824]: I1211 02:06:17.970563 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5407364a-e2b6-43b4-9eb0-deb12a022289-kube-api-access-4x4bd" (OuterVolumeSpecName: "kube-api-access-4x4bd") pod "5407364a-e2b6-43b4-9eb0-deb12a022289" (UID: "5407364a-e2b6-43b4-9eb0-deb12a022289"). InnerVolumeSpecName "kube-api-access-4x4bd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:06:17 crc kubenswrapper[4824]: I1211 02:06:17.972539 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5407364a-e2b6-43b4-9eb0-deb12a022289-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5407364a-e2b6-43b4-9eb0-deb12a022289" (UID: "5407364a-e2b6-43b4-9eb0-deb12a022289"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:06:17 crc kubenswrapper[4824]: I1211 02:06:17.985097 4824 scope.go:117] "RemoveContainer" containerID="5a3bbef81765195bec5e85dc7ab9b24eec8b0ff2accaa3e59b2312b6d539413d" Dec 11 02:06:17 crc kubenswrapper[4824]: E1211 02:06:17.988876 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5a3bbef81765195bec5e85dc7ab9b24eec8b0ff2accaa3e59b2312b6d539413d\": container with ID starting with 5a3bbef81765195bec5e85dc7ab9b24eec8b0ff2accaa3e59b2312b6d539413d not found: ID does not exist" containerID="5a3bbef81765195bec5e85dc7ab9b24eec8b0ff2accaa3e59b2312b6d539413d" Dec 11 02:06:17 crc kubenswrapper[4824]: I1211 02:06:17.988977 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a3bbef81765195bec5e85dc7ab9b24eec8b0ff2accaa3e59b2312b6d539413d"} err="failed to get container status \"5a3bbef81765195bec5e85dc7ab9b24eec8b0ff2accaa3e59b2312b6d539413d\": rpc error: code = NotFound desc = could not find container \"5a3bbef81765195bec5e85dc7ab9b24eec8b0ff2accaa3e59b2312b6d539413d\": container with ID starting with 5a3bbef81765195bec5e85dc7ab9b24eec8b0ff2accaa3e59b2312b6d539413d not found: ID does not exist" Dec 11 02:06:17 crc kubenswrapper[4824]: I1211 02:06:17.993751 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-qr5xz"] Dec 11 02:06:18 crc kubenswrapper[4824]: I1211 02:06:18.007848 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-qr5xz"] Dec 11 02:06:18 crc kubenswrapper[4824]: I1211 02:06:18.066971 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5407364a-e2b6-43b4-9eb0-deb12a022289-config\") on node \"crc\" DevicePath \"\"" Dec 11 02:06:18 crc kubenswrapper[4824]: I1211 02:06:18.067016 4824 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5407364a-e2b6-43b4-9eb0-deb12a022289-client-ca\") on node \"crc\" DevicePath \"\"" Dec 11 02:06:18 crc kubenswrapper[4824]: I1211 02:06:18.067027 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4x4bd\" (UniqueName: \"kubernetes.io/projected/5407364a-e2b6-43b4-9eb0-deb12a022289-kube-api-access-4x4bd\") on node \"crc\" DevicePath \"\"" Dec 11 02:06:18 crc kubenswrapper[4824]: I1211 02:06:18.067037 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5407364a-e2b6-43b4-9eb0-deb12a022289-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 02:06:18 crc kubenswrapper[4824]: I1211 02:06:18.093012 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-5f76685dbc-k5nss"] Dec 11 02:06:18 crc kubenswrapper[4824]: E1211 02:06:18.093259 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5407364a-e2b6-43b4-9eb0-deb12a022289" containerName="route-controller-manager" Dec 11 02:06:18 crc kubenswrapper[4824]: I1211 02:06:18.093277 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="5407364a-e2b6-43b4-9eb0-deb12a022289" containerName="route-controller-manager" Dec 11 02:06:18 crc kubenswrapper[4824]: E1211 02:06:18.093296 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e099899-aaf7-4bd8-b37e-8616b5060e33" containerName="controller-manager" Dec 11 02:06:18 crc kubenswrapper[4824]: I1211 02:06:18.093303 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e099899-aaf7-4bd8-b37e-8616b5060e33" containerName="controller-manager" Dec 11 02:06:18 crc kubenswrapper[4824]: I1211 02:06:18.093395 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e099899-aaf7-4bd8-b37e-8616b5060e33" containerName="controller-manager" Dec 11 02:06:18 crc kubenswrapper[4824]: I1211 02:06:18.093409 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="5407364a-e2b6-43b4-9eb0-deb12a022289" containerName="route-controller-manager" Dec 11 02:06:18 crc kubenswrapper[4824]: I1211 02:06:18.094194 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5f76685dbc-k5nss" Dec 11 02:06:18 crc kubenswrapper[4824]: I1211 02:06:18.095875 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8678c97bff-pbrkn"] Dec 11 02:06:18 crc kubenswrapper[4824]: I1211 02:06:18.096475 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-8678c97bff-pbrkn" Dec 11 02:06:18 crc kubenswrapper[4824]: I1211 02:06:18.100497 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 11 02:06:18 crc kubenswrapper[4824]: I1211 02:06:18.102673 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 11 02:06:18 crc kubenswrapper[4824]: I1211 02:06:18.102703 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 11 02:06:18 crc kubenswrapper[4824]: I1211 02:06:18.102767 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 11 02:06:18 crc kubenswrapper[4824]: I1211 02:06:18.102965 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 11 02:06:18 crc kubenswrapper[4824]: I1211 02:06:18.103138 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 11 02:06:18 crc kubenswrapper[4824]: I1211 02:06:18.107516 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5f76685dbc-k5nss"] Dec 11 02:06:18 crc kubenswrapper[4824]: I1211 02:06:18.109883 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 11 02:06:18 crc kubenswrapper[4824]: I1211 02:06:18.144987 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8678c97bff-pbrkn"] Dec 11 02:06:18 crc kubenswrapper[4824]: I1211 02:06:18.168700 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fdb352f0-9e8e-4966-adcf-de078c33017a-client-ca\") pod \"controller-manager-5f76685dbc-k5nss\" (UID: \"fdb352f0-9e8e-4966-adcf-de078c33017a\") " pod="openshift-controller-manager/controller-manager-5f76685dbc-k5nss" Dec 11 02:06:18 crc kubenswrapper[4824]: I1211 02:06:18.168913 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fdb352f0-9e8e-4966-adcf-de078c33017a-config\") pod \"controller-manager-5f76685dbc-k5nss\" (UID: \"fdb352f0-9e8e-4966-adcf-de078c33017a\") " pod="openshift-controller-manager/controller-manager-5f76685dbc-k5nss" Dec 11 02:06:18 crc kubenswrapper[4824]: I1211 02:06:18.169048 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15b3302c-517d-4508-bd7c-047c457e7b8f-config\") pod \"route-controller-manager-8678c97bff-pbrkn\" (UID: \"15b3302c-517d-4508-bd7c-047c457e7b8f\") " pod="openshift-route-controller-manager/route-controller-manager-8678c97bff-pbrkn" Dec 11 02:06:18 crc kubenswrapper[4824]: I1211 02:06:18.169175 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fdb352f0-9e8e-4966-adcf-de078c33017a-serving-cert\") pod \"controller-manager-5f76685dbc-k5nss\" (UID: \"fdb352f0-9e8e-4966-adcf-de078c33017a\") " pod="openshift-controller-manager/controller-manager-5f76685dbc-k5nss" Dec 11 02:06:18 crc kubenswrapper[4824]: I1211 02:06:18.169341 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/15b3302c-517d-4508-bd7c-047c457e7b8f-client-ca\") pod \"route-controller-manager-8678c97bff-pbrkn\" (UID: \"15b3302c-517d-4508-bd7c-047c457e7b8f\") " pod="openshift-route-controller-manager/route-controller-manager-8678c97bff-pbrkn" Dec 11 02:06:18 crc kubenswrapper[4824]: I1211 02:06:18.169471 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjhfx\" (UniqueName: \"kubernetes.io/projected/15b3302c-517d-4508-bd7c-047c457e7b8f-kube-api-access-bjhfx\") pod \"route-controller-manager-8678c97bff-pbrkn\" (UID: \"15b3302c-517d-4508-bd7c-047c457e7b8f\") " pod="openshift-route-controller-manager/route-controller-manager-8678c97bff-pbrkn" Dec 11 02:06:18 crc kubenswrapper[4824]: I1211 02:06:18.169598 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mtztj\" (UniqueName: \"kubernetes.io/projected/fdb352f0-9e8e-4966-adcf-de078c33017a-kube-api-access-mtztj\") pod \"controller-manager-5f76685dbc-k5nss\" (UID: \"fdb352f0-9e8e-4966-adcf-de078c33017a\") " pod="openshift-controller-manager/controller-manager-5f76685dbc-k5nss" Dec 11 02:06:18 crc kubenswrapper[4824]: I1211 02:06:18.169717 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/fdb352f0-9e8e-4966-adcf-de078c33017a-proxy-ca-bundles\") pod \"controller-manager-5f76685dbc-k5nss\" (UID: \"fdb352f0-9e8e-4966-adcf-de078c33017a\") " pod="openshift-controller-manager/controller-manager-5f76685dbc-k5nss" Dec 11 02:06:18 crc kubenswrapper[4824]: I1211 02:06:18.169835 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/15b3302c-517d-4508-bd7c-047c457e7b8f-serving-cert\") pod \"route-controller-manager-8678c97bff-pbrkn\" (UID: \"15b3302c-517d-4508-bd7c-047c457e7b8f\") " pod="openshift-route-controller-manager/route-controller-manager-8678c97bff-pbrkn" Dec 11 02:06:18 crc kubenswrapper[4824]: I1211 02:06:18.270750 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15b3302c-517d-4508-bd7c-047c457e7b8f-config\") pod \"route-controller-manager-8678c97bff-pbrkn\" (UID: \"15b3302c-517d-4508-bd7c-047c457e7b8f\") " pod="openshift-route-controller-manager/route-controller-manager-8678c97bff-pbrkn" Dec 11 02:06:18 crc kubenswrapper[4824]: I1211 02:06:18.270987 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fdb352f0-9e8e-4966-adcf-de078c33017a-serving-cert\") pod \"controller-manager-5f76685dbc-k5nss\" (UID: \"fdb352f0-9e8e-4966-adcf-de078c33017a\") " pod="openshift-controller-manager/controller-manager-5f76685dbc-k5nss" Dec 11 02:06:18 crc kubenswrapper[4824]: I1211 02:06:18.272639 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/15b3302c-517d-4508-bd7c-047c457e7b8f-client-ca\") pod \"route-controller-manager-8678c97bff-pbrkn\" (UID: \"15b3302c-517d-4508-bd7c-047c457e7b8f\") " pod="openshift-route-controller-manager/route-controller-manager-8678c97bff-pbrkn" Dec 11 02:06:18 crc kubenswrapper[4824]: I1211 02:06:18.272751 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bjhfx\" (UniqueName: \"kubernetes.io/projected/15b3302c-517d-4508-bd7c-047c457e7b8f-kube-api-access-bjhfx\") pod \"route-controller-manager-8678c97bff-pbrkn\" (UID: \"15b3302c-517d-4508-bd7c-047c457e7b8f\") " pod="openshift-route-controller-manager/route-controller-manager-8678c97bff-pbrkn" Dec 11 02:06:18 crc kubenswrapper[4824]: I1211 02:06:18.272849 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mtztj\" (UniqueName: \"kubernetes.io/projected/fdb352f0-9e8e-4966-adcf-de078c33017a-kube-api-access-mtztj\") pod \"controller-manager-5f76685dbc-k5nss\" (UID: \"fdb352f0-9e8e-4966-adcf-de078c33017a\") " pod="openshift-controller-manager/controller-manager-5f76685dbc-k5nss" Dec 11 02:06:18 crc kubenswrapper[4824]: I1211 02:06:18.272961 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/fdb352f0-9e8e-4966-adcf-de078c33017a-proxy-ca-bundles\") pod \"controller-manager-5f76685dbc-k5nss\" (UID: \"fdb352f0-9e8e-4966-adcf-de078c33017a\") " pod="openshift-controller-manager/controller-manager-5f76685dbc-k5nss" Dec 11 02:06:18 crc kubenswrapper[4824]: I1211 02:06:18.273093 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/15b3302c-517d-4508-bd7c-047c457e7b8f-serving-cert\") pod \"route-controller-manager-8678c97bff-pbrkn\" (UID: \"15b3302c-517d-4508-bd7c-047c457e7b8f\") " pod="openshift-route-controller-manager/route-controller-manager-8678c97bff-pbrkn" Dec 11 02:06:18 crc kubenswrapper[4824]: I1211 02:06:18.273289 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fdb352f0-9e8e-4966-adcf-de078c33017a-client-ca\") pod \"controller-manager-5f76685dbc-k5nss\" (UID: \"fdb352f0-9e8e-4966-adcf-de078c33017a\") " pod="openshift-controller-manager/controller-manager-5f76685dbc-k5nss" Dec 11 02:06:18 crc kubenswrapper[4824]: I1211 02:06:18.273407 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fdb352f0-9e8e-4966-adcf-de078c33017a-config\") pod \"controller-manager-5f76685dbc-k5nss\" (UID: \"fdb352f0-9e8e-4966-adcf-de078c33017a\") " pod="openshift-controller-manager/controller-manager-5f76685dbc-k5nss" Dec 11 02:06:18 crc kubenswrapper[4824]: I1211 02:06:18.272220 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15b3302c-517d-4508-bd7c-047c457e7b8f-config\") pod \"route-controller-manager-8678c97bff-pbrkn\" (UID: \"15b3302c-517d-4508-bd7c-047c457e7b8f\") " pod="openshift-route-controller-manager/route-controller-manager-8678c97bff-pbrkn" Dec 11 02:06:18 crc kubenswrapper[4824]: I1211 02:06:18.274764 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/15b3302c-517d-4508-bd7c-047c457e7b8f-client-ca\") pod \"route-controller-manager-8678c97bff-pbrkn\" (UID: \"15b3302c-517d-4508-bd7c-047c457e7b8f\") " pod="openshift-route-controller-manager/route-controller-manager-8678c97bff-pbrkn" Dec 11 02:06:18 crc kubenswrapper[4824]: I1211 02:06:18.275170 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fdb352f0-9e8e-4966-adcf-de078c33017a-client-ca\") pod \"controller-manager-5f76685dbc-k5nss\" (UID: \"fdb352f0-9e8e-4966-adcf-de078c33017a\") " pod="openshift-controller-manager/controller-manager-5f76685dbc-k5nss" Dec 11 02:06:18 crc kubenswrapper[4824]: I1211 02:06:18.279956 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/fdb352f0-9e8e-4966-adcf-de078c33017a-proxy-ca-bundles\") pod \"controller-manager-5f76685dbc-k5nss\" (UID: \"fdb352f0-9e8e-4966-adcf-de078c33017a\") " pod="openshift-controller-manager/controller-manager-5f76685dbc-k5nss" Dec 11 02:06:18 crc kubenswrapper[4824]: I1211 02:06:18.282571 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fdb352f0-9e8e-4966-adcf-de078c33017a-config\") pod \"controller-manager-5f76685dbc-k5nss\" (UID: \"fdb352f0-9e8e-4966-adcf-de078c33017a\") " pod="openshift-controller-manager/controller-manager-5f76685dbc-k5nss" Dec 11 02:06:18 crc kubenswrapper[4824]: I1211 02:06:18.286903 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-msljv"] Dec 11 02:06:18 crc kubenswrapper[4824]: I1211 02:06:18.287751 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fdb352f0-9e8e-4966-adcf-de078c33017a-serving-cert\") pod \"controller-manager-5f76685dbc-k5nss\" (UID: \"fdb352f0-9e8e-4966-adcf-de078c33017a\") " pod="openshift-controller-manager/controller-manager-5f76685dbc-k5nss" Dec 11 02:06:18 crc kubenswrapper[4824]: I1211 02:06:18.292833 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/15b3302c-517d-4508-bd7c-047c457e7b8f-serving-cert\") pod \"route-controller-manager-8678c97bff-pbrkn\" (UID: \"15b3302c-517d-4508-bd7c-047c457e7b8f\") " pod="openshift-route-controller-manager/route-controller-manager-8678c97bff-pbrkn" Dec 11 02:06:18 crc kubenswrapper[4824]: I1211 02:06:18.294396 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-msljv"] Dec 11 02:06:18 crc kubenswrapper[4824]: I1211 02:06:18.303516 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bjhfx\" (UniqueName: \"kubernetes.io/projected/15b3302c-517d-4508-bd7c-047c457e7b8f-kube-api-access-bjhfx\") pod \"route-controller-manager-8678c97bff-pbrkn\" (UID: \"15b3302c-517d-4508-bd7c-047c457e7b8f\") " pod="openshift-route-controller-manager/route-controller-manager-8678c97bff-pbrkn" Dec 11 02:06:18 crc kubenswrapper[4824]: I1211 02:06:18.309344 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mtztj\" (UniqueName: \"kubernetes.io/projected/fdb352f0-9e8e-4966-adcf-de078c33017a-kube-api-access-mtztj\") pod \"controller-manager-5f76685dbc-k5nss\" (UID: \"fdb352f0-9e8e-4966-adcf-de078c33017a\") " pod="openshift-controller-manager/controller-manager-5f76685dbc-k5nss" Dec 11 02:06:18 crc kubenswrapper[4824]: I1211 02:06:18.418815 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5f76685dbc-k5nss" Dec 11 02:06:18 crc kubenswrapper[4824]: I1211 02:06:18.434836 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-8678c97bff-pbrkn" Dec 11 02:06:18 crc kubenswrapper[4824]: I1211 02:06:18.628328 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5f76685dbc-k5nss"] Dec 11 02:06:18 crc kubenswrapper[4824]: I1211 02:06:18.647609 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5407364a-e2b6-43b4-9eb0-deb12a022289" path="/var/lib/kubelet/pods/5407364a-e2b6-43b4-9eb0-deb12a022289/volumes" Dec 11 02:06:18 crc kubenswrapper[4824]: I1211 02:06:18.648909 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e099899-aaf7-4bd8-b37e-8616b5060e33" path="/var/lib/kubelet/pods/9e099899-aaf7-4bd8-b37e-8616b5060e33/volumes" Dec 11 02:06:18 crc kubenswrapper[4824]: I1211 02:06:18.687394 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8678c97bff-pbrkn"] Dec 11 02:06:18 crc kubenswrapper[4824]: W1211 02:06:18.699786 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod15b3302c_517d_4508_bd7c_047c457e7b8f.slice/crio-d439974e8b30ab4f9871877b701df11c7486e59818d3cced1f45e668300c8738 WatchSource:0}: Error finding container d439974e8b30ab4f9871877b701df11c7486e59818d3cced1f45e668300c8738: Status 404 returned error can't find the container with id d439974e8b30ab4f9871877b701df11c7486e59818d3cced1f45e668300c8738 Dec 11 02:06:18 crc kubenswrapper[4824]: I1211 02:06:18.963468 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5f76685dbc-k5nss" event={"ID":"fdb352f0-9e8e-4966-adcf-de078c33017a","Type":"ContainerStarted","Data":"4b3066c01d8e4d563965bcdfbeabd49094dd198cf3ebef20e82420e19d474d83"} Dec 11 02:06:18 crc kubenswrapper[4824]: I1211 02:06:18.963769 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5f76685dbc-k5nss" event={"ID":"fdb352f0-9e8e-4966-adcf-de078c33017a","Type":"ContainerStarted","Data":"bfeeef06a69c4cdd0398fdf8ce49b259e408e29ab781a3d45b4ca5ca82682718"} Dec 11 02:06:18 crc kubenswrapper[4824]: I1211 02:06:18.964203 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-5f76685dbc-k5nss" Dec 11 02:06:18 crc kubenswrapper[4824]: I1211 02:06:18.965474 4824 patch_prober.go:28] interesting pod/controller-manager-5f76685dbc-k5nss container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.57:8443/healthz\": dial tcp 10.217.0.57:8443: connect: connection refused" start-of-body= Dec 11 02:06:18 crc kubenswrapper[4824]: I1211 02:06:18.965516 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-5f76685dbc-k5nss" podUID="fdb352f0-9e8e-4966-adcf-de078c33017a" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.57:8443/healthz\": dial tcp 10.217.0.57:8443: connect: connection refused" Dec 11 02:06:18 crc kubenswrapper[4824]: I1211 02:06:18.966193 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-8678c97bff-pbrkn" event={"ID":"15b3302c-517d-4508-bd7c-047c457e7b8f","Type":"ContainerStarted","Data":"f5e18d920abc0e36ec82e319d135182204e4d04c1ed4a29acba5a5156b0fd381"} Dec 11 02:06:18 crc kubenswrapper[4824]: I1211 02:06:18.966238 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-8678c97bff-pbrkn" event={"ID":"15b3302c-517d-4508-bd7c-047c457e7b8f","Type":"ContainerStarted","Data":"d439974e8b30ab4f9871877b701df11c7486e59818d3cced1f45e668300c8738"} Dec 11 02:06:18 crc kubenswrapper[4824]: I1211 02:06:18.966366 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-8678c97bff-pbrkn" Dec 11 02:06:18 crc kubenswrapper[4824]: I1211 02:06:18.967599 4824 patch_prober.go:28] interesting pod/route-controller-manager-8678c97bff-pbrkn container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.58:8443/healthz\": dial tcp 10.217.0.58:8443: connect: connection refused" start-of-body= Dec 11 02:06:18 crc kubenswrapper[4824]: I1211 02:06:18.967636 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-8678c97bff-pbrkn" podUID="15b3302c-517d-4508-bd7c-047c457e7b8f" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.58:8443/healthz\": dial tcp 10.217.0.58:8443: connect: connection refused" Dec 11 02:06:18 crc kubenswrapper[4824]: I1211 02:06:18.988807 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-5f76685dbc-k5nss" podStartSLOduration=2.988786357 podStartE2EDuration="2.988786357s" podCreationTimestamp="2025-12-11 02:06:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:06:18.983481067 +0000 UTC m=+320.672518466" watchObservedRunningTime="2025-12-11 02:06:18.988786357 +0000 UTC m=+320.677823746" Dec 11 02:06:19 crc kubenswrapper[4824]: I1211 02:06:19.000684 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-8678c97bff-pbrkn" podStartSLOduration=3.00066131 podStartE2EDuration="3.00066131s" podCreationTimestamp="2025-12-11 02:06:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:06:18.999680693 +0000 UTC m=+320.688718082" watchObservedRunningTime="2025-12-11 02:06:19.00066131 +0000 UTC m=+320.689698709" Dec 11 02:06:19 crc kubenswrapper[4824]: I1211 02:06:19.980665 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-5f76685dbc-k5nss" Dec 11 02:06:19 crc kubenswrapper[4824]: I1211 02:06:19.980736 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-8678c97bff-pbrkn" Dec 11 02:06:33 crc kubenswrapper[4824]: I1211 02:06:33.352978 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8678c97bff-pbrkn"] Dec 11 02:06:33 crc kubenswrapper[4824]: I1211 02:06:33.354103 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-8678c97bff-pbrkn" podUID="15b3302c-517d-4508-bd7c-047c457e7b8f" containerName="route-controller-manager" containerID="cri-o://f5e18d920abc0e36ec82e319d135182204e4d04c1ed4a29acba5a5156b0fd381" gracePeriod=30 Dec 11 02:06:34 crc kubenswrapper[4824]: I1211 02:06:34.046885 4824 generic.go:334] "Generic (PLEG): container finished" podID="15b3302c-517d-4508-bd7c-047c457e7b8f" containerID="f5e18d920abc0e36ec82e319d135182204e4d04c1ed4a29acba5a5156b0fd381" exitCode=0 Dec 11 02:06:34 crc kubenswrapper[4824]: I1211 02:06:34.046924 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-8678c97bff-pbrkn" event={"ID":"15b3302c-517d-4508-bd7c-047c457e7b8f","Type":"ContainerDied","Data":"f5e18d920abc0e36ec82e319d135182204e4d04c1ed4a29acba5a5156b0fd381"} Dec 11 02:06:34 crc kubenswrapper[4824]: I1211 02:06:34.809640 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-8678c97bff-pbrkn" Dec 11 02:06:34 crc kubenswrapper[4824]: I1211 02:06:34.849291 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7688c94445-969fh"] Dec 11 02:06:34 crc kubenswrapper[4824]: E1211 02:06:34.849601 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15b3302c-517d-4508-bd7c-047c457e7b8f" containerName="route-controller-manager" Dec 11 02:06:34 crc kubenswrapper[4824]: I1211 02:06:34.849622 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="15b3302c-517d-4508-bd7c-047c457e7b8f" containerName="route-controller-manager" Dec 11 02:06:34 crc kubenswrapper[4824]: I1211 02:06:34.849740 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="15b3302c-517d-4508-bd7c-047c457e7b8f" containerName="route-controller-manager" Dec 11 02:06:34 crc kubenswrapper[4824]: I1211 02:06:34.850205 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7688c94445-969fh" Dec 11 02:06:34 crc kubenswrapper[4824]: I1211 02:06:34.854305 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7688c94445-969fh"] Dec 11 02:06:34 crc kubenswrapper[4824]: I1211 02:06:34.893192 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/15b3302c-517d-4508-bd7c-047c457e7b8f-client-ca\") pod \"15b3302c-517d-4508-bd7c-047c457e7b8f\" (UID: \"15b3302c-517d-4508-bd7c-047c457e7b8f\") " Dec 11 02:06:34 crc kubenswrapper[4824]: I1211 02:06:34.893281 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15b3302c-517d-4508-bd7c-047c457e7b8f-config\") pod \"15b3302c-517d-4508-bd7c-047c457e7b8f\" (UID: \"15b3302c-517d-4508-bd7c-047c457e7b8f\") " Dec 11 02:06:34 crc kubenswrapper[4824]: I1211 02:06:34.893334 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bjhfx\" (UniqueName: \"kubernetes.io/projected/15b3302c-517d-4508-bd7c-047c457e7b8f-kube-api-access-bjhfx\") pod \"15b3302c-517d-4508-bd7c-047c457e7b8f\" (UID: \"15b3302c-517d-4508-bd7c-047c457e7b8f\") " Dec 11 02:06:34 crc kubenswrapper[4824]: I1211 02:06:34.893382 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/15b3302c-517d-4508-bd7c-047c457e7b8f-serving-cert\") pod \"15b3302c-517d-4508-bd7c-047c457e7b8f\" (UID: \"15b3302c-517d-4508-bd7c-047c457e7b8f\") " Dec 11 02:06:34 crc kubenswrapper[4824]: I1211 02:06:34.893579 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/53944f76-cb51-42e3-8be4-b44597818fd5-config\") pod \"route-controller-manager-7688c94445-969fh\" (UID: \"53944f76-cb51-42e3-8be4-b44597818fd5\") " pod="openshift-route-controller-manager/route-controller-manager-7688c94445-969fh" Dec 11 02:06:34 crc kubenswrapper[4824]: I1211 02:06:34.893611 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/53944f76-cb51-42e3-8be4-b44597818fd5-client-ca\") pod \"route-controller-manager-7688c94445-969fh\" (UID: \"53944f76-cb51-42e3-8be4-b44597818fd5\") " pod="openshift-route-controller-manager/route-controller-manager-7688c94445-969fh" Dec 11 02:06:34 crc kubenswrapper[4824]: I1211 02:06:34.893695 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/53944f76-cb51-42e3-8be4-b44597818fd5-serving-cert\") pod \"route-controller-manager-7688c94445-969fh\" (UID: \"53944f76-cb51-42e3-8be4-b44597818fd5\") " pod="openshift-route-controller-manager/route-controller-manager-7688c94445-969fh" Dec 11 02:06:34 crc kubenswrapper[4824]: I1211 02:06:34.893767 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hzkt6\" (UniqueName: \"kubernetes.io/projected/53944f76-cb51-42e3-8be4-b44597818fd5-kube-api-access-hzkt6\") pod \"route-controller-manager-7688c94445-969fh\" (UID: \"53944f76-cb51-42e3-8be4-b44597818fd5\") " pod="openshift-route-controller-manager/route-controller-manager-7688c94445-969fh" Dec 11 02:06:34 crc kubenswrapper[4824]: I1211 02:06:34.894021 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/15b3302c-517d-4508-bd7c-047c457e7b8f-client-ca" (OuterVolumeSpecName: "client-ca") pod "15b3302c-517d-4508-bd7c-047c457e7b8f" (UID: "15b3302c-517d-4508-bd7c-047c457e7b8f"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:06:34 crc kubenswrapper[4824]: I1211 02:06:34.894527 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/15b3302c-517d-4508-bd7c-047c457e7b8f-config" (OuterVolumeSpecName: "config") pod "15b3302c-517d-4508-bd7c-047c457e7b8f" (UID: "15b3302c-517d-4508-bd7c-047c457e7b8f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:06:34 crc kubenswrapper[4824]: I1211 02:06:34.899478 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15b3302c-517d-4508-bd7c-047c457e7b8f-kube-api-access-bjhfx" (OuterVolumeSpecName: "kube-api-access-bjhfx") pod "15b3302c-517d-4508-bd7c-047c457e7b8f" (UID: "15b3302c-517d-4508-bd7c-047c457e7b8f"). InnerVolumeSpecName "kube-api-access-bjhfx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:06:34 crc kubenswrapper[4824]: I1211 02:06:34.899695 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/15b3302c-517d-4508-bd7c-047c457e7b8f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "15b3302c-517d-4508-bd7c-047c457e7b8f" (UID: "15b3302c-517d-4508-bd7c-047c457e7b8f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:06:34 crc kubenswrapper[4824]: I1211 02:06:34.994680 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/53944f76-cb51-42e3-8be4-b44597818fd5-config\") pod \"route-controller-manager-7688c94445-969fh\" (UID: \"53944f76-cb51-42e3-8be4-b44597818fd5\") " pod="openshift-route-controller-manager/route-controller-manager-7688c94445-969fh" Dec 11 02:06:34 crc kubenswrapper[4824]: I1211 02:06:34.994753 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/53944f76-cb51-42e3-8be4-b44597818fd5-client-ca\") pod \"route-controller-manager-7688c94445-969fh\" (UID: \"53944f76-cb51-42e3-8be4-b44597818fd5\") " pod="openshift-route-controller-manager/route-controller-manager-7688c94445-969fh" Dec 11 02:06:34 crc kubenswrapper[4824]: I1211 02:06:34.994849 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/53944f76-cb51-42e3-8be4-b44597818fd5-serving-cert\") pod \"route-controller-manager-7688c94445-969fh\" (UID: \"53944f76-cb51-42e3-8be4-b44597818fd5\") " pod="openshift-route-controller-manager/route-controller-manager-7688c94445-969fh" Dec 11 02:06:34 crc kubenswrapper[4824]: I1211 02:06:34.994886 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hzkt6\" (UniqueName: \"kubernetes.io/projected/53944f76-cb51-42e3-8be4-b44597818fd5-kube-api-access-hzkt6\") pod \"route-controller-manager-7688c94445-969fh\" (UID: \"53944f76-cb51-42e3-8be4-b44597818fd5\") " pod="openshift-route-controller-manager/route-controller-manager-7688c94445-969fh" Dec 11 02:06:34 crc kubenswrapper[4824]: I1211 02:06:34.994968 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bjhfx\" (UniqueName: \"kubernetes.io/projected/15b3302c-517d-4508-bd7c-047c457e7b8f-kube-api-access-bjhfx\") on node \"crc\" DevicePath \"\"" Dec 11 02:06:34 crc kubenswrapper[4824]: I1211 02:06:34.994988 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/15b3302c-517d-4508-bd7c-047c457e7b8f-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 02:06:34 crc kubenswrapper[4824]: I1211 02:06:34.995006 4824 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/15b3302c-517d-4508-bd7c-047c457e7b8f-client-ca\") on node \"crc\" DevicePath \"\"" Dec 11 02:06:34 crc kubenswrapper[4824]: I1211 02:06:34.995067 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15b3302c-517d-4508-bd7c-047c457e7b8f-config\") on node \"crc\" DevicePath \"\"" Dec 11 02:06:34 crc kubenswrapper[4824]: I1211 02:06:34.996897 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/53944f76-cb51-42e3-8be4-b44597818fd5-config\") pod \"route-controller-manager-7688c94445-969fh\" (UID: \"53944f76-cb51-42e3-8be4-b44597818fd5\") " pod="openshift-route-controller-manager/route-controller-manager-7688c94445-969fh" Dec 11 02:06:34 crc kubenswrapper[4824]: I1211 02:06:34.999308 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/53944f76-cb51-42e3-8be4-b44597818fd5-client-ca\") pod \"route-controller-manager-7688c94445-969fh\" (UID: \"53944f76-cb51-42e3-8be4-b44597818fd5\") " pod="openshift-route-controller-manager/route-controller-manager-7688c94445-969fh" Dec 11 02:06:35 crc kubenswrapper[4824]: I1211 02:06:35.001886 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/53944f76-cb51-42e3-8be4-b44597818fd5-serving-cert\") pod \"route-controller-manager-7688c94445-969fh\" (UID: \"53944f76-cb51-42e3-8be4-b44597818fd5\") " pod="openshift-route-controller-manager/route-controller-manager-7688c94445-969fh" Dec 11 02:06:35 crc kubenswrapper[4824]: I1211 02:06:35.014402 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hzkt6\" (UniqueName: \"kubernetes.io/projected/53944f76-cb51-42e3-8be4-b44597818fd5-kube-api-access-hzkt6\") pod \"route-controller-manager-7688c94445-969fh\" (UID: \"53944f76-cb51-42e3-8be4-b44597818fd5\") " pod="openshift-route-controller-manager/route-controller-manager-7688c94445-969fh" Dec 11 02:06:35 crc kubenswrapper[4824]: I1211 02:06:35.064917 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-8678c97bff-pbrkn" event={"ID":"15b3302c-517d-4508-bd7c-047c457e7b8f","Type":"ContainerDied","Data":"d439974e8b30ab4f9871877b701df11c7486e59818d3cced1f45e668300c8738"} Dec 11 02:06:35 crc kubenswrapper[4824]: I1211 02:06:35.064992 4824 scope.go:117] "RemoveContainer" containerID="f5e18d920abc0e36ec82e319d135182204e4d04c1ed4a29acba5a5156b0fd381" Dec 11 02:06:35 crc kubenswrapper[4824]: I1211 02:06:35.065359 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-8678c97bff-pbrkn" Dec 11 02:06:35 crc kubenswrapper[4824]: I1211 02:06:35.102707 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8678c97bff-pbrkn"] Dec 11 02:06:35 crc kubenswrapper[4824]: I1211 02:06:35.107686 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8678c97bff-pbrkn"] Dec 11 02:06:35 crc kubenswrapper[4824]: I1211 02:06:35.166338 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7688c94445-969fh" Dec 11 02:06:35 crc kubenswrapper[4824]: I1211 02:06:35.583494 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7688c94445-969fh"] Dec 11 02:06:35 crc kubenswrapper[4824]: W1211 02:06:35.593033 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod53944f76_cb51_42e3_8be4_b44597818fd5.slice/crio-df701b236f191c42ee2c3f183cffe6b62351dcdd93f334243bb6681154375e70 WatchSource:0}: Error finding container df701b236f191c42ee2c3f183cffe6b62351dcdd93f334243bb6681154375e70: Status 404 returned error can't find the container with id df701b236f191c42ee2c3f183cffe6b62351dcdd93f334243bb6681154375e70 Dec 11 02:06:36 crc kubenswrapper[4824]: I1211 02:06:36.071620 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7688c94445-969fh" event={"ID":"53944f76-cb51-42e3-8be4-b44597818fd5","Type":"ContainerStarted","Data":"3f0c2531a167ebbe938ef08b2875eb388f9e037581099a552931dcce867812b0"} Dec 11 02:06:36 crc kubenswrapper[4824]: I1211 02:06:36.071665 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7688c94445-969fh" event={"ID":"53944f76-cb51-42e3-8be4-b44597818fd5","Type":"ContainerStarted","Data":"df701b236f191c42ee2c3f183cffe6b62351dcdd93f334243bb6681154375e70"} Dec 11 02:06:36 crc kubenswrapper[4824]: I1211 02:06:36.071964 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-7688c94445-969fh" Dec 11 02:06:36 crc kubenswrapper[4824]: I1211 02:06:36.095243 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-7688c94445-969fh" podStartSLOduration=3.095221251 podStartE2EDuration="3.095221251s" podCreationTimestamp="2025-12-11 02:06:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:06:36.093155013 +0000 UTC m=+337.782192412" watchObservedRunningTime="2025-12-11 02:06:36.095221251 +0000 UTC m=+337.784258630" Dec 11 02:06:36 crc kubenswrapper[4824]: I1211 02:06:36.205822 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-7688c94445-969fh" Dec 11 02:06:36 crc kubenswrapper[4824]: I1211 02:06:36.643206 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15b3302c-517d-4508-bd7c-047c457e7b8f" path="/var/lib/kubelet/pods/15b3302c-517d-4508-bd7c-047c457e7b8f/volumes" Dec 11 02:06:46 crc kubenswrapper[4824]: I1211 02:06:46.910504 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nqjh5"] Dec 11 02:06:46 crc kubenswrapper[4824]: I1211 02:06:46.911337 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-nqjh5" podUID="d9806a3c-cb3d-4b3e-8889-11f8cf49c030" containerName="registry-server" containerID="cri-o://0f6c6d568fd4ead23e24d85c0ced23037ff04686489fd1752516ab279b1e63b3" gracePeriod=30 Dec 11 02:06:46 crc kubenswrapper[4824]: I1211 02:06:46.929520 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-86mf9"] Dec 11 02:06:46 crc kubenswrapper[4824]: I1211 02:06:46.930003 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-86mf9" podUID="b2b7c06f-4350-41be-b472-c75d21558911" containerName="registry-server" containerID="cri-o://15910645eb6869cc7816a7bae3d43c849e26e86197c452fe3b6a9e0df7d31b82" gracePeriod=30 Dec 11 02:06:46 crc kubenswrapper[4824]: I1211 02:06:46.936063 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-z42g5"] Dec 11 02:06:46 crc kubenswrapper[4824]: I1211 02:06:46.936400 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-z42g5" podUID="ffb4acd4-543f-4509-b993-249e3656048d" containerName="marketplace-operator" containerID="cri-o://60693fc1fe023dcf8dbae48f9d5d679fb604f2a41ce2a4e0c21987b053e0c4f1" gracePeriod=30 Dec 11 02:06:46 crc kubenswrapper[4824]: I1211 02:06:46.942932 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jc49g"] Dec 11 02:06:46 crc kubenswrapper[4824]: I1211 02:06:46.943240 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-jc49g" podUID="6127fb96-7f67-453d-8d16-40fac34e6482" containerName="registry-server" containerID="cri-o://bd6609faa0c263d1acb6b286455bfc86a0fb4b396c42cf8eeb5c1a12469e1ec6" gracePeriod=30 Dec 11 02:06:46 crc kubenswrapper[4824]: I1211 02:06:46.948754 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rnmtp"] Dec 11 02:06:46 crc kubenswrapper[4824]: I1211 02:06:46.949142 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-rnmtp" podUID="0b3b02dc-e8f5-4629-b1bf-72febdb40843" containerName="registry-server" containerID="cri-o://844683dffb7882a5cc33ed4c458f332f8ffa61080cfdae10d2ce9819e505dc34" gracePeriod=30 Dec 11 02:06:46 crc kubenswrapper[4824]: I1211 02:06:46.959676 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-xj7qj"] Dec 11 02:06:46 crc kubenswrapper[4824]: I1211 02:06:46.960628 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-xj7qj" Dec 11 02:06:46 crc kubenswrapper[4824]: I1211 02:06:46.978746 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-xj7qj"] Dec 11 02:06:47 crc kubenswrapper[4824]: I1211 02:06:47.045731 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-msrlm\" (UniqueName: \"kubernetes.io/projected/19528072-20f2-4741-bfd9-b7c44430b834-kube-api-access-msrlm\") pod \"marketplace-operator-79b997595-xj7qj\" (UID: \"19528072-20f2-4741-bfd9-b7c44430b834\") " pod="openshift-marketplace/marketplace-operator-79b997595-xj7qj" Dec 11 02:06:47 crc kubenswrapper[4824]: I1211 02:06:47.045808 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/19528072-20f2-4741-bfd9-b7c44430b834-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-xj7qj\" (UID: \"19528072-20f2-4741-bfd9-b7c44430b834\") " pod="openshift-marketplace/marketplace-operator-79b997595-xj7qj" Dec 11 02:06:47 crc kubenswrapper[4824]: I1211 02:06:47.045977 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/19528072-20f2-4741-bfd9-b7c44430b834-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-xj7qj\" (UID: \"19528072-20f2-4741-bfd9-b7c44430b834\") " pod="openshift-marketplace/marketplace-operator-79b997595-xj7qj" Dec 11 02:06:47 crc kubenswrapper[4824]: I1211 02:06:47.146906 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-msrlm\" (UniqueName: \"kubernetes.io/projected/19528072-20f2-4741-bfd9-b7c44430b834-kube-api-access-msrlm\") pod \"marketplace-operator-79b997595-xj7qj\" (UID: \"19528072-20f2-4741-bfd9-b7c44430b834\") " pod="openshift-marketplace/marketplace-operator-79b997595-xj7qj" Dec 11 02:06:47 crc kubenswrapper[4824]: I1211 02:06:47.146967 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/19528072-20f2-4741-bfd9-b7c44430b834-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-xj7qj\" (UID: \"19528072-20f2-4741-bfd9-b7c44430b834\") " pod="openshift-marketplace/marketplace-operator-79b997595-xj7qj" Dec 11 02:06:47 crc kubenswrapper[4824]: I1211 02:06:47.146996 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/19528072-20f2-4741-bfd9-b7c44430b834-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-xj7qj\" (UID: \"19528072-20f2-4741-bfd9-b7c44430b834\") " pod="openshift-marketplace/marketplace-operator-79b997595-xj7qj" Dec 11 02:06:47 crc kubenswrapper[4824]: I1211 02:06:47.148529 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/19528072-20f2-4741-bfd9-b7c44430b834-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-xj7qj\" (UID: \"19528072-20f2-4741-bfd9-b7c44430b834\") " pod="openshift-marketplace/marketplace-operator-79b997595-xj7qj" Dec 11 02:06:47 crc kubenswrapper[4824]: I1211 02:06:47.152797 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/19528072-20f2-4741-bfd9-b7c44430b834-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-xj7qj\" (UID: \"19528072-20f2-4741-bfd9-b7c44430b834\") " pod="openshift-marketplace/marketplace-operator-79b997595-xj7qj" Dec 11 02:06:47 crc kubenswrapper[4824]: I1211 02:06:47.163039 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-msrlm\" (UniqueName: \"kubernetes.io/projected/19528072-20f2-4741-bfd9-b7c44430b834-kube-api-access-msrlm\") pod \"marketplace-operator-79b997595-xj7qj\" (UID: \"19528072-20f2-4741-bfd9-b7c44430b834\") " pod="openshift-marketplace/marketplace-operator-79b997595-xj7qj" Dec 11 02:06:47 crc kubenswrapper[4824]: I1211 02:06:47.172671 4824 generic.go:334] "Generic (PLEG): container finished" podID="ffb4acd4-543f-4509-b993-249e3656048d" containerID="60693fc1fe023dcf8dbae48f9d5d679fb604f2a41ce2a4e0c21987b053e0c4f1" exitCode=0 Dec 11 02:06:47 crc kubenswrapper[4824]: I1211 02:06:47.172738 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-z42g5" event={"ID":"ffb4acd4-543f-4509-b993-249e3656048d","Type":"ContainerDied","Data":"60693fc1fe023dcf8dbae48f9d5d679fb604f2a41ce2a4e0c21987b053e0c4f1"} Dec 11 02:06:47 crc kubenswrapper[4824]: I1211 02:06:47.176484 4824 generic.go:334] "Generic (PLEG): container finished" podID="0b3b02dc-e8f5-4629-b1bf-72febdb40843" containerID="844683dffb7882a5cc33ed4c458f332f8ffa61080cfdae10d2ce9819e505dc34" exitCode=0 Dec 11 02:06:47 crc kubenswrapper[4824]: I1211 02:06:47.176549 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rnmtp" event={"ID":"0b3b02dc-e8f5-4629-b1bf-72febdb40843","Type":"ContainerDied","Data":"844683dffb7882a5cc33ed4c458f332f8ffa61080cfdae10d2ce9819e505dc34"} Dec 11 02:06:47 crc kubenswrapper[4824]: I1211 02:06:47.179384 4824 generic.go:334] "Generic (PLEG): container finished" podID="6127fb96-7f67-453d-8d16-40fac34e6482" containerID="bd6609faa0c263d1acb6b286455bfc86a0fb4b396c42cf8eeb5c1a12469e1ec6" exitCode=0 Dec 11 02:06:47 crc kubenswrapper[4824]: I1211 02:06:47.179437 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jc49g" event={"ID":"6127fb96-7f67-453d-8d16-40fac34e6482","Type":"ContainerDied","Data":"bd6609faa0c263d1acb6b286455bfc86a0fb4b396c42cf8eeb5c1a12469e1ec6"} Dec 11 02:06:47 crc kubenswrapper[4824]: I1211 02:06:47.185760 4824 generic.go:334] "Generic (PLEG): container finished" podID="d9806a3c-cb3d-4b3e-8889-11f8cf49c030" containerID="0f6c6d568fd4ead23e24d85c0ced23037ff04686489fd1752516ab279b1e63b3" exitCode=0 Dec 11 02:06:47 crc kubenswrapper[4824]: I1211 02:06:47.185866 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nqjh5" event={"ID":"d9806a3c-cb3d-4b3e-8889-11f8cf49c030","Type":"ContainerDied","Data":"0f6c6d568fd4ead23e24d85c0ced23037ff04686489fd1752516ab279b1e63b3"} Dec 11 02:06:47 crc kubenswrapper[4824]: I1211 02:06:47.190312 4824 generic.go:334] "Generic (PLEG): container finished" podID="b2b7c06f-4350-41be-b472-c75d21558911" containerID="15910645eb6869cc7816a7bae3d43c849e26e86197c452fe3b6a9e0df7d31b82" exitCode=0 Dec 11 02:06:47 crc kubenswrapper[4824]: I1211 02:06:47.190351 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-86mf9" event={"ID":"b2b7c06f-4350-41be-b472-c75d21558911","Type":"ContainerDied","Data":"15910645eb6869cc7816a7bae3d43c849e26e86197c452fe3b6a9e0df7d31b82"} Dec 11 02:06:47 crc kubenswrapper[4824]: I1211 02:06:47.283427 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-xj7qj" Dec 11 02:06:47 crc kubenswrapper[4824]: I1211 02:06:47.407567 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nqjh5" Dec 11 02:06:47 crc kubenswrapper[4824]: I1211 02:06:47.451256 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4xhtb\" (UniqueName: \"kubernetes.io/projected/d9806a3c-cb3d-4b3e-8889-11f8cf49c030-kube-api-access-4xhtb\") pod \"d9806a3c-cb3d-4b3e-8889-11f8cf49c030\" (UID: \"d9806a3c-cb3d-4b3e-8889-11f8cf49c030\") " Dec 11 02:06:47 crc kubenswrapper[4824]: I1211 02:06:47.451312 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9806a3c-cb3d-4b3e-8889-11f8cf49c030-utilities\") pod \"d9806a3c-cb3d-4b3e-8889-11f8cf49c030\" (UID: \"d9806a3c-cb3d-4b3e-8889-11f8cf49c030\") " Dec 11 02:06:47 crc kubenswrapper[4824]: I1211 02:06:47.451384 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9806a3c-cb3d-4b3e-8889-11f8cf49c030-catalog-content\") pod \"d9806a3c-cb3d-4b3e-8889-11f8cf49c030\" (UID: \"d9806a3c-cb3d-4b3e-8889-11f8cf49c030\") " Dec 11 02:06:47 crc kubenswrapper[4824]: I1211 02:06:47.452540 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d9806a3c-cb3d-4b3e-8889-11f8cf49c030-utilities" (OuterVolumeSpecName: "utilities") pod "d9806a3c-cb3d-4b3e-8889-11f8cf49c030" (UID: "d9806a3c-cb3d-4b3e-8889-11f8cf49c030"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:06:47 crc kubenswrapper[4824]: I1211 02:06:47.457440 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9806a3c-cb3d-4b3e-8889-11f8cf49c030-kube-api-access-4xhtb" (OuterVolumeSpecName: "kube-api-access-4xhtb") pod "d9806a3c-cb3d-4b3e-8889-11f8cf49c030" (UID: "d9806a3c-cb3d-4b3e-8889-11f8cf49c030"). InnerVolumeSpecName "kube-api-access-4xhtb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:06:47 crc kubenswrapper[4824]: I1211 02:06:47.500324 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d9806a3c-cb3d-4b3e-8889-11f8cf49c030-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d9806a3c-cb3d-4b3e-8889-11f8cf49c030" (UID: "d9806a3c-cb3d-4b3e-8889-11f8cf49c030"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:06:47 crc kubenswrapper[4824]: I1211 02:06:47.551852 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9806a3c-cb3d-4b3e-8889-11f8cf49c030-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 02:06:47 crc kubenswrapper[4824]: I1211 02:06:47.551880 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4xhtb\" (UniqueName: \"kubernetes.io/projected/d9806a3c-cb3d-4b3e-8889-11f8cf49c030-kube-api-access-4xhtb\") on node \"crc\" DevicePath \"\"" Dec 11 02:06:47 crc kubenswrapper[4824]: I1211 02:06:47.551891 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9806a3c-cb3d-4b3e-8889-11f8cf49c030-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 02:06:47 crc kubenswrapper[4824]: I1211 02:06:47.552587 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-86mf9" Dec 11 02:06:47 crc kubenswrapper[4824]: I1211 02:06:47.591721 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-z42g5" Dec 11 02:06:47 crc kubenswrapper[4824]: I1211 02:06:47.652466 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2b7c06f-4350-41be-b472-c75d21558911-catalog-content\") pod \"b2b7c06f-4350-41be-b472-c75d21558911\" (UID: \"b2b7c06f-4350-41be-b472-c75d21558911\") " Dec 11 02:06:47 crc kubenswrapper[4824]: I1211 02:06:47.652806 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2gdpn\" (UniqueName: \"kubernetes.io/projected/b2b7c06f-4350-41be-b472-c75d21558911-kube-api-access-2gdpn\") pod \"b2b7c06f-4350-41be-b472-c75d21558911\" (UID: \"b2b7c06f-4350-41be-b472-c75d21558911\") " Dec 11 02:06:47 crc kubenswrapper[4824]: I1211 02:06:47.652832 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ffb4acd4-543f-4509-b993-249e3656048d-marketplace-trusted-ca\") pod \"ffb4acd4-543f-4509-b993-249e3656048d\" (UID: \"ffb4acd4-543f-4509-b993-249e3656048d\") " Dec 11 02:06:47 crc kubenswrapper[4824]: I1211 02:06:47.652867 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kq4zx\" (UniqueName: \"kubernetes.io/projected/ffb4acd4-543f-4509-b993-249e3656048d-kube-api-access-kq4zx\") pod \"ffb4acd4-543f-4509-b993-249e3656048d\" (UID: \"ffb4acd4-543f-4509-b993-249e3656048d\") " Dec 11 02:06:47 crc kubenswrapper[4824]: I1211 02:06:47.652892 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2b7c06f-4350-41be-b472-c75d21558911-utilities\") pod \"b2b7c06f-4350-41be-b472-c75d21558911\" (UID: \"b2b7c06f-4350-41be-b472-c75d21558911\") " Dec 11 02:06:47 crc kubenswrapper[4824]: I1211 02:06:47.652916 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ffb4acd4-543f-4509-b993-249e3656048d-marketplace-operator-metrics\") pod \"ffb4acd4-543f-4509-b993-249e3656048d\" (UID: \"ffb4acd4-543f-4509-b993-249e3656048d\") " Dec 11 02:06:47 crc kubenswrapper[4824]: I1211 02:06:47.654401 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ffb4acd4-543f-4509-b993-249e3656048d-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "ffb4acd4-543f-4509-b993-249e3656048d" (UID: "ffb4acd4-543f-4509-b993-249e3656048d"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:06:47 crc kubenswrapper[4824]: I1211 02:06:47.656658 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ffb4acd4-543f-4509-b993-249e3656048d-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "ffb4acd4-543f-4509-b993-249e3656048d" (UID: "ffb4acd4-543f-4509-b993-249e3656048d"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:06:47 crc kubenswrapper[4824]: I1211 02:06:47.657793 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2b7c06f-4350-41be-b472-c75d21558911-kube-api-access-2gdpn" (OuterVolumeSpecName: "kube-api-access-2gdpn") pod "b2b7c06f-4350-41be-b472-c75d21558911" (UID: "b2b7c06f-4350-41be-b472-c75d21558911"). InnerVolumeSpecName "kube-api-access-2gdpn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:06:47 crc kubenswrapper[4824]: I1211 02:06:47.658688 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b2b7c06f-4350-41be-b472-c75d21558911-utilities" (OuterVolumeSpecName: "utilities") pod "b2b7c06f-4350-41be-b472-c75d21558911" (UID: "b2b7c06f-4350-41be-b472-c75d21558911"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:06:47 crc kubenswrapper[4824]: I1211 02:06:47.661160 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ffb4acd4-543f-4509-b993-249e3656048d-kube-api-access-kq4zx" (OuterVolumeSpecName: "kube-api-access-kq4zx") pod "ffb4acd4-543f-4509-b993-249e3656048d" (UID: "ffb4acd4-543f-4509-b993-249e3656048d"). InnerVolumeSpecName "kube-api-access-kq4zx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:06:47 crc kubenswrapper[4824]: I1211 02:06:47.717270 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b2b7c06f-4350-41be-b472-c75d21558911-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b2b7c06f-4350-41be-b472-c75d21558911" (UID: "b2b7c06f-4350-41be-b472-c75d21558911"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:06:47 crc kubenswrapper[4824]: I1211 02:06:47.753802 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kq4zx\" (UniqueName: \"kubernetes.io/projected/ffb4acd4-543f-4509-b993-249e3656048d-kube-api-access-kq4zx\") on node \"crc\" DevicePath \"\"" Dec 11 02:06:47 crc kubenswrapper[4824]: I1211 02:06:47.753839 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2b7c06f-4350-41be-b472-c75d21558911-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 02:06:47 crc kubenswrapper[4824]: I1211 02:06:47.753851 4824 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ffb4acd4-543f-4509-b993-249e3656048d-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 11 02:06:47 crc kubenswrapper[4824]: I1211 02:06:47.753860 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2b7c06f-4350-41be-b472-c75d21558911-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 02:06:47 crc kubenswrapper[4824]: I1211 02:06:47.753869 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2gdpn\" (UniqueName: \"kubernetes.io/projected/b2b7c06f-4350-41be-b472-c75d21558911-kube-api-access-2gdpn\") on node \"crc\" DevicePath \"\"" Dec 11 02:06:47 crc kubenswrapper[4824]: I1211 02:06:47.753878 4824 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ffb4acd4-543f-4509-b993-249e3656048d-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 11 02:06:47 crc kubenswrapper[4824]: I1211 02:06:47.772627 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jc49g" Dec 11 02:06:47 crc kubenswrapper[4824]: I1211 02:06:47.785778 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rnmtp" Dec 11 02:06:47 crc kubenswrapper[4824]: I1211 02:06:47.850134 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-xj7qj"] Dec 11 02:06:47 crc kubenswrapper[4824]: I1211 02:06:47.854322 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgcxd\" (UniqueName: \"kubernetes.io/projected/0b3b02dc-e8f5-4629-b1bf-72febdb40843-kube-api-access-zgcxd\") pod \"0b3b02dc-e8f5-4629-b1bf-72febdb40843\" (UID: \"0b3b02dc-e8f5-4629-b1bf-72febdb40843\") " Dec 11 02:06:47 crc kubenswrapper[4824]: I1211 02:06:47.854379 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6127fb96-7f67-453d-8d16-40fac34e6482-catalog-content\") pod \"6127fb96-7f67-453d-8d16-40fac34e6482\" (UID: \"6127fb96-7f67-453d-8d16-40fac34e6482\") " Dec 11 02:06:47 crc kubenswrapper[4824]: I1211 02:06:47.854492 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b3b02dc-e8f5-4629-b1bf-72febdb40843-utilities\") pod \"0b3b02dc-e8f5-4629-b1bf-72febdb40843\" (UID: \"0b3b02dc-e8f5-4629-b1bf-72febdb40843\") " Dec 11 02:06:47 crc kubenswrapper[4824]: I1211 02:06:47.854550 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b3b02dc-e8f5-4629-b1bf-72febdb40843-catalog-content\") pod \"0b3b02dc-e8f5-4629-b1bf-72febdb40843\" (UID: \"0b3b02dc-e8f5-4629-b1bf-72febdb40843\") " Dec 11 02:06:47 crc kubenswrapper[4824]: I1211 02:06:47.854586 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-64k6f\" (UniqueName: \"kubernetes.io/projected/6127fb96-7f67-453d-8d16-40fac34e6482-kube-api-access-64k6f\") pod \"6127fb96-7f67-453d-8d16-40fac34e6482\" (UID: \"6127fb96-7f67-453d-8d16-40fac34e6482\") " Dec 11 02:06:47 crc kubenswrapper[4824]: I1211 02:06:47.854611 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6127fb96-7f67-453d-8d16-40fac34e6482-utilities\") pod \"6127fb96-7f67-453d-8d16-40fac34e6482\" (UID: \"6127fb96-7f67-453d-8d16-40fac34e6482\") " Dec 11 02:06:47 crc kubenswrapper[4824]: I1211 02:06:47.855823 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6127fb96-7f67-453d-8d16-40fac34e6482-utilities" (OuterVolumeSpecName: "utilities") pod "6127fb96-7f67-453d-8d16-40fac34e6482" (UID: "6127fb96-7f67-453d-8d16-40fac34e6482"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:06:47 crc kubenswrapper[4824]: I1211 02:06:47.857735 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b3b02dc-e8f5-4629-b1bf-72febdb40843-kube-api-access-zgcxd" (OuterVolumeSpecName: "kube-api-access-zgcxd") pod "0b3b02dc-e8f5-4629-b1bf-72febdb40843" (UID: "0b3b02dc-e8f5-4629-b1bf-72febdb40843"). InnerVolumeSpecName "kube-api-access-zgcxd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:06:47 crc kubenswrapper[4824]: I1211 02:06:47.860523 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6127fb96-7f67-453d-8d16-40fac34e6482-kube-api-access-64k6f" (OuterVolumeSpecName: "kube-api-access-64k6f") pod "6127fb96-7f67-453d-8d16-40fac34e6482" (UID: "6127fb96-7f67-453d-8d16-40fac34e6482"). InnerVolumeSpecName "kube-api-access-64k6f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:06:47 crc kubenswrapper[4824]: I1211 02:06:47.870689 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0b3b02dc-e8f5-4629-b1bf-72febdb40843-utilities" (OuterVolumeSpecName: "utilities") pod "0b3b02dc-e8f5-4629-b1bf-72febdb40843" (UID: "0b3b02dc-e8f5-4629-b1bf-72febdb40843"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:06:47 crc kubenswrapper[4824]: I1211 02:06:47.880500 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6127fb96-7f67-453d-8d16-40fac34e6482-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6127fb96-7f67-453d-8d16-40fac34e6482" (UID: "6127fb96-7f67-453d-8d16-40fac34e6482"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:06:47 crc kubenswrapper[4824]: I1211 02:06:47.955934 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgcxd\" (UniqueName: \"kubernetes.io/projected/0b3b02dc-e8f5-4629-b1bf-72febdb40843-kube-api-access-zgcxd\") on node \"crc\" DevicePath \"\"" Dec 11 02:06:47 crc kubenswrapper[4824]: I1211 02:06:47.955979 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6127fb96-7f67-453d-8d16-40fac34e6482-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 02:06:47 crc kubenswrapper[4824]: I1211 02:06:47.955993 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b3b02dc-e8f5-4629-b1bf-72febdb40843-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 02:06:47 crc kubenswrapper[4824]: I1211 02:06:47.956007 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-64k6f\" (UniqueName: \"kubernetes.io/projected/6127fb96-7f67-453d-8d16-40fac34e6482-kube-api-access-64k6f\") on node \"crc\" DevicePath \"\"" Dec 11 02:06:47 crc kubenswrapper[4824]: I1211 02:06:47.956019 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6127fb96-7f67-453d-8d16-40fac34e6482-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 02:06:47 crc kubenswrapper[4824]: I1211 02:06:47.983525 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0b3b02dc-e8f5-4629-b1bf-72febdb40843-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0b3b02dc-e8f5-4629-b1bf-72febdb40843" (UID: "0b3b02dc-e8f5-4629-b1bf-72febdb40843"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:06:48 crc kubenswrapper[4824]: I1211 02:06:48.057187 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b3b02dc-e8f5-4629-b1bf-72febdb40843-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 02:06:48 crc kubenswrapper[4824]: I1211 02:06:48.196938 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-86mf9" event={"ID":"b2b7c06f-4350-41be-b472-c75d21558911","Type":"ContainerDied","Data":"b5fe6fcc461fc458da49c42d8a8cc74b4e28b5c1a118a0a667331b5d36b1cc04"} Dec 11 02:06:48 crc kubenswrapper[4824]: I1211 02:06:48.197002 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-86mf9" Dec 11 02:06:48 crc kubenswrapper[4824]: I1211 02:06:48.197018 4824 scope.go:117] "RemoveContainer" containerID="15910645eb6869cc7816a7bae3d43c849e26e86197c452fe3b6a9e0df7d31b82" Dec 11 02:06:48 crc kubenswrapper[4824]: I1211 02:06:48.200508 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rnmtp" event={"ID":"0b3b02dc-e8f5-4629-b1bf-72febdb40843","Type":"ContainerDied","Data":"aced56de9a90fc5c8e1f819e7e3e6eb8fa5199de5104caef53f1d721f12fa3f1"} Dec 11 02:06:48 crc kubenswrapper[4824]: I1211 02:06:48.200523 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rnmtp" Dec 11 02:06:48 crc kubenswrapper[4824]: I1211 02:06:48.203810 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-z42g5" Dec 11 02:06:48 crc kubenswrapper[4824]: I1211 02:06:48.204345 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-z42g5" event={"ID":"ffb4acd4-543f-4509-b993-249e3656048d","Type":"ContainerDied","Data":"08bb8882304545419153546c5cbe5f2043cb256198be52a6208a7a5118123a30"} Dec 11 02:06:48 crc kubenswrapper[4824]: I1211 02:06:48.207764 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jc49g" event={"ID":"6127fb96-7f67-453d-8d16-40fac34e6482","Type":"ContainerDied","Data":"f0492d0355db6825fede043db8c18469767b36b313a1a0718163d8e0c2a9baa2"} Dec 11 02:06:48 crc kubenswrapper[4824]: I1211 02:06:48.208173 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jc49g" Dec 11 02:06:48 crc kubenswrapper[4824]: I1211 02:06:48.214074 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nqjh5" Dec 11 02:06:48 crc kubenswrapper[4824]: I1211 02:06:48.215203 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nqjh5" event={"ID":"d9806a3c-cb3d-4b3e-8889-11f8cf49c030","Type":"ContainerDied","Data":"e5c7e4ed64f3273f14c292ac3b3d8ff1470dbb633aa74733de56db3daab5bde2"} Dec 11 02:06:48 crc kubenswrapper[4824]: I1211 02:06:48.215274 4824 scope.go:117] "RemoveContainer" containerID="bdab89470dc1e06cdad086566602a8595a319b4bed0607fc722d7a808b90c2f8" Dec 11 02:06:48 crc kubenswrapper[4824]: I1211 02:06:48.217573 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-xj7qj" event={"ID":"19528072-20f2-4741-bfd9-b7c44430b834","Type":"ContainerStarted","Data":"5f7e2c2da46332e3fd8c4ad1dc55240c907611cf64fe7b4bf1e48d9cb54e8289"} Dec 11 02:06:48 crc kubenswrapper[4824]: I1211 02:06:48.217610 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-xj7qj" event={"ID":"19528072-20f2-4741-bfd9-b7c44430b834","Type":"ContainerStarted","Data":"c2e2d7a3ea68336f2b95de3da16b927f5387cd6949d53e0fe8e349b7a9474b87"} Dec 11 02:06:48 crc kubenswrapper[4824]: I1211 02:06:48.217889 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-xj7qj" Dec 11 02:06:48 crc kubenswrapper[4824]: I1211 02:06:48.219859 4824 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-xj7qj container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.61:8080/healthz\": dial tcp 10.217.0.61:8080: connect: connection refused" start-of-body= Dec 11 02:06:48 crc kubenswrapper[4824]: I1211 02:06:48.219922 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-xj7qj" podUID="19528072-20f2-4741-bfd9-b7c44430b834" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.61:8080/healthz\": dial tcp 10.217.0.61:8080: connect: connection refused" Dec 11 02:06:48 crc kubenswrapper[4824]: I1211 02:06:48.238086 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-86mf9"] Dec 11 02:06:48 crc kubenswrapper[4824]: I1211 02:06:48.249685 4824 scope.go:117] "RemoveContainer" containerID="143b00c17a13c20a70e1b91e408ede84b97addff9037deb82bc4f487dbcc0a69" Dec 11 02:06:48 crc kubenswrapper[4824]: I1211 02:06:48.250650 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-86mf9"] Dec 11 02:06:48 crc kubenswrapper[4824]: I1211 02:06:48.253397 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-xj7qj" podStartSLOduration=2.253374431 podStartE2EDuration="2.253374431s" podCreationTimestamp="2025-12-11 02:06:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:06:48.247212357 +0000 UTC m=+349.936249736" watchObservedRunningTime="2025-12-11 02:06:48.253374431 +0000 UTC m=+349.942411810" Dec 11 02:06:48 crc kubenswrapper[4824]: I1211 02:06:48.266174 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jc49g"] Dec 11 02:06:48 crc kubenswrapper[4824]: I1211 02:06:48.273929 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-jc49g"] Dec 11 02:06:48 crc kubenswrapper[4824]: I1211 02:06:48.278141 4824 scope.go:117] "RemoveContainer" containerID="844683dffb7882a5cc33ed4c458f332f8ffa61080cfdae10d2ce9819e505dc34" Dec 11 02:06:48 crc kubenswrapper[4824]: I1211 02:06:48.279265 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nqjh5"] Dec 11 02:06:48 crc kubenswrapper[4824]: I1211 02:06:48.285503 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-nqjh5"] Dec 11 02:06:48 crc kubenswrapper[4824]: I1211 02:06:48.288537 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-z42g5"] Dec 11 02:06:48 crc kubenswrapper[4824]: I1211 02:06:48.292023 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-z42g5"] Dec 11 02:06:48 crc kubenswrapper[4824]: I1211 02:06:48.299522 4824 scope.go:117] "RemoveContainer" containerID="2b69d5e0865f31b41962f4290865c9d2a83767b4752253b77987606cc86a9bb3" Dec 11 02:06:48 crc kubenswrapper[4824]: I1211 02:06:48.301077 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rnmtp"] Dec 11 02:06:48 crc kubenswrapper[4824]: I1211 02:06:48.304671 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-rnmtp"] Dec 11 02:06:48 crc kubenswrapper[4824]: I1211 02:06:48.317099 4824 scope.go:117] "RemoveContainer" containerID="57a96e8e6e5d272df43e6d59a414c321970a47564eb5dde745abddba9723e046" Dec 11 02:06:48 crc kubenswrapper[4824]: I1211 02:06:48.336659 4824 scope.go:117] "RemoveContainer" containerID="60693fc1fe023dcf8dbae48f9d5d679fb604f2a41ce2a4e0c21987b053e0c4f1" Dec 11 02:06:48 crc kubenswrapper[4824]: I1211 02:06:48.354134 4824 scope.go:117] "RemoveContainer" containerID="bd6609faa0c263d1acb6b286455bfc86a0fb4b396c42cf8eeb5c1a12469e1ec6" Dec 11 02:06:48 crc kubenswrapper[4824]: I1211 02:06:48.372385 4824 scope.go:117] "RemoveContainer" containerID="e284d9b7c962ec7af9fd503606c0088dbf83b7e2646b38c7bbb32a9ecf3b1445" Dec 11 02:06:48 crc kubenswrapper[4824]: I1211 02:06:48.389363 4824 scope.go:117] "RemoveContainer" containerID="7513e3598adb92ec0a08cf216ce6210836909455aaddb7f47d6ab801d14e1588" Dec 11 02:06:48 crc kubenswrapper[4824]: I1211 02:06:48.403938 4824 scope.go:117] "RemoveContainer" containerID="0f6c6d568fd4ead23e24d85c0ced23037ff04686489fd1752516ab279b1e63b3" Dec 11 02:06:48 crc kubenswrapper[4824]: I1211 02:06:48.418636 4824 scope.go:117] "RemoveContainer" containerID="2a640e46ccf1141c27a92fc5e06a0e9ee5df1690b4df836c0188d0f31938d90d" Dec 11 02:06:48 crc kubenswrapper[4824]: I1211 02:06:48.433946 4824 scope.go:117] "RemoveContainer" containerID="84b098c9bd03d479948fdfc8972a1077980e5c30874d36863df78bfcb45acfb1" Dec 11 02:06:48 crc kubenswrapper[4824]: I1211 02:06:48.647256 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b3b02dc-e8f5-4629-b1bf-72febdb40843" path="/var/lib/kubelet/pods/0b3b02dc-e8f5-4629-b1bf-72febdb40843/volumes" Dec 11 02:06:48 crc kubenswrapper[4824]: I1211 02:06:48.648813 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6127fb96-7f67-453d-8d16-40fac34e6482" path="/var/lib/kubelet/pods/6127fb96-7f67-453d-8d16-40fac34e6482/volumes" Dec 11 02:06:48 crc kubenswrapper[4824]: I1211 02:06:48.654894 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2b7c06f-4350-41be-b472-c75d21558911" path="/var/lib/kubelet/pods/b2b7c06f-4350-41be-b472-c75d21558911/volumes" Dec 11 02:06:48 crc kubenswrapper[4824]: I1211 02:06:48.655685 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d9806a3c-cb3d-4b3e-8889-11f8cf49c030" path="/var/lib/kubelet/pods/d9806a3c-cb3d-4b3e-8889-11f8cf49c030/volumes" Dec 11 02:06:48 crc kubenswrapper[4824]: I1211 02:06:48.657080 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ffb4acd4-543f-4509-b993-249e3656048d" path="/var/lib/kubelet/pods/ffb4acd4-543f-4509-b993-249e3656048d/volumes" Dec 11 02:06:49 crc kubenswrapper[4824]: I1211 02:06:49.129138 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-k4msm"] Dec 11 02:06:49 crc kubenswrapper[4824]: E1211 02:06:49.129370 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b3b02dc-e8f5-4629-b1bf-72febdb40843" containerName="extract-utilities" Dec 11 02:06:49 crc kubenswrapper[4824]: I1211 02:06:49.129388 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b3b02dc-e8f5-4629-b1bf-72febdb40843" containerName="extract-utilities" Dec 11 02:06:49 crc kubenswrapper[4824]: E1211 02:06:49.129401 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2b7c06f-4350-41be-b472-c75d21558911" containerName="extract-content" Dec 11 02:06:49 crc kubenswrapper[4824]: I1211 02:06:49.129409 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2b7c06f-4350-41be-b472-c75d21558911" containerName="extract-content" Dec 11 02:06:49 crc kubenswrapper[4824]: E1211 02:06:49.129425 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffb4acd4-543f-4509-b993-249e3656048d" containerName="marketplace-operator" Dec 11 02:06:49 crc kubenswrapper[4824]: I1211 02:06:49.129434 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffb4acd4-543f-4509-b993-249e3656048d" containerName="marketplace-operator" Dec 11 02:06:49 crc kubenswrapper[4824]: E1211 02:06:49.129451 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6127fb96-7f67-453d-8d16-40fac34e6482" containerName="extract-utilities" Dec 11 02:06:49 crc kubenswrapper[4824]: I1211 02:06:49.129459 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="6127fb96-7f67-453d-8d16-40fac34e6482" containerName="extract-utilities" Dec 11 02:06:49 crc kubenswrapper[4824]: E1211 02:06:49.129471 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b3b02dc-e8f5-4629-b1bf-72febdb40843" containerName="registry-server" Dec 11 02:06:49 crc kubenswrapper[4824]: I1211 02:06:49.129478 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b3b02dc-e8f5-4629-b1bf-72febdb40843" containerName="registry-server" Dec 11 02:06:49 crc kubenswrapper[4824]: E1211 02:06:49.129486 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6127fb96-7f67-453d-8d16-40fac34e6482" containerName="registry-server" Dec 11 02:06:49 crc kubenswrapper[4824]: I1211 02:06:49.129494 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="6127fb96-7f67-453d-8d16-40fac34e6482" containerName="registry-server" Dec 11 02:06:49 crc kubenswrapper[4824]: E1211 02:06:49.129504 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2b7c06f-4350-41be-b472-c75d21558911" containerName="extract-utilities" Dec 11 02:06:49 crc kubenswrapper[4824]: I1211 02:06:49.129514 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2b7c06f-4350-41be-b472-c75d21558911" containerName="extract-utilities" Dec 11 02:06:49 crc kubenswrapper[4824]: E1211 02:06:49.129522 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9806a3c-cb3d-4b3e-8889-11f8cf49c030" containerName="extract-content" Dec 11 02:06:49 crc kubenswrapper[4824]: I1211 02:06:49.129529 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9806a3c-cb3d-4b3e-8889-11f8cf49c030" containerName="extract-content" Dec 11 02:06:49 crc kubenswrapper[4824]: E1211 02:06:49.129543 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9806a3c-cb3d-4b3e-8889-11f8cf49c030" containerName="extract-utilities" Dec 11 02:06:49 crc kubenswrapper[4824]: I1211 02:06:49.129550 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9806a3c-cb3d-4b3e-8889-11f8cf49c030" containerName="extract-utilities" Dec 11 02:06:49 crc kubenswrapper[4824]: E1211 02:06:49.129560 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b3b02dc-e8f5-4629-b1bf-72febdb40843" containerName="extract-content" Dec 11 02:06:49 crc kubenswrapper[4824]: I1211 02:06:49.129567 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b3b02dc-e8f5-4629-b1bf-72febdb40843" containerName="extract-content" Dec 11 02:06:49 crc kubenswrapper[4824]: E1211 02:06:49.129578 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2b7c06f-4350-41be-b472-c75d21558911" containerName="registry-server" Dec 11 02:06:49 crc kubenswrapper[4824]: I1211 02:06:49.129589 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2b7c06f-4350-41be-b472-c75d21558911" containerName="registry-server" Dec 11 02:06:49 crc kubenswrapper[4824]: E1211 02:06:49.129600 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6127fb96-7f67-453d-8d16-40fac34e6482" containerName="extract-content" Dec 11 02:06:49 crc kubenswrapper[4824]: I1211 02:06:49.129608 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="6127fb96-7f67-453d-8d16-40fac34e6482" containerName="extract-content" Dec 11 02:06:49 crc kubenswrapper[4824]: E1211 02:06:49.129619 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9806a3c-cb3d-4b3e-8889-11f8cf49c030" containerName="registry-server" Dec 11 02:06:49 crc kubenswrapper[4824]: I1211 02:06:49.129627 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9806a3c-cb3d-4b3e-8889-11f8cf49c030" containerName="registry-server" Dec 11 02:06:49 crc kubenswrapper[4824]: I1211 02:06:49.129737 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="ffb4acd4-543f-4509-b993-249e3656048d" containerName="marketplace-operator" Dec 11 02:06:49 crc kubenswrapper[4824]: I1211 02:06:49.129751 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="6127fb96-7f67-453d-8d16-40fac34e6482" containerName="registry-server" Dec 11 02:06:49 crc kubenswrapper[4824]: I1211 02:06:49.129761 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9806a3c-cb3d-4b3e-8889-11f8cf49c030" containerName="registry-server" Dec 11 02:06:49 crc kubenswrapper[4824]: I1211 02:06:49.129775 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2b7c06f-4350-41be-b472-c75d21558911" containerName="registry-server" Dec 11 02:06:49 crc kubenswrapper[4824]: I1211 02:06:49.129784 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b3b02dc-e8f5-4629-b1bf-72febdb40843" containerName="registry-server" Dec 11 02:06:49 crc kubenswrapper[4824]: I1211 02:06:49.130682 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k4msm" Dec 11 02:06:49 crc kubenswrapper[4824]: I1211 02:06:49.133106 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 11 02:06:49 crc kubenswrapper[4824]: I1211 02:06:49.181096 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-k4msm"] Dec 11 02:06:49 crc kubenswrapper[4824]: I1211 02:06:49.232543 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-xj7qj" Dec 11 02:06:49 crc kubenswrapper[4824]: I1211 02:06:49.278741 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38e53ad7-42f0-4b59-a721-b5e38b6e1904-utilities\") pod \"redhat-marketplace-k4msm\" (UID: \"38e53ad7-42f0-4b59-a721-b5e38b6e1904\") " pod="openshift-marketplace/redhat-marketplace-k4msm" Dec 11 02:06:49 crc kubenswrapper[4824]: I1211 02:06:49.279205 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-226kk\" (UniqueName: \"kubernetes.io/projected/38e53ad7-42f0-4b59-a721-b5e38b6e1904-kube-api-access-226kk\") pod \"redhat-marketplace-k4msm\" (UID: \"38e53ad7-42f0-4b59-a721-b5e38b6e1904\") " pod="openshift-marketplace/redhat-marketplace-k4msm" Dec 11 02:06:49 crc kubenswrapper[4824]: I1211 02:06:49.279568 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38e53ad7-42f0-4b59-a721-b5e38b6e1904-catalog-content\") pod \"redhat-marketplace-k4msm\" (UID: \"38e53ad7-42f0-4b59-a721-b5e38b6e1904\") " pod="openshift-marketplace/redhat-marketplace-k4msm" Dec 11 02:06:49 crc kubenswrapper[4824]: I1211 02:06:49.321519 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-df294"] Dec 11 02:06:49 crc kubenswrapper[4824]: I1211 02:06:49.322765 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-df294" Dec 11 02:06:49 crc kubenswrapper[4824]: I1211 02:06:49.324972 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 11 02:06:49 crc kubenswrapper[4824]: I1211 02:06:49.333338 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-df294"] Dec 11 02:06:49 crc kubenswrapper[4824]: I1211 02:06:49.380281 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-226kk\" (UniqueName: \"kubernetes.io/projected/38e53ad7-42f0-4b59-a721-b5e38b6e1904-kube-api-access-226kk\") pod \"redhat-marketplace-k4msm\" (UID: \"38e53ad7-42f0-4b59-a721-b5e38b6e1904\") " pod="openshift-marketplace/redhat-marketplace-k4msm" Dec 11 02:06:49 crc kubenswrapper[4824]: I1211 02:06:49.380343 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38e53ad7-42f0-4b59-a721-b5e38b6e1904-catalog-content\") pod \"redhat-marketplace-k4msm\" (UID: \"38e53ad7-42f0-4b59-a721-b5e38b6e1904\") " pod="openshift-marketplace/redhat-marketplace-k4msm" Dec 11 02:06:49 crc kubenswrapper[4824]: I1211 02:06:49.380392 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38e53ad7-42f0-4b59-a721-b5e38b6e1904-utilities\") pod \"redhat-marketplace-k4msm\" (UID: \"38e53ad7-42f0-4b59-a721-b5e38b6e1904\") " pod="openshift-marketplace/redhat-marketplace-k4msm" Dec 11 02:06:49 crc kubenswrapper[4824]: I1211 02:06:49.381440 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38e53ad7-42f0-4b59-a721-b5e38b6e1904-utilities\") pod \"redhat-marketplace-k4msm\" (UID: \"38e53ad7-42f0-4b59-a721-b5e38b6e1904\") " pod="openshift-marketplace/redhat-marketplace-k4msm" Dec 11 02:06:49 crc kubenswrapper[4824]: I1211 02:06:49.381465 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38e53ad7-42f0-4b59-a721-b5e38b6e1904-catalog-content\") pod \"redhat-marketplace-k4msm\" (UID: \"38e53ad7-42f0-4b59-a721-b5e38b6e1904\") " pod="openshift-marketplace/redhat-marketplace-k4msm" Dec 11 02:06:49 crc kubenswrapper[4824]: I1211 02:06:49.401553 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-226kk\" (UniqueName: \"kubernetes.io/projected/38e53ad7-42f0-4b59-a721-b5e38b6e1904-kube-api-access-226kk\") pod \"redhat-marketplace-k4msm\" (UID: \"38e53ad7-42f0-4b59-a721-b5e38b6e1904\") " pod="openshift-marketplace/redhat-marketplace-k4msm" Dec 11 02:06:49 crc kubenswrapper[4824]: I1211 02:06:49.481449 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/582641c9-665f-4d61-9374-d50d139762ab-catalog-content\") pod \"redhat-operators-df294\" (UID: \"582641c9-665f-4d61-9374-d50d139762ab\") " pod="openshift-marketplace/redhat-operators-df294" Dec 11 02:06:49 crc kubenswrapper[4824]: I1211 02:06:49.481495 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vw8xj\" (UniqueName: \"kubernetes.io/projected/582641c9-665f-4d61-9374-d50d139762ab-kube-api-access-vw8xj\") pod \"redhat-operators-df294\" (UID: \"582641c9-665f-4d61-9374-d50d139762ab\") " pod="openshift-marketplace/redhat-operators-df294" Dec 11 02:06:49 crc kubenswrapper[4824]: I1211 02:06:49.481530 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/582641c9-665f-4d61-9374-d50d139762ab-utilities\") pod \"redhat-operators-df294\" (UID: \"582641c9-665f-4d61-9374-d50d139762ab\") " pod="openshift-marketplace/redhat-operators-df294" Dec 11 02:06:49 crc kubenswrapper[4824]: I1211 02:06:49.497413 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k4msm" Dec 11 02:06:49 crc kubenswrapper[4824]: I1211 02:06:49.586939 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/582641c9-665f-4d61-9374-d50d139762ab-utilities\") pod \"redhat-operators-df294\" (UID: \"582641c9-665f-4d61-9374-d50d139762ab\") " pod="openshift-marketplace/redhat-operators-df294" Dec 11 02:06:49 crc kubenswrapper[4824]: I1211 02:06:49.587249 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/582641c9-665f-4d61-9374-d50d139762ab-catalog-content\") pod \"redhat-operators-df294\" (UID: \"582641c9-665f-4d61-9374-d50d139762ab\") " pod="openshift-marketplace/redhat-operators-df294" Dec 11 02:06:49 crc kubenswrapper[4824]: I1211 02:06:49.587270 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vw8xj\" (UniqueName: \"kubernetes.io/projected/582641c9-665f-4d61-9374-d50d139762ab-kube-api-access-vw8xj\") pod \"redhat-operators-df294\" (UID: \"582641c9-665f-4d61-9374-d50d139762ab\") " pod="openshift-marketplace/redhat-operators-df294" Dec 11 02:06:49 crc kubenswrapper[4824]: I1211 02:06:49.587430 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/582641c9-665f-4d61-9374-d50d139762ab-utilities\") pod \"redhat-operators-df294\" (UID: \"582641c9-665f-4d61-9374-d50d139762ab\") " pod="openshift-marketplace/redhat-operators-df294" Dec 11 02:06:49 crc kubenswrapper[4824]: I1211 02:06:49.588567 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/582641c9-665f-4d61-9374-d50d139762ab-catalog-content\") pod \"redhat-operators-df294\" (UID: \"582641c9-665f-4d61-9374-d50d139762ab\") " pod="openshift-marketplace/redhat-operators-df294" Dec 11 02:06:49 crc kubenswrapper[4824]: I1211 02:06:49.605735 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vw8xj\" (UniqueName: \"kubernetes.io/projected/582641c9-665f-4d61-9374-d50d139762ab-kube-api-access-vw8xj\") pod \"redhat-operators-df294\" (UID: \"582641c9-665f-4d61-9374-d50d139762ab\") " pod="openshift-marketplace/redhat-operators-df294" Dec 11 02:06:49 crc kubenswrapper[4824]: I1211 02:06:49.643763 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-df294" Dec 11 02:06:49 crc kubenswrapper[4824]: I1211 02:06:49.899580 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-k4msm"] Dec 11 02:06:49 crc kubenswrapper[4824]: W1211 02:06:49.909462 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod38e53ad7_42f0_4b59_a721_b5e38b6e1904.slice/crio-deb566da62e64f9c06d1db17f6961e9a38bc9a1d78f7a43805be8c409456009d WatchSource:0}: Error finding container deb566da62e64f9c06d1db17f6961e9a38bc9a1d78f7a43805be8c409456009d: Status 404 returned error can't find the container with id deb566da62e64f9c06d1db17f6961e9a38bc9a1d78f7a43805be8c409456009d Dec 11 02:06:50 crc kubenswrapper[4824]: I1211 02:06:50.052706 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-df294"] Dec 11 02:06:50 crc kubenswrapper[4824]: W1211 02:06:50.059747 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod582641c9_665f_4d61_9374_d50d139762ab.slice/crio-d59c1b9a1b23fee8e1fca1b79490ab8f0d72576f8d5021cc0f03a041d36b58d1 WatchSource:0}: Error finding container d59c1b9a1b23fee8e1fca1b79490ab8f0d72576f8d5021cc0f03a041d36b58d1: Status 404 returned error can't find the container with id d59c1b9a1b23fee8e1fca1b79490ab8f0d72576f8d5021cc0f03a041d36b58d1 Dec 11 02:06:50 crc kubenswrapper[4824]: I1211 02:06:50.235439 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-df294" event={"ID":"582641c9-665f-4d61-9374-d50d139762ab","Type":"ContainerStarted","Data":"d59c1b9a1b23fee8e1fca1b79490ab8f0d72576f8d5021cc0f03a041d36b58d1"} Dec 11 02:06:50 crc kubenswrapper[4824]: I1211 02:06:50.237151 4824 generic.go:334] "Generic (PLEG): container finished" podID="38e53ad7-42f0-4b59-a721-b5e38b6e1904" containerID="e96d6b4b1ca5416cf6a41cd6cfb9464507974d59949b2b88a62e58eb117e90bd" exitCode=0 Dec 11 02:06:50 crc kubenswrapper[4824]: I1211 02:06:50.238207 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k4msm" event={"ID":"38e53ad7-42f0-4b59-a721-b5e38b6e1904","Type":"ContainerDied","Data":"e96d6b4b1ca5416cf6a41cd6cfb9464507974d59949b2b88a62e58eb117e90bd"} Dec 11 02:06:50 crc kubenswrapper[4824]: I1211 02:06:50.238257 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k4msm" event={"ID":"38e53ad7-42f0-4b59-a721-b5e38b6e1904","Type":"ContainerStarted","Data":"deb566da62e64f9c06d1db17f6961e9a38bc9a1d78f7a43805be8c409456009d"} Dec 11 02:06:51 crc kubenswrapper[4824]: I1211 02:06:51.245068 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k4msm" event={"ID":"38e53ad7-42f0-4b59-a721-b5e38b6e1904","Type":"ContainerStarted","Data":"44c6c5241acfe42345875d3bea631fc95d916fe529a6d24a519518b0335245ae"} Dec 11 02:06:51 crc kubenswrapper[4824]: I1211 02:06:51.246161 4824 generic.go:334] "Generic (PLEG): container finished" podID="582641c9-665f-4d61-9374-d50d139762ab" containerID="40bfe40ea92f2b6e73afe1db9e6c91dc3ffa2799d2c8e505adecb83a2c6deb57" exitCode=0 Dec 11 02:06:51 crc kubenswrapper[4824]: I1211 02:06:51.246208 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-df294" event={"ID":"582641c9-665f-4d61-9374-d50d139762ab","Type":"ContainerDied","Data":"40bfe40ea92f2b6e73afe1db9e6c91dc3ffa2799d2c8e505adecb83a2c6deb57"} Dec 11 02:06:51 crc kubenswrapper[4824]: I1211 02:06:51.735282 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-4qn4g"] Dec 11 02:06:51 crc kubenswrapper[4824]: I1211 02:06:51.739241 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4qn4g" Dec 11 02:06:51 crc kubenswrapper[4824]: I1211 02:06:51.743102 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 11 02:06:51 crc kubenswrapper[4824]: I1211 02:06:51.748926 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4qn4g"] Dec 11 02:06:51 crc kubenswrapper[4824]: I1211 02:06:51.758531 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5e0df10-f888-47f3-87a9-889529a89473-catalog-content\") pod \"community-operators-4qn4g\" (UID: \"d5e0df10-f888-47f3-87a9-889529a89473\") " pod="openshift-marketplace/community-operators-4qn4g" Dec 11 02:06:51 crc kubenswrapper[4824]: I1211 02:06:51.758640 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2hfhp\" (UniqueName: \"kubernetes.io/projected/d5e0df10-f888-47f3-87a9-889529a89473-kube-api-access-2hfhp\") pod \"community-operators-4qn4g\" (UID: \"d5e0df10-f888-47f3-87a9-889529a89473\") " pod="openshift-marketplace/community-operators-4qn4g" Dec 11 02:06:51 crc kubenswrapper[4824]: I1211 02:06:51.758664 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5e0df10-f888-47f3-87a9-889529a89473-utilities\") pod \"community-operators-4qn4g\" (UID: \"d5e0df10-f888-47f3-87a9-889529a89473\") " pod="openshift-marketplace/community-operators-4qn4g" Dec 11 02:06:51 crc kubenswrapper[4824]: I1211 02:06:51.859737 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5e0df10-f888-47f3-87a9-889529a89473-catalog-content\") pod \"community-operators-4qn4g\" (UID: \"d5e0df10-f888-47f3-87a9-889529a89473\") " pod="openshift-marketplace/community-operators-4qn4g" Dec 11 02:06:51 crc kubenswrapper[4824]: I1211 02:06:51.859812 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2hfhp\" (UniqueName: \"kubernetes.io/projected/d5e0df10-f888-47f3-87a9-889529a89473-kube-api-access-2hfhp\") pod \"community-operators-4qn4g\" (UID: \"d5e0df10-f888-47f3-87a9-889529a89473\") " pod="openshift-marketplace/community-operators-4qn4g" Dec 11 02:06:51 crc kubenswrapper[4824]: I1211 02:06:51.859830 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5e0df10-f888-47f3-87a9-889529a89473-utilities\") pod \"community-operators-4qn4g\" (UID: \"d5e0df10-f888-47f3-87a9-889529a89473\") " pod="openshift-marketplace/community-operators-4qn4g" Dec 11 02:06:51 crc kubenswrapper[4824]: I1211 02:06:51.860338 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5e0df10-f888-47f3-87a9-889529a89473-utilities\") pod \"community-operators-4qn4g\" (UID: \"d5e0df10-f888-47f3-87a9-889529a89473\") " pod="openshift-marketplace/community-operators-4qn4g" Dec 11 02:06:51 crc kubenswrapper[4824]: I1211 02:06:51.860797 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5e0df10-f888-47f3-87a9-889529a89473-catalog-content\") pod \"community-operators-4qn4g\" (UID: \"d5e0df10-f888-47f3-87a9-889529a89473\") " pod="openshift-marketplace/community-operators-4qn4g" Dec 11 02:06:51 crc kubenswrapper[4824]: I1211 02:06:51.887387 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2hfhp\" (UniqueName: \"kubernetes.io/projected/d5e0df10-f888-47f3-87a9-889529a89473-kube-api-access-2hfhp\") pod \"community-operators-4qn4g\" (UID: \"d5e0df10-f888-47f3-87a9-889529a89473\") " pod="openshift-marketplace/community-operators-4qn4g" Dec 11 02:06:51 crc kubenswrapper[4824]: I1211 02:06:51.927655 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-7dwps"] Dec 11 02:06:51 crc kubenswrapper[4824]: I1211 02:06:51.929002 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7dwps" Dec 11 02:06:51 crc kubenswrapper[4824]: I1211 02:06:51.934582 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 11 02:06:51 crc kubenswrapper[4824]: I1211 02:06:51.960655 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/beb14a1d-6a61-4787-8e2b-ee44942e2828-utilities\") pod \"certified-operators-7dwps\" (UID: \"beb14a1d-6a61-4787-8e2b-ee44942e2828\") " pod="openshift-marketplace/certified-operators-7dwps" Dec 11 02:06:51 crc kubenswrapper[4824]: I1211 02:06:51.960723 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jnfch\" (UniqueName: \"kubernetes.io/projected/beb14a1d-6a61-4787-8e2b-ee44942e2828-kube-api-access-jnfch\") pod \"certified-operators-7dwps\" (UID: \"beb14a1d-6a61-4787-8e2b-ee44942e2828\") " pod="openshift-marketplace/certified-operators-7dwps" Dec 11 02:06:51 crc kubenswrapper[4824]: I1211 02:06:51.960862 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/beb14a1d-6a61-4787-8e2b-ee44942e2828-catalog-content\") pod \"certified-operators-7dwps\" (UID: \"beb14a1d-6a61-4787-8e2b-ee44942e2828\") " pod="openshift-marketplace/certified-operators-7dwps" Dec 11 02:06:51 crc kubenswrapper[4824]: I1211 02:06:51.979150 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7dwps"] Dec 11 02:06:52 crc kubenswrapper[4824]: I1211 02:06:52.061763 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/beb14a1d-6a61-4787-8e2b-ee44942e2828-catalog-content\") pod \"certified-operators-7dwps\" (UID: \"beb14a1d-6a61-4787-8e2b-ee44942e2828\") " pod="openshift-marketplace/certified-operators-7dwps" Dec 11 02:06:52 crc kubenswrapper[4824]: I1211 02:06:52.061855 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/beb14a1d-6a61-4787-8e2b-ee44942e2828-utilities\") pod \"certified-operators-7dwps\" (UID: \"beb14a1d-6a61-4787-8e2b-ee44942e2828\") " pod="openshift-marketplace/certified-operators-7dwps" Dec 11 02:06:52 crc kubenswrapper[4824]: I1211 02:06:52.061883 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jnfch\" (UniqueName: \"kubernetes.io/projected/beb14a1d-6a61-4787-8e2b-ee44942e2828-kube-api-access-jnfch\") pod \"certified-operators-7dwps\" (UID: \"beb14a1d-6a61-4787-8e2b-ee44942e2828\") " pod="openshift-marketplace/certified-operators-7dwps" Dec 11 02:06:52 crc kubenswrapper[4824]: I1211 02:06:52.062167 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/beb14a1d-6a61-4787-8e2b-ee44942e2828-catalog-content\") pod \"certified-operators-7dwps\" (UID: \"beb14a1d-6a61-4787-8e2b-ee44942e2828\") " pod="openshift-marketplace/certified-operators-7dwps" Dec 11 02:06:52 crc kubenswrapper[4824]: I1211 02:06:52.062472 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/beb14a1d-6a61-4787-8e2b-ee44942e2828-utilities\") pod \"certified-operators-7dwps\" (UID: \"beb14a1d-6a61-4787-8e2b-ee44942e2828\") " pod="openshift-marketplace/certified-operators-7dwps" Dec 11 02:06:52 crc kubenswrapper[4824]: I1211 02:06:52.073632 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4qn4g" Dec 11 02:06:52 crc kubenswrapper[4824]: I1211 02:06:52.096094 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jnfch\" (UniqueName: \"kubernetes.io/projected/beb14a1d-6a61-4787-8e2b-ee44942e2828-kube-api-access-jnfch\") pod \"certified-operators-7dwps\" (UID: \"beb14a1d-6a61-4787-8e2b-ee44942e2828\") " pod="openshift-marketplace/certified-operators-7dwps" Dec 11 02:06:52 crc kubenswrapper[4824]: I1211 02:06:52.256719 4824 generic.go:334] "Generic (PLEG): container finished" podID="38e53ad7-42f0-4b59-a721-b5e38b6e1904" containerID="44c6c5241acfe42345875d3bea631fc95d916fe529a6d24a519518b0335245ae" exitCode=0 Dec 11 02:06:52 crc kubenswrapper[4824]: I1211 02:06:52.257020 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k4msm" event={"ID":"38e53ad7-42f0-4b59-a721-b5e38b6e1904","Type":"ContainerDied","Data":"44c6c5241acfe42345875d3bea631fc95d916fe529a6d24a519518b0335245ae"} Dec 11 02:06:52 crc kubenswrapper[4824]: I1211 02:06:52.260380 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-df294" event={"ID":"582641c9-665f-4d61-9374-d50d139762ab","Type":"ContainerStarted","Data":"974bc6bccabbf3e088c5ae9fbfd3f06cc72e3d912dcccb4739b8ab20633e3ad1"} Dec 11 02:06:52 crc kubenswrapper[4824]: I1211 02:06:52.291634 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7dwps" Dec 11 02:06:52 crc kubenswrapper[4824]: W1211 02:06:52.484837 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd5e0df10_f888_47f3_87a9_889529a89473.slice/crio-27ff133ab0f020d044d44f83ec83e2d32e77f0102e9c2d0c294a2afc396e734f WatchSource:0}: Error finding container 27ff133ab0f020d044d44f83ec83e2d32e77f0102e9c2d0c294a2afc396e734f: Status 404 returned error can't find the container with id 27ff133ab0f020d044d44f83ec83e2d32e77f0102e9c2d0c294a2afc396e734f Dec 11 02:06:52 crc kubenswrapper[4824]: I1211 02:06:52.486854 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4qn4g"] Dec 11 02:06:52 crc kubenswrapper[4824]: I1211 02:06:52.694057 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7dwps"] Dec 11 02:06:52 crc kubenswrapper[4824]: W1211 02:06:52.713732 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbeb14a1d_6a61_4787_8e2b_ee44942e2828.slice/crio-2fad8d7713b6adeea66bfeb0cb53bd982b39af5afb63a89207bec2af30f6a337 WatchSource:0}: Error finding container 2fad8d7713b6adeea66bfeb0cb53bd982b39af5afb63a89207bec2af30f6a337: Status 404 returned error can't find the container with id 2fad8d7713b6adeea66bfeb0cb53bd982b39af5afb63a89207bec2af30f6a337 Dec 11 02:06:52 crc kubenswrapper[4824]: E1211 02:06:52.903971 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod582641c9_665f_4d61_9374_d50d139762ab.slice/crio-974bc6bccabbf3e088c5ae9fbfd3f06cc72e3d912dcccb4739b8ab20633e3ad1.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbeb14a1d_6a61_4787_8e2b_ee44942e2828.slice/crio-conmon-4382dff63b8a8c27a32d6fa7e0519472b18363149789d5febd67109c44d6c190.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod582641c9_665f_4d61_9374_d50d139762ab.slice/crio-conmon-974bc6bccabbf3e088c5ae9fbfd3f06cc72e3d912dcccb4739b8ab20633e3ad1.scope\": RecentStats: unable to find data in memory cache]" Dec 11 02:06:53 crc kubenswrapper[4824]: I1211 02:06:53.266605 4824 generic.go:334] "Generic (PLEG): container finished" podID="beb14a1d-6a61-4787-8e2b-ee44942e2828" containerID="4382dff63b8a8c27a32d6fa7e0519472b18363149789d5febd67109c44d6c190" exitCode=0 Dec 11 02:06:53 crc kubenswrapper[4824]: I1211 02:06:53.266725 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7dwps" event={"ID":"beb14a1d-6a61-4787-8e2b-ee44942e2828","Type":"ContainerDied","Data":"4382dff63b8a8c27a32d6fa7e0519472b18363149789d5febd67109c44d6c190"} Dec 11 02:06:53 crc kubenswrapper[4824]: I1211 02:06:53.267003 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7dwps" event={"ID":"beb14a1d-6a61-4787-8e2b-ee44942e2828","Type":"ContainerStarted","Data":"2fad8d7713b6adeea66bfeb0cb53bd982b39af5afb63a89207bec2af30f6a337"} Dec 11 02:06:53 crc kubenswrapper[4824]: I1211 02:06:53.269049 4824 generic.go:334] "Generic (PLEG): container finished" podID="582641c9-665f-4d61-9374-d50d139762ab" containerID="974bc6bccabbf3e088c5ae9fbfd3f06cc72e3d912dcccb4739b8ab20633e3ad1" exitCode=0 Dec 11 02:06:53 crc kubenswrapper[4824]: I1211 02:06:53.269163 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-df294" event={"ID":"582641c9-665f-4d61-9374-d50d139762ab","Type":"ContainerDied","Data":"974bc6bccabbf3e088c5ae9fbfd3f06cc72e3d912dcccb4739b8ab20633e3ad1"} Dec 11 02:06:53 crc kubenswrapper[4824]: I1211 02:06:53.271748 4824 generic.go:334] "Generic (PLEG): container finished" podID="d5e0df10-f888-47f3-87a9-889529a89473" containerID="83059c1c13609cbb389b02eba3e252fb6689e4b6f4434749385fd9943a685ebf" exitCode=0 Dec 11 02:06:53 crc kubenswrapper[4824]: I1211 02:06:53.271803 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4qn4g" event={"ID":"d5e0df10-f888-47f3-87a9-889529a89473","Type":"ContainerDied","Data":"83059c1c13609cbb389b02eba3e252fb6689e4b6f4434749385fd9943a685ebf"} Dec 11 02:06:53 crc kubenswrapper[4824]: I1211 02:06:53.271820 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4qn4g" event={"ID":"d5e0df10-f888-47f3-87a9-889529a89473","Type":"ContainerStarted","Data":"27ff133ab0f020d044d44f83ec83e2d32e77f0102e9c2d0c294a2afc396e734f"} Dec 11 02:06:53 crc kubenswrapper[4824]: I1211 02:06:53.274269 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k4msm" event={"ID":"38e53ad7-42f0-4b59-a721-b5e38b6e1904","Type":"ContainerStarted","Data":"81c9f0b9a3d2003a93954ed2b1edc894cc0d1e8e6f7856d70a711daa8d8c5cfd"} Dec 11 02:06:53 crc kubenswrapper[4824]: I1211 02:06:53.312686 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-k4msm" podStartSLOduration=1.689931742 podStartE2EDuration="4.31266377s" podCreationTimestamp="2025-12-11 02:06:49 +0000 UTC" firstStartedPulling="2025-12-11 02:06:50.239305423 +0000 UTC m=+351.928342802" lastFinishedPulling="2025-12-11 02:06:52.862037451 +0000 UTC m=+354.551074830" observedRunningTime="2025-12-11 02:06:53.308520193 +0000 UTC m=+354.997557602" watchObservedRunningTime="2025-12-11 02:06:53.31266377 +0000 UTC m=+355.001701179" Dec 11 02:06:54 crc kubenswrapper[4824]: I1211 02:06:54.280728 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-df294" event={"ID":"582641c9-665f-4d61-9374-d50d139762ab","Type":"ContainerStarted","Data":"b49fa7d611d18c7a26fada0e82cea44d92b8ccecf9ca8b3a58290dfcb4d65125"} Dec 11 02:06:54 crc kubenswrapper[4824]: I1211 02:06:54.283210 4824 generic.go:334] "Generic (PLEG): container finished" podID="d5e0df10-f888-47f3-87a9-889529a89473" containerID="8afbe339f9aee1fee98e6b4c0d8e4ab7b89fe8426387c974a7ac48a27ce57a79" exitCode=0 Dec 11 02:06:54 crc kubenswrapper[4824]: I1211 02:06:54.283570 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4qn4g" event={"ID":"d5e0df10-f888-47f3-87a9-889529a89473","Type":"ContainerDied","Data":"8afbe339f9aee1fee98e6b4c0d8e4ab7b89fe8426387c974a7ac48a27ce57a79"} Dec 11 02:06:54 crc kubenswrapper[4824]: I1211 02:06:54.290582 4824 generic.go:334] "Generic (PLEG): container finished" podID="beb14a1d-6a61-4787-8e2b-ee44942e2828" containerID="08ed96f115420fde53df358887000e830716bd70e42ff1a54148fe8dbf8439b0" exitCode=0 Dec 11 02:06:54 crc kubenswrapper[4824]: I1211 02:06:54.290688 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7dwps" event={"ID":"beb14a1d-6a61-4787-8e2b-ee44942e2828","Type":"ContainerDied","Data":"08ed96f115420fde53df358887000e830716bd70e42ff1a54148fe8dbf8439b0"} Dec 11 02:06:54 crc kubenswrapper[4824]: I1211 02:06:54.302763 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-df294" podStartSLOduration=2.721532794 podStartE2EDuration="5.302744185s" podCreationTimestamp="2025-12-11 02:06:49 +0000 UTC" firstStartedPulling="2025-12-11 02:06:51.248552107 +0000 UTC m=+352.937589486" lastFinishedPulling="2025-12-11 02:06:53.829763498 +0000 UTC m=+355.518800877" observedRunningTime="2025-12-11 02:06:54.299560586 +0000 UTC m=+355.988597965" watchObservedRunningTime="2025-12-11 02:06:54.302744185 +0000 UTC m=+355.991781564" Dec 11 02:06:55 crc kubenswrapper[4824]: I1211 02:06:55.317237 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4qn4g" event={"ID":"d5e0df10-f888-47f3-87a9-889529a89473","Type":"ContainerStarted","Data":"0947f359dd667b5225082e88ad254dfc05363007f9113524f9fe7c4848f47b83"} Dec 11 02:06:55 crc kubenswrapper[4824]: I1211 02:06:55.336970 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-4qn4g" podStartSLOduration=2.927422206 podStartE2EDuration="4.336952132s" podCreationTimestamp="2025-12-11 02:06:51 +0000 UTC" firstStartedPulling="2025-12-11 02:06:53.272951672 +0000 UTC m=+354.961989081" lastFinishedPulling="2025-12-11 02:06:54.682481628 +0000 UTC m=+356.371519007" observedRunningTime="2025-12-11 02:06:55.333205117 +0000 UTC m=+357.022242526" watchObservedRunningTime="2025-12-11 02:06:55.336952132 +0000 UTC m=+357.025989511" Dec 11 02:06:57 crc kubenswrapper[4824]: I1211 02:06:57.329990 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7dwps" event={"ID":"beb14a1d-6a61-4787-8e2b-ee44942e2828","Type":"ContainerStarted","Data":"ca34783218eccde1b6d0a57e038f1523378c0e97d1e5c3b27664aeed3696dac2"} Dec 11 02:06:57 crc kubenswrapper[4824]: I1211 02:06:57.362375 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-7dwps" podStartSLOduration=4.541839036 podStartE2EDuration="6.362352835s" podCreationTimestamp="2025-12-11 02:06:51 +0000 UTC" firstStartedPulling="2025-12-11 02:06:53.267977082 +0000 UTC m=+354.957014461" lastFinishedPulling="2025-12-11 02:06:55.088490881 +0000 UTC m=+356.777528260" observedRunningTime="2025-12-11 02:06:57.358071734 +0000 UTC m=+359.047109113" watchObservedRunningTime="2025-12-11 02:06:57.362352835 +0000 UTC m=+359.051390254" Dec 11 02:06:59 crc kubenswrapper[4824]: I1211 02:06:59.497720 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-k4msm" Dec 11 02:06:59 crc kubenswrapper[4824]: I1211 02:06:59.498063 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-k4msm" Dec 11 02:06:59 crc kubenswrapper[4824]: I1211 02:06:59.552165 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-k4msm" Dec 11 02:06:59 crc kubenswrapper[4824]: I1211 02:06:59.644778 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-df294" Dec 11 02:06:59 crc kubenswrapper[4824]: I1211 02:06:59.644843 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-df294" Dec 11 02:07:00 crc kubenswrapper[4824]: I1211 02:07:00.052225 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-4gv5b"] Dec 11 02:07:00 crc kubenswrapper[4824]: I1211 02:07:00.053052 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-4gv5b" Dec 11 02:07:00 crc kubenswrapper[4824]: I1211 02:07:00.067605 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-4gv5b"] Dec 11 02:07:00 crc kubenswrapper[4824]: I1211 02:07:00.170714 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/3b68a89b-4e99-4b08-8bf7-6b0987cff1df-registry-tls\") pod \"image-registry-66df7c8f76-4gv5b\" (UID: \"3b68a89b-4e99-4b08-8bf7-6b0987cff1df\") " pod="openshift-image-registry/image-registry-66df7c8f76-4gv5b" Dec 11 02:07:00 crc kubenswrapper[4824]: I1211 02:07:00.170780 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-4gv5b\" (UID: \"3b68a89b-4e99-4b08-8bf7-6b0987cff1df\") " pod="openshift-image-registry/image-registry-66df7c8f76-4gv5b" Dec 11 02:07:00 crc kubenswrapper[4824]: I1211 02:07:00.170801 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-th8zs\" (UniqueName: \"kubernetes.io/projected/3b68a89b-4e99-4b08-8bf7-6b0987cff1df-kube-api-access-th8zs\") pod \"image-registry-66df7c8f76-4gv5b\" (UID: \"3b68a89b-4e99-4b08-8bf7-6b0987cff1df\") " pod="openshift-image-registry/image-registry-66df7c8f76-4gv5b" Dec 11 02:07:00 crc kubenswrapper[4824]: I1211 02:07:00.170820 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/3b68a89b-4e99-4b08-8bf7-6b0987cff1df-installation-pull-secrets\") pod \"image-registry-66df7c8f76-4gv5b\" (UID: \"3b68a89b-4e99-4b08-8bf7-6b0987cff1df\") " pod="openshift-image-registry/image-registry-66df7c8f76-4gv5b" Dec 11 02:07:00 crc kubenswrapper[4824]: I1211 02:07:00.170841 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3b68a89b-4e99-4b08-8bf7-6b0987cff1df-trusted-ca\") pod \"image-registry-66df7c8f76-4gv5b\" (UID: \"3b68a89b-4e99-4b08-8bf7-6b0987cff1df\") " pod="openshift-image-registry/image-registry-66df7c8f76-4gv5b" Dec 11 02:07:00 crc kubenswrapper[4824]: I1211 02:07:00.170860 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3b68a89b-4e99-4b08-8bf7-6b0987cff1df-bound-sa-token\") pod \"image-registry-66df7c8f76-4gv5b\" (UID: \"3b68a89b-4e99-4b08-8bf7-6b0987cff1df\") " pod="openshift-image-registry/image-registry-66df7c8f76-4gv5b" Dec 11 02:07:00 crc kubenswrapper[4824]: I1211 02:07:00.171145 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/3b68a89b-4e99-4b08-8bf7-6b0987cff1df-ca-trust-extracted\") pod \"image-registry-66df7c8f76-4gv5b\" (UID: \"3b68a89b-4e99-4b08-8bf7-6b0987cff1df\") " pod="openshift-image-registry/image-registry-66df7c8f76-4gv5b" Dec 11 02:07:00 crc kubenswrapper[4824]: I1211 02:07:00.171277 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/3b68a89b-4e99-4b08-8bf7-6b0987cff1df-registry-certificates\") pod \"image-registry-66df7c8f76-4gv5b\" (UID: \"3b68a89b-4e99-4b08-8bf7-6b0987cff1df\") " pod="openshift-image-registry/image-registry-66df7c8f76-4gv5b" Dec 11 02:07:00 crc kubenswrapper[4824]: I1211 02:07:00.204706 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-4gv5b\" (UID: \"3b68a89b-4e99-4b08-8bf7-6b0987cff1df\") " pod="openshift-image-registry/image-registry-66df7c8f76-4gv5b" Dec 11 02:07:00 crc kubenswrapper[4824]: I1211 02:07:00.272646 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/3b68a89b-4e99-4b08-8bf7-6b0987cff1df-ca-trust-extracted\") pod \"image-registry-66df7c8f76-4gv5b\" (UID: \"3b68a89b-4e99-4b08-8bf7-6b0987cff1df\") " pod="openshift-image-registry/image-registry-66df7c8f76-4gv5b" Dec 11 02:07:00 crc kubenswrapper[4824]: I1211 02:07:00.272719 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/3b68a89b-4e99-4b08-8bf7-6b0987cff1df-registry-certificates\") pod \"image-registry-66df7c8f76-4gv5b\" (UID: \"3b68a89b-4e99-4b08-8bf7-6b0987cff1df\") " pod="openshift-image-registry/image-registry-66df7c8f76-4gv5b" Dec 11 02:07:00 crc kubenswrapper[4824]: I1211 02:07:00.272762 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/3b68a89b-4e99-4b08-8bf7-6b0987cff1df-registry-tls\") pod \"image-registry-66df7c8f76-4gv5b\" (UID: \"3b68a89b-4e99-4b08-8bf7-6b0987cff1df\") " pod="openshift-image-registry/image-registry-66df7c8f76-4gv5b" Dec 11 02:07:00 crc kubenswrapper[4824]: I1211 02:07:00.272789 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-th8zs\" (UniqueName: \"kubernetes.io/projected/3b68a89b-4e99-4b08-8bf7-6b0987cff1df-kube-api-access-th8zs\") pod \"image-registry-66df7c8f76-4gv5b\" (UID: \"3b68a89b-4e99-4b08-8bf7-6b0987cff1df\") " pod="openshift-image-registry/image-registry-66df7c8f76-4gv5b" Dec 11 02:07:00 crc kubenswrapper[4824]: I1211 02:07:00.272814 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/3b68a89b-4e99-4b08-8bf7-6b0987cff1df-installation-pull-secrets\") pod \"image-registry-66df7c8f76-4gv5b\" (UID: \"3b68a89b-4e99-4b08-8bf7-6b0987cff1df\") " pod="openshift-image-registry/image-registry-66df7c8f76-4gv5b" Dec 11 02:07:00 crc kubenswrapper[4824]: I1211 02:07:00.272838 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3b68a89b-4e99-4b08-8bf7-6b0987cff1df-trusted-ca\") pod \"image-registry-66df7c8f76-4gv5b\" (UID: \"3b68a89b-4e99-4b08-8bf7-6b0987cff1df\") " pod="openshift-image-registry/image-registry-66df7c8f76-4gv5b" Dec 11 02:07:00 crc kubenswrapper[4824]: I1211 02:07:00.272858 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3b68a89b-4e99-4b08-8bf7-6b0987cff1df-bound-sa-token\") pod \"image-registry-66df7c8f76-4gv5b\" (UID: \"3b68a89b-4e99-4b08-8bf7-6b0987cff1df\") " pod="openshift-image-registry/image-registry-66df7c8f76-4gv5b" Dec 11 02:07:00 crc kubenswrapper[4824]: I1211 02:07:00.276184 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/3b68a89b-4e99-4b08-8bf7-6b0987cff1df-ca-trust-extracted\") pod \"image-registry-66df7c8f76-4gv5b\" (UID: \"3b68a89b-4e99-4b08-8bf7-6b0987cff1df\") " pod="openshift-image-registry/image-registry-66df7c8f76-4gv5b" Dec 11 02:07:00 crc kubenswrapper[4824]: I1211 02:07:00.276904 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/3b68a89b-4e99-4b08-8bf7-6b0987cff1df-registry-certificates\") pod \"image-registry-66df7c8f76-4gv5b\" (UID: \"3b68a89b-4e99-4b08-8bf7-6b0987cff1df\") " pod="openshift-image-registry/image-registry-66df7c8f76-4gv5b" Dec 11 02:07:00 crc kubenswrapper[4824]: I1211 02:07:00.277524 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3b68a89b-4e99-4b08-8bf7-6b0987cff1df-trusted-ca\") pod \"image-registry-66df7c8f76-4gv5b\" (UID: \"3b68a89b-4e99-4b08-8bf7-6b0987cff1df\") " pod="openshift-image-registry/image-registry-66df7c8f76-4gv5b" Dec 11 02:07:00 crc kubenswrapper[4824]: I1211 02:07:00.283019 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/3b68a89b-4e99-4b08-8bf7-6b0987cff1df-registry-tls\") pod \"image-registry-66df7c8f76-4gv5b\" (UID: \"3b68a89b-4e99-4b08-8bf7-6b0987cff1df\") " pod="openshift-image-registry/image-registry-66df7c8f76-4gv5b" Dec 11 02:07:00 crc kubenswrapper[4824]: I1211 02:07:00.285940 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/3b68a89b-4e99-4b08-8bf7-6b0987cff1df-installation-pull-secrets\") pod \"image-registry-66df7c8f76-4gv5b\" (UID: \"3b68a89b-4e99-4b08-8bf7-6b0987cff1df\") " pod="openshift-image-registry/image-registry-66df7c8f76-4gv5b" Dec 11 02:07:00 crc kubenswrapper[4824]: I1211 02:07:00.290426 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3b68a89b-4e99-4b08-8bf7-6b0987cff1df-bound-sa-token\") pod \"image-registry-66df7c8f76-4gv5b\" (UID: \"3b68a89b-4e99-4b08-8bf7-6b0987cff1df\") " pod="openshift-image-registry/image-registry-66df7c8f76-4gv5b" Dec 11 02:07:00 crc kubenswrapper[4824]: I1211 02:07:00.292613 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-th8zs\" (UniqueName: \"kubernetes.io/projected/3b68a89b-4e99-4b08-8bf7-6b0987cff1df-kube-api-access-th8zs\") pod \"image-registry-66df7c8f76-4gv5b\" (UID: \"3b68a89b-4e99-4b08-8bf7-6b0987cff1df\") " pod="openshift-image-registry/image-registry-66df7c8f76-4gv5b" Dec 11 02:07:00 crc kubenswrapper[4824]: I1211 02:07:00.389262 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-4gv5b" Dec 11 02:07:00 crc kubenswrapper[4824]: I1211 02:07:00.418402 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-k4msm" Dec 11 02:07:00 crc kubenswrapper[4824]: I1211 02:07:00.700871 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-df294" podUID="582641c9-665f-4d61-9374-d50d139762ab" containerName="registry-server" probeResult="failure" output=< Dec 11 02:07:00 crc kubenswrapper[4824]: timeout: failed to connect service ":50051" within 1s Dec 11 02:07:00 crc kubenswrapper[4824]: > Dec 11 02:07:00 crc kubenswrapper[4824]: I1211 02:07:00.809994 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-4gv5b"] Dec 11 02:07:00 crc kubenswrapper[4824]: W1211 02:07:00.816801 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b68a89b_4e99_4b08_8bf7_6b0987cff1df.slice/crio-9c521ddcf63b1144e4d251c12474bd961af6f6e39a7e1048edec198d3d2628f8 WatchSource:0}: Error finding container 9c521ddcf63b1144e4d251c12474bd961af6f6e39a7e1048edec198d3d2628f8: Status 404 returned error can't find the container with id 9c521ddcf63b1144e4d251c12474bd961af6f6e39a7e1048edec198d3d2628f8 Dec 11 02:07:01 crc kubenswrapper[4824]: I1211 02:07:01.355744 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-4gv5b" event={"ID":"3b68a89b-4e99-4b08-8bf7-6b0987cff1df","Type":"ContainerStarted","Data":"23dcfebaa5f27e697e4db967620a1916b14e86bb03e0a905bb20ff720f476558"} Dec 11 02:07:01 crc kubenswrapper[4824]: I1211 02:07:01.355797 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-4gv5b" event={"ID":"3b68a89b-4e99-4b08-8bf7-6b0987cff1df","Type":"ContainerStarted","Data":"9c521ddcf63b1144e4d251c12474bd961af6f6e39a7e1048edec198d3d2628f8"} Dec 11 02:07:01 crc kubenswrapper[4824]: I1211 02:07:01.373637 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-4gv5b" podStartSLOduration=1.373601108 podStartE2EDuration="1.373601108s" podCreationTimestamp="2025-12-11 02:07:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:07:01.372664491 +0000 UTC m=+363.061701900" watchObservedRunningTime="2025-12-11 02:07:01.373601108 +0000 UTC m=+363.062638487" Dec 11 02:07:02 crc kubenswrapper[4824]: I1211 02:07:02.074235 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-4qn4g" Dec 11 02:07:02 crc kubenswrapper[4824]: I1211 02:07:02.074289 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-4qn4g" Dec 11 02:07:02 crc kubenswrapper[4824]: I1211 02:07:02.140565 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-4qn4g" Dec 11 02:07:02 crc kubenswrapper[4824]: I1211 02:07:02.292576 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-7dwps" Dec 11 02:07:02 crc kubenswrapper[4824]: I1211 02:07:02.292630 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-7dwps" Dec 11 02:07:02 crc kubenswrapper[4824]: I1211 02:07:02.339626 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-7dwps" Dec 11 02:07:02 crc kubenswrapper[4824]: I1211 02:07:02.362156 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-4gv5b" Dec 11 02:07:02 crc kubenswrapper[4824]: I1211 02:07:02.400504 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-4qn4g" Dec 11 02:07:02 crc kubenswrapper[4824]: I1211 02:07:02.401317 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-7dwps" Dec 11 02:07:07 crc kubenswrapper[4824]: I1211 02:07:07.250867 4824 patch_prober.go:28] interesting pod/machine-config-daemon-gx6xt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 02:07:07 crc kubenswrapper[4824]: I1211 02:07:07.251417 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 02:07:09 crc kubenswrapper[4824]: I1211 02:07:09.713430 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-df294" Dec 11 02:07:09 crc kubenswrapper[4824]: I1211 02:07:09.781985 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-df294" Dec 11 02:07:20 crc kubenswrapper[4824]: I1211 02:07:20.400244 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-4gv5b" Dec 11 02:07:20 crc kubenswrapper[4824]: I1211 02:07:20.494260 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-254w2"] Dec 11 02:07:37 crc kubenswrapper[4824]: I1211 02:07:37.250908 4824 patch_prober.go:28] interesting pod/machine-config-daemon-gx6xt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 02:07:37 crc kubenswrapper[4824]: I1211 02:07:37.251549 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 02:07:45 crc kubenswrapper[4824]: I1211 02:07:45.618704 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-254w2" podUID="18d74f73-0f04-4f3c-88c4-a431d490db9d" containerName="registry" containerID="cri-o://3f56af012e4c19f54832d8b45fdab91a708a1cf9f8d8ad236c76736800cf9a4e" gracePeriod=30 Dec 11 02:07:46 crc kubenswrapper[4824]: I1211 02:07:46.107178 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-254w2" Dec 11 02:07:46 crc kubenswrapper[4824]: I1211 02:07:46.229041 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/18d74f73-0f04-4f3c-88c4-a431d490db9d-registry-certificates\") pod \"18d74f73-0f04-4f3c-88c4-a431d490db9d\" (UID: \"18d74f73-0f04-4f3c-88c4-a431d490db9d\") " Dec 11 02:07:46 crc kubenswrapper[4824]: I1211 02:07:46.229167 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/18d74f73-0f04-4f3c-88c4-a431d490db9d-installation-pull-secrets\") pod \"18d74f73-0f04-4f3c-88c4-a431d490db9d\" (UID: \"18d74f73-0f04-4f3c-88c4-a431d490db9d\") " Dec 11 02:07:46 crc kubenswrapper[4824]: I1211 02:07:46.229234 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/18d74f73-0f04-4f3c-88c4-a431d490db9d-trusted-ca\") pod \"18d74f73-0f04-4f3c-88c4-a431d490db9d\" (UID: \"18d74f73-0f04-4f3c-88c4-a431d490db9d\") " Dec 11 02:07:46 crc kubenswrapper[4824]: I1211 02:07:46.229267 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/18d74f73-0f04-4f3c-88c4-a431d490db9d-bound-sa-token\") pod \"18d74f73-0f04-4f3c-88c4-a431d490db9d\" (UID: \"18d74f73-0f04-4f3c-88c4-a431d490db9d\") " Dec 11 02:07:46 crc kubenswrapper[4824]: I1211 02:07:46.230812 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/18d74f73-0f04-4f3c-88c4-a431d490db9d-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "18d74f73-0f04-4f3c-88c4-a431d490db9d" (UID: "18d74f73-0f04-4f3c-88c4-a431d490db9d"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:07:46 crc kubenswrapper[4824]: I1211 02:07:46.230987 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"18d74f73-0f04-4f3c-88c4-a431d490db9d\" (UID: \"18d74f73-0f04-4f3c-88c4-a431d490db9d\") " Dec 11 02:07:46 crc kubenswrapper[4824]: I1211 02:07:46.231047 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/18d74f73-0f04-4f3c-88c4-a431d490db9d-registry-tls\") pod \"18d74f73-0f04-4f3c-88c4-a431d490db9d\" (UID: \"18d74f73-0f04-4f3c-88c4-a431d490db9d\") " Dec 11 02:07:46 crc kubenswrapper[4824]: I1211 02:07:46.231138 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z4d6h\" (UniqueName: \"kubernetes.io/projected/18d74f73-0f04-4f3c-88c4-a431d490db9d-kube-api-access-z4d6h\") pod \"18d74f73-0f04-4f3c-88c4-a431d490db9d\" (UID: \"18d74f73-0f04-4f3c-88c4-a431d490db9d\") " Dec 11 02:07:46 crc kubenswrapper[4824]: I1211 02:07:46.231192 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/18d74f73-0f04-4f3c-88c4-a431d490db9d-ca-trust-extracted\") pod \"18d74f73-0f04-4f3c-88c4-a431d490db9d\" (UID: \"18d74f73-0f04-4f3c-88c4-a431d490db9d\") " Dec 11 02:07:46 crc kubenswrapper[4824]: I1211 02:07:46.231516 4824 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/18d74f73-0f04-4f3c-88c4-a431d490db9d-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 11 02:07:46 crc kubenswrapper[4824]: I1211 02:07:46.240077 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18d74f73-0f04-4f3c-88c4-a431d490db9d-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "18d74f73-0f04-4f3c-88c4-a431d490db9d" (UID: "18d74f73-0f04-4f3c-88c4-a431d490db9d"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:07:46 crc kubenswrapper[4824]: I1211 02:07:46.244835 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18d74f73-0f04-4f3c-88c4-a431d490db9d-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "18d74f73-0f04-4f3c-88c4-a431d490db9d" (UID: "18d74f73-0f04-4f3c-88c4-a431d490db9d"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:07:46 crc kubenswrapper[4824]: I1211 02:07:46.251203 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/18d74f73-0f04-4f3c-88c4-a431d490db9d-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "18d74f73-0f04-4f3c-88c4-a431d490db9d" (UID: "18d74f73-0f04-4f3c-88c4-a431d490db9d"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:07:46 crc kubenswrapper[4824]: I1211 02:07:46.253198 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "18d74f73-0f04-4f3c-88c4-a431d490db9d" (UID: "18d74f73-0f04-4f3c-88c4-a431d490db9d"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 11 02:07:46 crc kubenswrapper[4824]: I1211 02:07:46.256390 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18d74f73-0f04-4f3c-88c4-a431d490db9d-kube-api-access-z4d6h" (OuterVolumeSpecName: "kube-api-access-z4d6h") pod "18d74f73-0f04-4f3c-88c4-a431d490db9d" (UID: "18d74f73-0f04-4f3c-88c4-a431d490db9d"). InnerVolumeSpecName "kube-api-access-z4d6h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:07:46 crc kubenswrapper[4824]: I1211 02:07:46.257952 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18d74f73-0f04-4f3c-88c4-a431d490db9d-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "18d74f73-0f04-4f3c-88c4-a431d490db9d" (UID: "18d74f73-0f04-4f3c-88c4-a431d490db9d"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:07:46 crc kubenswrapper[4824]: I1211 02:07:46.296798 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/18d74f73-0f04-4f3c-88c4-a431d490db9d-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "18d74f73-0f04-4f3c-88c4-a431d490db9d" (UID: "18d74f73-0f04-4f3c-88c4-a431d490db9d"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:07:46 crc kubenswrapper[4824]: I1211 02:07:46.333361 4824 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/18d74f73-0f04-4f3c-88c4-a431d490db9d-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 11 02:07:46 crc kubenswrapper[4824]: I1211 02:07:46.333421 4824 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/18d74f73-0f04-4f3c-88c4-a431d490db9d-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 11 02:07:46 crc kubenswrapper[4824]: I1211 02:07:46.333475 4824 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/18d74f73-0f04-4f3c-88c4-a431d490db9d-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 11 02:07:46 crc kubenswrapper[4824]: I1211 02:07:46.333495 4824 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/18d74f73-0f04-4f3c-88c4-a431d490db9d-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 11 02:07:46 crc kubenswrapper[4824]: I1211 02:07:46.333514 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z4d6h\" (UniqueName: \"kubernetes.io/projected/18d74f73-0f04-4f3c-88c4-a431d490db9d-kube-api-access-z4d6h\") on node \"crc\" DevicePath \"\"" Dec 11 02:07:46 crc kubenswrapper[4824]: I1211 02:07:46.333532 4824 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/18d74f73-0f04-4f3c-88c4-a431d490db9d-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 11 02:07:46 crc kubenswrapper[4824]: I1211 02:07:46.680640 4824 generic.go:334] "Generic (PLEG): container finished" podID="18d74f73-0f04-4f3c-88c4-a431d490db9d" containerID="3f56af012e4c19f54832d8b45fdab91a708a1cf9f8d8ad236c76736800cf9a4e" exitCode=0 Dec 11 02:07:46 crc kubenswrapper[4824]: I1211 02:07:46.680690 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-254w2" event={"ID":"18d74f73-0f04-4f3c-88c4-a431d490db9d","Type":"ContainerDied","Data":"3f56af012e4c19f54832d8b45fdab91a708a1cf9f8d8ad236c76736800cf9a4e"} Dec 11 02:07:46 crc kubenswrapper[4824]: I1211 02:07:46.680724 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-254w2" event={"ID":"18d74f73-0f04-4f3c-88c4-a431d490db9d","Type":"ContainerDied","Data":"f6e961ed0d22d4ebf17f47d28193643c29f600b6086ac18952ba7840db02ccc3"} Dec 11 02:07:46 crc kubenswrapper[4824]: I1211 02:07:46.680747 4824 scope.go:117] "RemoveContainer" containerID="3f56af012e4c19f54832d8b45fdab91a708a1cf9f8d8ad236c76736800cf9a4e" Dec 11 02:07:46 crc kubenswrapper[4824]: I1211 02:07:46.680787 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-254w2" Dec 11 02:07:46 crc kubenswrapper[4824]: I1211 02:07:46.710851 4824 scope.go:117] "RemoveContainer" containerID="3f56af012e4c19f54832d8b45fdab91a708a1cf9f8d8ad236c76736800cf9a4e" Dec 11 02:07:46 crc kubenswrapper[4824]: E1211 02:07:46.711834 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f56af012e4c19f54832d8b45fdab91a708a1cf9f8d8ad236c76736800cf9a4e\": container with ID starting with 3f56af012e4c19f54832d8b45fdab91a708a1cf9f8d8ad236c76736800cf9a4e not found: ID does not exist" containerID="3f56af012e4c19f54832d8b45fdab91a708a1cf9f8d8ad236c76736800cf9a4e" Dec 11 02:07:46 crc kubenswrapper[4824]: I1211 02:07:46.711912 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f56af012e4c19f54832d8b45fdab91a708a1cf9f8d8ad236c76736800cf9a4e"} err="failed to get container status \"3f56af012e4c19f54832d8b45fdab91a708a1cf9f8d8ad236c76736800cf9a4e\": rpc error: code = NotFound desc = could not find container \"3f56af012e4c19f54832d8b45fdab91a708a1cf9f8d8ad236c76736800cf9a4e\": container with ID starting with 3f56af012e4c19f54832d8b45fdab91a708a1cf9f8d8ad236c76736800cf9a4e not found: ID does not exist" Dec 11 02:07:46 crc kubenswrapper[4824]: I1211 02:07:46.715033 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-254w2"] Dec 11 02:07:46 crc kubenswrapper[4824]: I1211 02:07:46.721842 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-254w2"] Dec 11 02:07:48 crc kubenswrapper[4824]: I1211 02:07:48.641020 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="18d74f73-0f04-4f3c-88c4-a431d490db9d" path="/var/lib/kubelet/pods/18d74f73-0f04-4f3c-88c4-a431d490db9d/volumes" Dec 11 02:08:07 crc kubenswrapper[4824]: I1211 02:08:07.251650 4824 patch_prober.go:28] interesting pod/machine-config-daemon-gx6xt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 02:08:07 crc kubenswrapper[4824]: I1211 02:08:07.252290 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 02:08:07 crc kubenswrapper[4824]: I1211 02:08:07.252371 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" Dec 11 02:08:07 crc kubenswrapper[4824]: I1211 02:08:07.253240 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e5bb671b5b5757c6365db5e523a726c313eb514acdb624a11ae5ddd89c2d5112"} pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 11 02:08:07 crc kubenswrapper[4824]: I1211 02:08:07.253537 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" containerName="machine-config-daemon" containerID="cri-o://e5bb671b5b5757c6365db5e523a726c313eb514acdb624a11ae5ddd89c2d5112" gracePeriod=600 Dec 11 02:08:07 crc kubenswrapper[4824]: I1211 02:08:07.835893 4824 generic.go:334] "Generic (PLEG): container finished" podID="44b156e4-64a4-4d45-aa5e-9b10a862faed" containerID="e5bb671b5b5757c6365db5e523a726c313eb514acdb624a11ae5ddd89c2d5112" exitCode=0 Dec 11 02:08:07 crc kubenswrapper[4824]: I1211 02:08:07.836006 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" event={"ID":"44b156e4-64a4-4d45-aa5e-9b10a862faed","Type":"ContainerDied","Data":"e5bb671b5b5757c6365db5e523a726c313eb514acdb624a11ae5ddd89c2d5112"} Dec 11 02:08:07 crc kubenswrapper[4824]: I1211 02:08:07.836377 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" event={"ID":"44b156e4-64a4-4d45-aa5e-9b10a862faed","Type":"ContainerStarted","Data":"820ab8fffa074e96d20e064678952b74a92387b5649face8929acb3791ff6e88"} Dec 11 02:08:07 crc kubenswrapper[4824]: I1211 02:08:07.836442 4824 scope.go:117] "RemoveContainer" containerID="dcf289ac27dbda4f24aabbde3fb5119329e7b4fc7d6e018168da9a69661a4c57" Dec 11 02:10:07 crc kubenswrapper[4824]: I1211 02:10:07.251333 4824 patch_prober.go:28] interesting pod/machine-config-daemon-gx6xt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 02:10:07 crc kubenswrapper[4824]: I1211 02:10:07.251969 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 02:10:37 crc kubenswrapper[4824]: I1211 02:10:37.251251 4824 patch_prober.go:28] interesting pod/machine-config-daemon-gx6xt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 02:10:37 crc kubenswrapper[4824]: I1211 02:10:37.252096 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 02:11:07 crc kubenswrapper[4824]: I1211 02:11:07.251166 4824 patch_prober.go:28] interesting pod/machine-config-daemon-gx6xt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 02:11:07 crc kubenswrapper[4824]: I1211 02:11:07.251916 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 02:11:07 crc kubenswrapper[4824]: I1211 02:11:07.251983 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" Dec 11 02:11:07 crc kubenswrapper[4824]: I1211 02:11:07.252877 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"820ab8fffa074e96d20e064678952b74a92387b5649face8929acb3791ff6e88"} pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 11 02:11:07 crc kubenswrapper[4824]: I1211 02:11:07.252972 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" containerName="machine-config-daemon" containerID="cri-o://820ab8fffa074e96d20e064678952b74a92387b5649face8929acb3791ff6e88" gracePeriod=600 Dec 11 02:11:08 crc kubenswrapper[4824]: I1211 02:11:08.244095 4824 generic.go:334] "Generic (PLEG): container finished" podID="44b156e4-64a4-4d45-aa5e-9b10a862faed" containerID="820ab8fffa074e96d20e064678952b74a92387b5649face8929acb3791ff6e88" exitCode=0 Dec 11 02:11:08 crc kubenswrapper[4824]: I1211 02:11:08.244174 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" event={"ID":"44b156e4-64a4-4d45-aa5e-9b10a862faed","Type":"ContainerDied","Data":"820ab8fffa074e96d20e064678952b74a92387b5649face8929acb3791ff6e88"} Dec 11 02:11:08 crc kubenswrapper[4824]: I1211 02:11:08.244708 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" event={"ID":"44b156e4-64a4-4d45-aa5e-9b10a862faed","Type":"ContainerStarted","Data":"db8b560cfc3483491a894f8e917ea1a5928fa44440c12ac1a044fd79b218d54a"} Dec 11 02:11:08 crc kubenswrapper[4824]: I1211 02:11:08.244748 4824 scope.go:117] "RemoveContainer" containerID="e5bb671b5b5757c6365db5e523a726c313eb514acdb624a11ae5ddd89c2d5112" Dec 11 02:13:07 crc kubenswrapper[4824]: I1211 02:13:07.250840 4824 patch_prober.go:28] interesting pod/machine-config-daemon-gx6xt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 02:13:07 crc kubenswrapper[4824]: I1211 02:13:07.251516 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 02:13:29 crc kubenswrapper[4824]: I1211 02:13:29.838435 4824 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 11 02:13:37 crc kubenswrapper[4824]: I1211 02:13:37.252368 4824 patch_prober.go:28] interesting pod/machine-config-daemon-gx6xt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 02:13:37 crc kubenswrapper[4824]: I1211 02:13:37.253089 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 02:13:55 crc kubenswrapper[4824]: I1211 02:13:55.883794 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-bstmj"] Dec 11 02:13:55 crc kubenswrapper[4824]: E1211 02:13:55.884638 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18d74f73-0f04-4f3c-88c4-a431d490db9d" containerName="registry" Dec 11 02:13:55 crc kubenswrapper[4824]: I1211 02:13:55.884654 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="18d74f73-0f04-4f3c-88c4-a431d490db9d" containerName="registry" Dec 11 02:13:55 crc kubenswrapper[4824]: I1211 02:13:55.884784 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="18d74f73-0f04-4f3c-88c4-a431d490db9d" containerName="registry" Dec 11 02:13:55 crc kubenswrapper[4824]: I1211 02:13:55.885266 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-bstmj" Dec 11 02:13:55 crc kubenswrapper[4824]: I1211 02:13:55.893959 4824 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-6xxrl" Dec 11 02:13:55 crc kubenswrapper[4824]: I1211 02:13:55.895033 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Dec 11 02:13:55 crc kubenswrapper[4824]: I1211 02:13:55.895738 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Dec 11 02:13:55 crc kubenswrapper[4824]: I1211 02:13:55.906857 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-7jv9b"] Dec 11 02:13:55 crc kubenswrapper[4824]: I1211 02:13:55.907755 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-7jv9b" Dec 11 02:13:55 crc kubenswrapper[4824]: I1211 02:13:55.910190 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-7jv9b"] Dec 11 02:13:55 crc kubenswrapper[4824]: I1211 02:13:55.914710 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-bstmj"] Dec 11 02:13:55 crc kubenswrapper[4824]: I1211 02:13:55.921418 4824 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-54dqj" Dec 11 02:13:55 crc kubenswrapper[4824]: I1211 02:13:55.922575 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r755m\" (UniqueName: \"kubernetes.io/projected/919ec639-d2fa-4447-86a4-a88b39d05ec3-kube-api-access-r755m\") pod \"cert-manager-cainjector-7f985d654d-bstmj\" (UID: \"919ec639-d2fa-4447-86a4-a88b39d05ec3\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-bstmj" Dec 11 02:13:56 crc kubenswrapper[4824]: I1211 02:13:56.002321 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-hxtz9"] Dec 11 02:13:56 crc kubenswrapper[4824]: I1211 02:13:56.002925 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-hxtz9" Dec 11 02:13:56 crc kubenswrapper[4824]: I1211 02:13:56.005915 4824 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-wl225" Dec 11 02:13:56 crc kubenswrapper[4824]: I1211 02:13:56.023364 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r755m\" (UniqueName: \"kubernetes.io/projected/919ec639-d2fa-4447-86a4-a88b39d05ec3-kube-api-access-r755m\") pod \"cert-manager-cainjector-7f985d654d-bstmj\" (UID: \"919ec639-d2fa-4447-86a4-a88b39d05ec3\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-bstmj" Dec 11 02:13:56 crc kubenswrapper[4824]: I1211 02:13:56.023613 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9t29l\" (UniqueName: \"kubernetes.io/projected/81bcc53d-3664-4987-8724-765fda6c8c09-kube-api-access-9t29l\") pod \"cert-manager-5b446d88c5-7jv9b\" (UID: \"81bcc53d-3664-4987-8724-765fda6c8c09\") " pod="cert-manager/cert-manager-5b446d88c5-7jv9b" Dec 11 02:13:56 crc kubenswrapper[4824]: I1211 02:13:56.029516 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-hxtz9"] Dec 11 02:13:56 crc kubenswrapper[4824]: I1211 02:13:56.051074 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r755m\" (UniqueName: \"kubernetes.io/projected/919ec639-d2fa-4447-86a4-a88b39d05ec3-kube-api-access-r755m\") pod \"cert-manager-cainjector-7f985d654d-bstmj\" (UID: \"919ec639-d2fa-4447-86a4-a88b39d05ec3\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-bstmj" Dec 11 02:13:56 crc kubenswrapper[4824]: I1211 02:13:56.125010 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xdqzp\" (UniqueName: \"kubernetes.io/projected/b9bca511-59e3-4650-86c8-a82c0175d51a-kube-api-access-xdqzp\") pod \"cert-manager-webhook-5655c58dd6-hxtz9\" (UID: \"b9bca511-59e3-4650-86c8-a82c0175d51a\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-hxtz9" Dec 11 02:13:56 crc kubenswrapper[4824]: I1211 02:13:56.125072 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9t29l\" (UniqueName: \"kubernetes.io/projected/81bcc53d-3664-4987-8724-765fda6c8c09-kube-api-access-9t29l\") pod \"cert-manager-5b446d88c5-7jv9b\" (UID: \"81bcc53d-3664-4987-8724-765fda6c8c09\") " pod="cert-manager/cert-manager-5b446d88c5-7jv9b" Dec 11 02:13:56 crc kubenswrapper[4824]: I1211 02:13:56.144388 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-4mrwp"] Dec 11 02:13:56 crc kubenswrapper[4824]: I1211 02:13:56.145311 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4mrwp" Dec 11 02:13:56 crc kubenswrapper[4824]: I1211 02:13:56.148386 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9t29l\" (UniqueName: \"kubernetes.io/projected/81bcc53d-3664-4987-8724-765fda6c8c09-kube-api-access-9t29l\") pod \"cert-manager-5b446d88c5-7jv9b\" (UID: \"81bcc53d-3664-4987-8724-765fda6c8c09\") " pod="cert-manager/cert-manager-5b446d88c5-7jv9b" Dec 11 02:13:56 crc kubenswrapper[4824]: I1211 02:13:56.157124 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4mrwp"] Dec 11 02:13:56 crc kubenswrapper[4824]: I1211 02:13:56.210989 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-bstmj" Dec 11 02:13:56 crc kubenswrapper[4824]: I1211 02:13:56.225788 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xdqzp\" (UniqueName: \"kubernetes.io/projected/b9bca511-59e3-4650-86c8-a82c0175d51a-kube-api-access-xdqzp\") pod \"cert-manager-webhook-5655c58dd6-hxtz9\" (UID: \"b9bca511-59e3-4650-86c8-a82c0175d51a\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-hxtz9" Dec 11 02:13:56 crc kubenswrapper[4824]: I1211 02:13:56.226025 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-7jv9b" Dec 11 02:13:56 crc kubenswrapper[4824]: I1211 02:13:56.247172 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xdqzp\" (UniqueName: \"kubernetes.io/projected/b9bca511-59e3-4650-86c8-a82c0175d51a-kube-api-access-xdqzp\") pod \"cert-manager-webhook-5655c58dd6-hxtz9\" (UID: \"b9bca511-59e3-4650-86c8-a82c0175d51a\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-hxtz9" Dec 11 02:13:56 crc kubenswrapper[4824]: I1211 02:13:56.317041 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-hxtz9" Dec 11 02:13:56 crc kubenswrapper[4824]: I1211 02:13:56.328070 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/11965819-0e2c-4107-a14c-2c7c2653701e-utilities\") pod \"redhat-marketplace-4mrwp\" (UID: \"11965819-0e2c-4107-a14c-2c7c2653701e\") " pod="openshift-marketplace/redhat-marketplace-4mrwp" Dec 11 02:13:56 crc kubenswrapper[4824]: I1211 02:13:56.328384 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqssf\" (UniqueName: \"kubernetes.io/projected/11965819-0e2c-4107-a14c-2c7c2653701e-kube-api-access-jqssf\") pod \"redhat-marketplace-4mrwp\" (UID: \"11965819-0e2c-4107-a14c-2c7c2653701e\") " pod="openshift-marketplace/redhat-marketplace-4mrwp" Dec 11 02:13:56 crc kubenswrapper[4824]: I1211 02:13:56.328981 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/11965819-0e2c-4107-a14c-2c7c2653701e-catalog-content\") pod \"redhat-marketplace-4mrwp\" (UID: \"11965819-0e2c-4107-a14c-2c7c2653701e\") " pod="openshift-marketplace/redhat-marketplace-4mrwp" Dec 11 02:13:56 crc kubenswrapper[4824]: I1211 02:13:56.430042 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/11965819-0e2c-4107-a14c-2c7c2653701e-catalog-content\") pod \"redhat-marketplace-4mrwp\" (UID: \"11965819-0e2c-4107-a14c-2c7c2653701e\") " pod="openshift-marketplace/redhat-marketplace-4mrwp" Dec 11 02:13:56 crc kubenswrapper[4824]: I1211 02:13:56.430158 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/11965819-0e2c-4107-a14c-2c7c2653701e-utilities\") pod \"redhat-marketplace-4mrwp\" (UID: \"11965819-0e2c-4107-a14c-2c7c2653701e\") " pod="openshift-marketplace/redhat-marketplace-4mrwp" Dec 11 02:13:56 crc kubenswrapper[4824]: I1211 02:13:56.430200 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jqssf\" (UniqueName: \"kubernetes.io/projected/11965819-0e2c-4107-a14c-2c7c2653701e-kube-api-access-jqssf\") pod \"redhat-marketplace-4mrwp\" (UID: \"11965819-0e2c-4107-a14c-2c7c2653701e\") " pod="openshift-marketplace/redhat-marketplace-4mrwp" Dec 11 02:13:56 crc kubenswrapper[4824]: I1211 02:13:56.430906 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/11965819-0e2c-4107-a14c-2c7c2653701e-catalog-content\") pod \"redhat-marketplace-4mrwp\" (UID: \"11965819-0e2c-4107-a14c-2c7c2653701e\") " pod="openshift-marketplace/redhat-marketplace-4mrwp" Dec 11 02:13:56 crc kubenswrapper[4824]: I1211 02:13:56.431041 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/11965819-0e2c-4107-a14c-2c7c2653701e-utilities\") pod \"redhat-marketplace-4mrwp\" (UID: \"11965819-0e2c-4107-a14c-2c7c2653701e\") " pod="openshift-marketplace/redhat-marketplace-4mrwp" Dec 11 02:13:56 crc kubenswrapper[4824]: I1211 02:13:56.452669 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jqssf\" (UniqueName: \"kubernetes.io/projected/11965819-0e2c-4107-a14c-2c7c2653701e-kube-api-access-jqssf\") pod \"redhat-marketplace-4mrwp\" (UID: \"11965819-0e2c-4107-a14c-2c7c2653701e\") " pod="openshift-marketplace/redhat-marketplace-4mrwp" Dec 11 02:13:56 crc kubenswrapper[4824]: I1211 02:13:56.514838 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-7jv9b"] Dec 11 02:13:56 crc kubenswrapper[4824]: I1211 02:13:56.521667 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4mrwp" Dec 11 02:13:56 crc kubenswrapper[4824]: I1211 02:13:56.532862 4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 11 02:13:56 crc kubenswrapper[4824]: I1211 02:13:56.584346 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-hxtz9"] Dec 11 02:13:56 crc kubenswrapper[4824]: W1211 02:13:56.588255 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb9bca511_59e3_4650_86c8_a82c0175d51a.slice/crio-ce2d9dec8aefa1a79a0462e6b4516516642d31c15643da663dae5cae05f502dc WatchSource:0}: Error finding container ce2d9dec8aefa1a79a0462e6b4516516642d31c15643da663dae5cae05f502dc: Status 404 returned error can't find the container with id ce2d9dec8aefa1a79a0462e6b4516516642d31c15643da663dae5cae05f502dc Dec 11 02:13:56 crc kubenswrapper[4824]: I1211 02:13:56.656469 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-bstmj"] Dec 11 02:13:56 crc kubenswrapper[4824]: I1211 02:13:56.927012 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4mrwp"] Dec 11 02:13:56 crc kubenswrapper[4824]: W1211 02:13:56.933329 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod11965819_0e2c_4107_a14c_2c7c2653701e.slice/crio-2c6371b30959fdcb41e9ff7a99ef960977653369f596d2087a68fdbc9064ee9f WatchSource:0}: Error finding container 2c6371b30959fdcb41e9ff7a99ef960977653369f596d2087a68fdbc9064ee9f: Status 404 returned error can't find the container with id 2c6371b30959fdcb41e9ff7a99ef960977653369f596d2087a68fdbc9064ee9f Dec 11 02:13:57 crc kubenswrapper[4824]: I1211 02:13:57.375438 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-7jv9b" event={"ID":"81bcc53d-3664-4987-8724-765fda6c8c09","Type":"ContainerStarted","Data":"b622f80cc1914f490238e610ac03b59d2f83dea3e076e0163432a6f13cd934df"} Dec 11 02:13:57 crc kubenswrapper[4824]: I1211 02:13:57.378295 4824 generic.go:334] "Generic (PLEG): container finished" podID="11965819-0e2c-4107-a14c-2c7c2653701e" containerID="75fec8e1f16bd2bed3219ed3ec1e9d024bdd535d556fd0f186ab3ef4e3fe1a25" exitCode=0 Dec 11 02:13:57 crc kubenswrapper[4824]: I1211 02:13:57.378420 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4mrwp" event={"ID":"11965819-0e2c-4107-a14c-2c7c2653701e","Type":"ContainerDied","Data":"75fec8e1f16bd2bed3219ed3ec1e9d024bdd535d556fd0f186ab3ef4e3fe1a25"} Dec 11 02:13:57 crc kubenswrapper[4824]: I1211 02:13:57.378510 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4mrwp" event={"ID":"11965819-0e2c-4107-a14c-2c7c2653701e","Type":"ContainerStarted","Data":"2c6371b30959fdcb41e9ff7a99ef960977653369f596d2087a68fdbc9064ee9f"} Dec 11 02:13:57 crc kubenswrapper[4824]: I1211 02:13:57.386144 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-hxtz9" event={"ID":"b9bca511-59e3-4650-86c8-a82c0175d51a","Type":"ContainerStarted","Data":"ce2d9dec8aefa1a79a0462e6b4516516642d31c15643da663dae5cae05f502dc"} Dec 11 02:13:57 crc kubenswrapper[4824]: I1211 02:13:57.388632 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-bstmj" event={"ID":"919ec639-d2fa-4447-86a4-a88b39d05ec3","Type":"ContainerStarted","Data":"684927ed5a088f926d843ef8f2f7ecf8fe45db9f9353fbbb607f11dccedc7de2"} Dec 11 02:14:00 crc kubenswrapper[4824]: I1211 02:14:00.413838 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4mrwp" event={"ID":"11965819-0e2c-4107-a14c-2c7c2653701e","Type":"ContainerStarted","Data":"c26153691c06d47721e32a065fe6744a4638a9a48a77897c876e81428f70b088"} Dec 11 02:14:00 crc kubenswrapper[4824]: I1211 02:14:00.416968 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-hxtz9" event={"ID":"b9bca511-59e3-4650-86c8-a82c0175d51a","Type":"ContainerStarted","Data":"a6724bb13bdb99b109f1dbf4fd2f2dcc9da0033ef3237c5d61c93e038a58f6f2"} Dec 11 02:14:00 crc kubenswrapper[4824]: I1211 02:14:00.417760 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-hxtz9" Dec 11 02:14:00 crc kubenswrapper[4824]: I1211 02:14:00.420300 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-bstmj" event={"ID":"919ec639-d2fa-4447-86a4-a88b39d05ec3","Type":"ContainerStarted","Data":"a0a6bd009a7a50327c45f1e642aa2442e3a1c3e7f13ee2e1d676bf7edce79216"} Dec 11 02:14:00 crc kubenswrapper[4824]: I1211 02:14:00.459180 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-bstmj" podStartSLOduration=1.9415470849999998 podStartE2EDuration="5.459158666s" podCreationTimestamp="2025-12-11 02:13:55 +0000 UTC" firstStartedPulling="2025-12-11 02:13:56.664537403 +0000 UTC m=+778.353574782" lastFinishedPulling="2025-12-11 02:14:00.182148964 +0000 UTC m=+781.871186363" observedRunningTime="2025-12-11 02:14:00.456521344 +0000 UTC m=+782.145558723" watchObservedRunningTime="2025-12-11 02:14:00.459158666 +0000 UTC m=+782.148196055" Dec 11 02:14:00 crc kubenswrapper[4824]: I1211 02:14:00.474405 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-hxtz9" podStartSLOduration=1.8101514079999999 podStartE2EDuration="5.474386237s" podCreationTimestamp="2025-12-11 02:13:55 +0000 UTC" firstStartedPulling="2025-12-11 02:13:56.590041253 +0000 UTC m=+778.279078632" lastFinishedPulling="2025-12-11 02:14:00.254276082 +0000 UTC m=+781.943313461" observedRunningTime="2025-12-11 02:14:00.469214504 +0000 UTC m=+782.158251873" watchObservedRunningTime="2025-12-11 02:14:00.474386237 +0000 UTC m=+782.163423626" Dec 11 02:14:01 crc kubenswrapper[4824]: I1211 02:14:01.430102 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-7jv9b" event={"ID":"81bcc53d-3664-4987-8724-765fda6c8c09","Type":"ContainerStarted","Data":"96c7325c9df4fdf9150ef0352ee5fa2cd56b48ff8993bd9259742a8ecc991c8c"} Dec 11 02:14:01 crc kubenswrapper[4824]: I1211 02:14:01.433174 4824 generic.go:334] "Generic (PLEG): container finished" podID="11965819-0e2c-4107-a14c-2c7c2653701e" containerID="c26153691c06d47721e32a065fe6744a4638a9a48a77897c876e81428f70b088" exitCode=0 Dec 11 02:14:01 crc kubenswrapper[4824]: I1211 02:14:01.433238 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4mrwp" event={"ID":"11965819-0e2c-4107-a14c-2c7c2653701e","Type":"ContainerDied","Data":"c26153691c06d47721e32a065fe6744a4638a9a48a77897c876e81428f70b088"} Dec 11 02:14:01 crc kubenswrapper[4824]: I1211 02:14:01.455712 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-7jv9b" podStartSLOduration=2.807925242 podStartE2EDuration="6.455685793s" podCreationTimestamp="2025-12-11 02:13:55 +0000 UTC" firstStartedPulling="2025-12-11 02:13:56.532587264 +0000 UTC m=+778.221624643" lastFinishedPulling="2025-12-11 02:14:00.180347815 +0000 UTC m=+781.869385194" observedRunningTime="2025-12-11 02:14:01.451797086 +0000 UTC m=+783.140834495" watchObservedRunningTime="2025-12-11 02:14:01.455685793 +0000 UTC m=+783.144723212" Dec 11 02:14:02 crc kubenswrapper[4824]: I1211 02:14:02.440722 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4mrwp" event={"ID":"11965819-0e2c-4107-a14c-2c7c2653701e","Type":"ContainerStarted","Data":"43d294562be066e780dc4992ad6c2822a7f104ac79ca294155feabd9d73fec4f"} Dec 11 02:14:02 crc kubenswrapper[4824]: I1211 02:14:02.469010 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-4mrwp" podStartSLOduration=1.9943285579999999 podStartE2EDuration="6.468991672s" podCreationTimestamp="2025-12-11 02:13:56 +0000 UTC" firstStartedPulling="2025-12-11 02:13:57.381051411 +0000 UTC m=+779.070088790" lastFinishedPulling="2025-12-11 02:14:01.855714495 +0000 UTC m=+783.544751904" observedRunningTime="2025-12-11 02:14:02.465711 +0000 UTC m=+784.154748389" watchObservedRunningTime="2025-12-11 02:14:02.468991672 +0000 UTC m=+784.158029061" Dec 11 02:14:06 crc kubenswrapper[4824]: I1211 02:14:06.324294 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-hxtz9" Dec 11 02:14:06 crc kubenswrapper[4824]: I1211 02:14:06.522502 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-4mrwp" Dec 11 02:14:06 crc kubenswrapper[4824]: I1211 02:14:06.522591 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-4mrwp" Dec 11 02:14:06 crc kubenswrapper[4824]: I1211 02:14:06.587227 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-4mrwp" Dec 11 02:14:06 crc kubenswrapper[4824]: I1211 02:14:06.659610 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-f6pgc"] Dec 11 02:14:06 crc kubenswrapper[4824]: I1211 02:14:06.660024 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" podUID="5b193bd2-b536-4056-92f1-94c9836ab2eb" containerName="ovn-controller" containerID="cri-o://4b4a3bc50eb80777188206adab774e8acad8b2e4952a8415f91b1c29ba9a0ffd" gracePeriod=30 Dec 11 02:14:06 crc kubenswrapper[4824]: I1211 02:14:06.660085 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" podUID="5b193bd2-b536-4056-92f1-94c9836ab2eb" containerName="nbdb" containerID="cri-o://4ed2ae40b1fe28111efc76d9d72345edecb5f3ea34522b238f3730812b29f2bd" gracePeriod=30 Dec 11 02:14:06 crc kubenswrapper[4824]: I1211 02:14:06.660154 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" podUID="5b193bd2-b536-4056-92f1-94c9836ab2eb" containerName="kube-rbac-proxy-node" containerID="cri-o://51f3c41d86ddc44f7eca118029876ad353317f753d0ad173ed4d30c2a15f10e6" gracePeriod=30 Dec 11 02:14:06 crc kubenswrapper[4824]: I1211 02:14:06.660204 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" podUID="5b193bd2-b536-4056-92f1-94c9836ab2eb" containerName="ovn-acl-logging" containerID="cri-o://6aa9fc1f1bff192c69bd59d32a05d4a72d09641d9ce4b265bc4517147a7e15dd" gracePeriod=30 Dec 11 02:14:06 crc kubenswrapper[4824]: I1211 02:14:06.660210 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" podUID="5b193bd2-b536-4056-92f1-94c9836ab2eb" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://c345850891744a0240781277e9a89a59f6899c4c944995f62e44650f807440a1" gracePeriod=30 Dec 11 02:14:06 crc kubenswrapper[4824]: I1211 02:14:06.660329 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" podUID="5b193bd2-b536-4056-92f1-94c9836ab2eb" containerName="sbdb" containerID="cri-o://9ead58694813806d307bbbc6a7b3cf41730d1c059b231a3002684bf09c10c7f4" gracePeriod=30 Dec 11 02:14:06 crc kubenswrapper[4824]: I1211 02:14:06.660319 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" podUID="5b193bd2-b536-4056-92f1-94c9836ab2eb" containerName="northd" containerID="cri-o://ce7aa078800e6f299cfe03770fa7e3b64fa3742fe0c0f383c87d84cf48569ddf" gracePeriod=30 Dec 11 02:14:06 crc kubenswrapper[4824]: I1211 02:14:06.703859 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" podUID="5b193bd2-b536-4056-92f1-94c9836ab2eb" containerName="ovnkube-controller" containerID="cri-o://8bb86ae130147feda9a5af751733489c3e0cc293e47e555879ae1fe0392c640e" gracePeriod=30 Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.251756 4824 patch_prober.go:28] interesting pod/machine-config-daemon-gx6xt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.251846 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.251908 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.252818 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"db8b560cfc3483491a894f8e917ea1a5928fa44440c12ac1a044fd79b218d54a"} pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.252915 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" containerName="machine-config-daemon" containerID="cri-o://db8b560cfc3483491a894f8e917ea1a5928fa44440c12ac1a044fd79b218d54a" gracePeriod=600 Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.475887 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f6pgc_5b193bd2-b536-4056-92f1-94c9836ab2eb/ovnkube-controller/3.log" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.478224 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f6pgc_5b193bd2-b536-4056-92f1-94c9836ab2eb/ovnkube-controller/3.log" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.480059 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f6pgc_5b193bd2-b536-4056-92f1-94c9836ab2eb/ovn-acl-logging/0.log" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.480691 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f6pgc_5b193bd2-b536-4056-92f1-94c9836ab2eb/ovn-controller/0.log" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.481420 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.481604 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f6pgc_5b193bd2-b536-4056-92f1-94c9836ab2eb/ovn-acl-logging/0.log" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.483289 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f6pgc_5b193bd2-b536-4056-92f1-94c9836ab2eb/ovn-controller/0.log" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.484136 4824 generic.go:334] "Generic (PLEG): container finished" podID="5b193bd2-b536-4056-92f1-94c9836ab2eb" containerID="8bb86ae130147feda9a5af751733489c3e0cc293e47e555879ae1fe0392c640e" exitCode=0 Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.484174 4824 generic.go:334] "Generic (PLEG): container finished" podID="5b193bd2-b536-4056-92f1-94c9836ab2eb" containerID="9ead58694813806d307bbbc6a7b3cf41730d1c059b231a3002684bf09c10c7f4" exitCode=0 Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.484187 4824 generic.go:334] "Generic (PLEG): container finished" podID="5b193bd2-b536-4056-92f1-94c9836ab2eb" containerID="4ed2ae40b1fe28111efc76d9d72345edecb5f3ea34522b238f3730812b29f2bd" exitCode=0 Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.484197 4824 generic.go:334] "Generic (PLEG): container finished" podID="5b193bd2-b536-4056-92f1-94c9836ab2eb" containerID="ce7aa078800e6f299cfe03770fa7e3b64fa3742fe0c0f383c87d84cf48569ddf" exitCode=0 Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.484205 4824 generic.go:334] "Generic (PLEG): container finished" podID="5b193bd2-b536-4056-92f1-94c9836ab2eb" containerID="c345850891744a0240781277e9a89a59f6899c4c944995f62e44650f807440a1" exitCode=0 Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.484213 4824 generic.go:334] "Generic (PLEG): container finished" podID="5b193bd2-b536-4056-92f1-94c9836ab2eb" containerID="51f3c41d86ddc44f7eca118029876ad353317f753d0ad173ed4d30c2a15f10e6" exitCode=0 Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.484221 4824 generic.go:334] "Generic (PLEG): container finished" podID="5b193bd2-b536-4056-92f1-94c9836ab2eb" containerID="6aa9fc1f1bff192c69bd59d32a05d4a72d09641d9ce4b265bc4517147a7e15dd" exitCode=143 Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.484230 4824 generic.go:334] "Generic (PLEG): container finished" podID="5b193bd2-b536-4056-92f1-94c9836ab2eb" containerID="4b4a3bc50eb80777188206adab774e8acad8b2e4952a8415f91b1c29ba9a0ffd" exitCode=143 Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.484225 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" event={"ID":"5b193bd2-b536-4056-92f1-94c9836ab2eb","Type":"ContainerDied","Data":"8bb86ae130147feda9a5af751733489c3e0cc293e47e555879ae1fe0392c640e"} Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.484281 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" event={"ID":"5b193bd2-b536-4056-92f1-94c9836ab2eb","Type":"ContainerDied","Data":"9ead58694813806d307bbbc6a7b3cf41730d1c059b231a3002684bf09c10c7f4"} Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.484299 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" event={"ID":"5b193bd2-b536-4056-92f1-94c9836ab2eb","Type":"ContainerDied","Data":"4ed2ae40b1fe28111efc76d9d72345edecb5f3ea34522b238f3730812b29f2bd"} Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.484314 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" event={"ID":"5b193bd2-b536-4056-92f1-94c9836ab2eb","Type":"ContainerDied","Data":"ce7aa078800e6f299cfe03770fa7e3b64fa3742fe0c0f383c87d84cf48569ddf"} Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.484326 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" event={"ID":"5b193bd2-b536-4056-92f1-94c9836ab2eb","Type":"ContainerDied","Data":"c345850891744a0240781277e9a89a59f6899c4c944995f62e44650f807440a1"} Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.484338 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" event={"ID":"5b193bd2-b536-4056-92f1-94c9836ab2eb","Type":"ContainerDied","Data":"51f3c41d86ddc44f7eca118029876ad353317f753d0ad173ed4d30c2a15f10e6"} Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.484351 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"154d82fd6acee749bfd37220547e52647dc7f9d86a73c28ce7d8577c9cadef6e"} Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.484364 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9ead58694813806d307bbbc6a7b3cf41730d1c059b231a3002684bf09c10c7f4"} Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.484374 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4ed2ae40b1fe28111efc76d9d72345edecb5f3ea34522b238f3730812b29f2bd"} Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.484379 4824 scope.go:117] "RemoveContainer" containerID="8bb86ae130147feda9a5af751733489c3e0cc293e47e555879ae1fe0392c640e" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.484382 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ce7aa078800e6f299cfe03770fa7e3b64fa3742fe0c0f383c87d84cf48569ddf"} Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.484477 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c345850891744a0240781277e9a89a59f6899c4c944995f62e44650f807440a1"} Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.484493 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"51f3c41d86ddc44f7eca118029876ad353317f753d0ad173ed4d30c2a15f10e6"} Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.484501 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6aa9fc1f1bff192c69bd59d32a05d4a72d09641d9ce4b265bc4517147a7e15dd"} Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.484508 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4b4a3bc50eb80777188206adab774e8acad8b2e4952a8415f91b1c29ba9a0ffd"} Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.484515 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df"} Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.484537 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" event={"ID":"5b193bd2-b536-4056-92f1-94c9836ab2eb","Type":"ContainerDied","Data":"6aa9fc1f1bff192c69bd59d32a05d4a72d09641d9ce4b265bc4517147a7e15dd"} Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.484558 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8bb86ae130147feda9a5af751733489c3e0cc293e47e555879ae1fe0392c640e"} Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.484566 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"154d82fd6acee749bfd37220547e52647dc7f9d86a73c28ce7d8577c9cadef6e"} Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.484574 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9ead58694813806d307bbbc6a7b3cf41730d1c059b231a3002684bf09c10c7f4"} Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.484581 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4ed2ae40b1fe28111efc76d9d72345edecb5f3ea34522b238f3730812b29f2bd"} Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.484588 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ce7aa078800e6f299cfe03770fa7e3b64fa3742fe0c0f383c87d84cf48569ddf"} Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.484607 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c345850891744a0240781277e9a89a59f6899c4c944995f62e44650f807440a1"} Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.484615 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"51f3c41d86ddc44f7eca118029876ad353317f753d0ad173ed4d30c2a15f10e6"} Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.484623 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6aa9fc1f1bff192c69bd59d32a05d4a72d09641d9ce4b265bc4517147a7e15dd"} Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.484631 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4b4a3bc50eb80777188206adab774e8acad8b2e4952a8415f91b1c29ba9a0ffd"} Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.484639 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df"} Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.484649 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" event={"ID":"5b193bd2-b536-4056-92f1-94c9836ab2eb","Type":"ContainerDied","Data":"4b4a3bc50eb80777188206adab774e8acad8b2e4952a8415f91b1c29ba9a0ffd"} Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.484661 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8bb86ae130147feda9a5af751733489c3e0cc293e47e555879ae1fe0392c640e"} Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.484670 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"154d82fd6acee749bfd37220547e52647dc7f9d86a73c28ce7d8577c9cadef6e"} Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.484677 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9ead58694813806d307bbbc6a7b3cf41730d1c059b231a3002684bf09c10c7f4"} Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.484684 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4ed2ae40b1fe28111efc76d9d72345edecb5f3ea34522b238f3730812b29f2bd"} Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.484691 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ce7aa078800e6f299cfe03770fa7e3b64fa3742fe0c0f383c87d84cf48569ddf"} Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.484698 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c345850891744a0240781277e9a89a59f6899c4c944995f62e44650f807440a1"} Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.484704 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"51f3c41d86ddc44f7eca118029876ad353317f753d0ad173ed4d30c2a15f10e6"} Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.484711 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6aa9fc1f1bff192c69bd59d32a05d4a72d09641d9ce4b265bc4517147a7e15dd"} Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.484717 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4b4a3bc50eb80777188206adab774e8acad8b2e4952a8415f91b1c29ba9a0ffd"} Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.484724 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df"} Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.484734 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" event={"ID":"5b193bd2-b536-4056-92f1-94c9836ab2eb","Type":"ContainerDied","Data":"b82eae5d23b57e21a6683458a036bb2e02cb72681d142360940afa5217ed72e3"} Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.484744 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8bb86ae130147feda9a5af751733489c3e0cc293e47e555879ae1fe0392c640e"} Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.484752 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"154d82fd6acee749bfd37220547e52647dc7f9d86a73c28ce7d8577c9cadef6e"} Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.484759 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9ead58694813806d307bbbc6a7b3cf41730d1c059b231a3002684bf09c10c7f4"} Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.484766 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4ed2ae40b1fe28111efc76d9d72345edecb5f3ea34522b238f3730812b29f2bd"} Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.484772 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ce7aa078800e6f299cfe03770fa7e3b64fa3742fe0c0f383c87d84cf48569ddf"} Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.484778 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c345850891744a0240781277e9a89a59f6899c4c944995f62e44650f807440a1"} Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.484785 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"51f3c41d86ddc44f7eca118029876ad353317f753d0ad173ed4d30c2a15f10e6"} Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.484792 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6aa9fc1f1bff192c69bd59d32a05d4a72d09641d9ce4b265bc4517147a7e15dd"} Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.484800 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4b4a3bc50eb80777188206adab774e8acad8b2e4952a8415f91b1c29ba9a0ffd"} Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.484808 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df"} Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.487098 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6bkc2_8022f696-d0e7-437b-848a-3cd25bd1f364/kube-multus/2.log" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.488033 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6bkc2_8022f696-d0e7-437b-848a-3cd25bd1f364/kube-multus/1.log" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.488159 4824 generic.go:334] "Generic (PLEG): container finished" podID="8022f696-d0e7-437b-848a-3cd25bd1f364" containerID="468cb1b4a3153f6a5f696a8cf76b784ea23f5a51db91c4c28a6994cbe54bd55c" exitCode=2 Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.488246 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-6bkc2" event={"ID":"8022f696-d0e7-437b-848a-3cd25bd1f364","Type":"ContainerDied","Data":"468cb1b4a3153f6a5f696a8cf76b784ea23f5a51db91c4c28a6994cbe54bd55c"} Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.488279 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f40719f59881e077e7edbab3adb9efd101359bd461e5434c365e119e8cdbddb2"} Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.489050 4824 scope.go:117] "RemoveContainer" containerID="468cb1b4a3153f6a5f696a8cf76b784ea23f5a51db91c4c28a6994cbe54bd55c" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.491633 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" event={"ID":"44b156e4-64a4-4d45-aa5e-9b10a862faed","Type":"ContainerDied","Data":"db8b560cfc3483491a894f8e917ea1a5928fa44440c12ac1a044fd79b218d54a"} Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.491675 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"820ab8fffa074e96d20e064678952b74a92387b5649face8929acb3791ff6e88"} Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.491554 4824 generic.go:334] "Generic (PLEG): container finished" podID="44b156e4-64a4-4d45-aa5e-9b10a862faed" containerID="db8b560cfc3483491a894f8e917ea1a5928fa44440c12ac1a044fd79b218d54a" exitCode=0 Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.527441 4824 scope.go:117] "RemoveContainer" containerID="154d82fd6acee749bfd37220547e52647dc7f9d86a73c28ce7d8577c9cadef6e" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.569416 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-65ngd"] Dec 11 02:14:07 crc kubenswrapper[4824]: E1211 02:14:07.570341 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b193bd2-b536-4056-92f1-94c9836ab2eb" containerName="ovnkube-controller" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.570369 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b193bd2-b536-4056-92f1-94c9836ab2eb" containerName="ovnkube-controller" Dec 11 02:14:07 crc kubenswrapper[4824]: E1211 02:14:07.570384 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b193bd2-b536-4056-92f1-94c9836ab2eb" containerName="northd" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.570400 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b193bd2-b536-4056-92f1-94c9836ab2eb" containerName="northd" Dec 11 02:14:07 crc kubenswrapper[4824]: E1211 02:14:07.570414 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b193bd2-b536-4056-92f1-94c9836ab2eb" containerName="ovn-controller" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.570425 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b193bd2-b536-4056-92f1-94c9836ab2eb" containerName="ovn-controller" Dec 11 02:14:07 crc kubenswrapper[4824]: E1211 02:14:07.570444 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b193bd2-b536-4056-92f1-94c9836ab2eb" containerName="nbdb" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.570453 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b193bd2-b536-4056-92f1-94c9836ab2eb" containerName="nbdb" Dec 11 02:14:07 crc kubenswrapper[4824]: E1211 02:14:07.570474 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b193bd2-b536-4056-92f1-94c9836ab2eb" containerName="kube-rbac-proxy-ovn-metrics" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.570488 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b193bd2-b536-4056-92f1-94c9836ab2eb" containerName="kube-rbac-proxy-ovn-metrics" Dec 11 02:14:07 crc kubenswrapper[4824]: E1211 02:14:07.570501 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b193bd2-b536-4056-92f1-94c9836ab2eb" containerName="ovnkube-controller" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.570597 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b193bd2-b536-4056-92f1-94c9836ab2eb" containerName="ovnkube-controller" Dec 11 02:14:07 crc kubenswrapper[4824]: E1211 02:14:07.570612 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b193bd2-b536-4056-92f1-94c9836ab2eb" containerName="ovnkube-controller" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.570628 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b193bd2-b536-4056-92f1-94c9836ab2eb" containerName="ovnkube-controller" Dec 11 02:14:07 crc kubenswrapper[4824]: E1211 02:14:07.570639 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b193bd2-b536-4056-92f1-94c9836ab2eb" containerName="sbdb" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.570648 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b193bd2-b536-4056-92f1-94c9836ab2eb" containerName="sbdb" Dec 11 02:14:07 crc kubenswrapper[4824]: E1211 02:14:07.570662 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b193bd2-b536-4056-92f1-94c9836ab2eb" containerName="kube-rbac-proxy-node" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.570671 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b193bd2-b536-4056-92f1-94c9836ab2eb" containerName="kube-rbac-proxy-node" Dec 11 02:14:07 crc kubenswrapper[4824]: E1211 02:14:07.570688 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b193bd2-b536-4056-92f1-94c9836ab2eb" containerName="ovn-acl-logging" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.570698 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b193bd2-b536-4056-92f1-94c9836ab2eb" containerName="ovn-acl-logging" Dec 11 02:14:07 crc kubenswrapper[4824]: E1211 02:14:07.570710 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b193bd2-b536-4056-92f1-94c9836ab2eb" containerName="kubecfg-setup" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.570720 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b193bd2-b536-4056-92f1-94c9836ab2eb" containerName="kubecfg-setup" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.571568 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b193bd2-b536-4056-92f1-94c9836ab2eb" containerName="ovnkube-controller" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.571595 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b193bd2-b536-4056-92f1-94c9836ab2eb" containerName="ovnkube-controller" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.571607 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b193bd2-b536-4056-92f1-94c9836ab2eb" containerName="northd" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.571635 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b193bd2-b536-4056-92f1-94c9836ab2eb" containerName="kube-rbac-proxy-node" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.571649 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b193bd2-b536-4056-92f1-94c9836ab2eb" containerName="ovn-acl-logging" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.571667 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b193bd2-b536-4056-92f1-94c9836ab2eb" containerName="ovn-controller" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.571682 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b193bd2-b536-4056-92f1-94c9836ab2eb" containerName="ovnkube-controller" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.571693 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b193bd2-b536-4056-92f1-94c9836ab2eb" containerName="kube-rbac-proxy-ovn-metrics" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.571704 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b193bd2-b536-4056-92f1-94c9836ab2eb" containerName="ovnkube-controller" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.571716 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b193bd2-b536-4056-92f1-94c9836ab2eb" containerName="sbdb" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.571736 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b193bd2-b536-4056-92f1-94c9836ab2eb" containerName="nbdb" Dec 11 02:14:07 crc kubenswrapper[4824]: E1211 02:14:07.572292 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b193bd2-b536-4056-92f1-94c9836ab2eb" containerName="ovnkube-controller" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.572305 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b193bd2-b536-4056-92f1-94c9836ab2eb" containerName="ovnkube-controller" Dec 11 02:14:07 crc kubenswrapper[4824]: E1211 02:14:07.572326 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b193bd2-b536-4056-92f1-94c9836ab2eb" containerName="ovnkube-controller" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.572336 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b193bd2-b536-4056-92f1-94c9836ab2eb" containerName="ovnkube-controller" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.573899 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b193bd2-b536-4056-92f1-94c9836ab2eb" containerName="ovnkube-controller" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.576521 4824 scope.go:117] "RemoveContainer" containerID="9ead58694813806d307bbbc6a7b3cf41730d1c059b231a3002684bf09c10c7f4" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.621942 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-4mrwp" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.622470 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-65ngd" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.658619 4824 scope.go:117] "RemoveContainer" containerID="4ed2ae40b1fe28111efc76d9d72345edecb5f3ea34522b238f3730812b29f2bd" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.659300 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-host-run-ovn-kubernetes\") pod \"5b193bd2-b536-4056-92f1-94c9836ab2eb\" (UID: \"5b193bd2-b536-4056-92f1-94c9836ab2eb\") " Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.659336 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-run-openvswitch\") pod \"5b193bd2-b536-4056-92f1-94c9836ab2eb\" (UID: \"5b193bd2-b536-4056-92f1-94c9836ab2eb\") " Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.659378 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-host-slash\") pod \"5b193bd2-b536-4056-92f1-94c9836ab2eb\" (UID: \"5b193bd2-b536-4056-92f1-94c9836ab2eb\") " Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.659393 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-host-var-lib-cni-networks-ovn-kubernetes\") pod \"5b193bd2-b536-4056-92f1-94c9836ab2eb\" (UID: \"5b193bd2-b536-4056-92f1-94c9836ab2eb\") " Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.659410 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-log-socket\") pod \"5b193bd2-b536-4056-92f1-94c9836ab2eb\" (UID: \"5b193bd2-b536-4056-92f1-94c9836ab2eb\") " Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.659448 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-systemd-units\") pod \"5b193bd2-b536-4056-92f1-94c9836ab2eb\" (UID: \"5b193bd2-b536-4056-92f1-94c9836ab2eb\") " Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.659468 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5b193bd2-b536-4056-92f1-94c9836ab2eb-ovn-node-metrics-cert\") pod \"5b193bd2-b536-4056-92f1-94c9836ab2eb\" (UID: \"5b193bd2-b536-4056-92f1-94c9836ab2eb\") " Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.659599 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-run-ovn\") pod \"5b193bd2-b536-4056-92f1-94c9836ab2eb\" (UID: \"5b193bd2-b536-4056-92f1-94c9836ab2eb\") " Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.659622 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/5b193bd2-b536-4056-92f1-94c9836ab2eb-ovnkube-script-lib\") pod \"5b193bd2-b536-4056-92f1-94c9836ab2eb\" (UID: \"5b193bd2-b536-4056-92f1-94c9836ab2eb\") " Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.659641 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-host-cni-netd\") pod \"5b193bd2-b536-4056-92f1-94c9836ab2eb\" (UID: \"5b193bd2-b536-4056-92f1-94c9836ab2eb\") " Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.659672 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-host-run-netns\") pod \"5b193bd2-b536-4056-92f1-94c9836ab2eb\" (UID: \"5b193bd2-b536-4056-92f1-94c9836ab2eb\") " Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.659692 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-var-lib-openvswitch\") pod \"5b193bd2-b536-4056-92f1-94c9836ab2eb\" (UID: \"5b193bd2-b536-4056-92f1-94c9836ab2eb\") " Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.659707 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5b193bd2-b536-4056-92f1-94c9836ab2eb-env-overrides\") pod \"5b193bd2-b536-4056-92f1-94c9836ab2eb\" (UID: \"5b193bd2-b536-4056-92f1-94c9836ab2eb\") " Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.659720 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-etc-openvswitch\") pod \"5b193bd2-b536-4056-92f1-94c9836ab2eb\" (UID: \"5b193bd2-b536-4056-92f1-94c9836ab2eb\") " Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.659753 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-host-cni-bin\") pod \"5b193bd2-b536-4056-92f1-94c9836ab2eb\" (UID: \"5b193bd2-b536-4056-92f1-94c9836ab2eb\") " Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.659779 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-host-kubelet\") pod \"5b193bd2-b536-4056-92f1-94c9836ab2eb\" (UID: \"5b193bd2-b536-4056-92f1-94c9836ab2eb\") " Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.659794 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-node-log\") pod \"5b193bd2-b536-4056-92f1-94c9836ab2eb\" (UID: \"5b193bd2-b536-4056-92f1-94c9836ab2eb\") " Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.659825 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-run-systemd\") pod \"5b193bd2-b536-4056-92f1-94c9836ab2eb\" (UID: \"5b193bd2-b536-4056-92f1-94c9836ab2eb\") " Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.659855 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5b193bd2-b536-4056-92f1-94c9836ab2eb-ovnkube-config\") pod \"5b193bd2-b536-4056-92f1-94c9836ab2eb\" (UID: \"5b193bd2-b536-4056-92f1-94c9836ab2eb\") " Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.659871 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qpqjk\" (UniqueName: \"kubernetes.io/projected/5b193bd2-b536-4056-92f1-94c9836ab2eb-kube-api-access-qpqjk\") pod \"5b193bd2-b536-4056-92f1-94c9836ab2eb\" (UID: \"5b193bd2-b536-4056-92f1-94c9836ab2eb\") " Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.660435 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "5b193bd2-b536-4056-92f1-94c9836ab2eb" (UID: "5b193bd2-b536-4056-92f1-94c9836ab2eb"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.660479 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "5b193bd2-b536-4056-92f1-94c9836ab2eb" (UID: "5b193bd2-b536-4056-92f1-94c9836ab2eb"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.660498 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-host-slash" (OuterVolumeSpecName: "host-slash") pod "5b193bd2-b536-4056-92f1-94c9836ab2eb" (UID: "5b193bd2-b536-4056-92f1-94c9836ab2eb"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.660513 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "5b193bd2-b536-4056-92f1-94c9836ab2eb" (UID: "5b193bd2-b536-4056-92f1-94c9836ab2eb"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.660546 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-log-socket" (OuterVolumeSpecName: "log-socket") pod "5b193bd2-b536-4056-92f1-94c9836ab2eb" (UID: "5b193bd2-b536-4056-92f1-94c9836ab2eb"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.660563 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "5b193bd2-b536-4056-92f1-94c9836ab2eb" (UID: "5b193bd2-b536-4056-92f1-94c9836ab2eb"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.662862 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "5b193bd2-b536-4056-92f1-94c9836ab2eb" (UID: "5b193bd2-b536-4056-92f1-94c9836ab2eb"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.662939 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "5b193bd2-b536-4056-92f1-94c9836ab2eb" (UID: "5b193bd2-b536-4056-92f1-94c9836ab2eb"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.663486 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5b193bd2-b536-4056-92f1-94c9836ab2eb-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "5b193bd2-b536-4056-92f1-94c9836ab2eb" (UID: "5b193bd2-b536-4056-92f1-94c9836ab2eb"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.663530 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "5b193bd2-b536-4056-92f1-94c9836ab2eb" (UID: "5b193bd2-b536-4056-92f1-94c9836ab2eb"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.663556 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "5b193bd2-b536-4056-92f1-94c9836ab2eb" (UID: "5b193bd2-b536-4056-92f1-94c9836ab2eb"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.663579 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "5b193bd2-b536-4056-92f1-94c9836ab2eb" (UID: "5b193bd2-b536-4056-92f1-94c9836ab2eb"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.664308 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5b193bd2-b536-4056-92f1-94c9836ab2eb-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "5b193bd2-b536-4056-92f1-94c9836ab2eb" (UID: "5b193bd2-b536-4056-92f1-94c9836ab2eb"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.664354 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "5b193bd2-b536-4056-92f1-94c9836ab2eb" (UID: "5b193bd2-b536-4056-92f1-94c9836ab2eb"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.665009 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5b193bd2-b536-4056-92f1-94c9836ab2eb-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "5b193bd2-b536-4056-92f1-94c9836ab2eb" (UID: "5b193bd2-b536-4056-92f1-94c9836ab2eb"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.665046 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "5b193bd2-b536-4056-92f1-94c9836ab2eb" (UID: "5b193bd2-b536-4056-92f1-94c9836ab2eb"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.665071 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-node-log" (OuterVolumeSpecName: "node-log") pod "5b193bd2-b536-4056-92f1-94c9836ab2eb" (UID: "5b193bd2-b536-4056-92f1-94c9836ab2eb"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.667207 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b193bd2-b536-4056-92f1-94c9836ab2eb-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "5b193bd2-b536-4056-92f1-94c9836ab2eb" (UID: "5b193bd2-b536-4056-92f1-94c9836ab2eb"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.670899 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b193bd2-b536-4056-92f1-94c9836ab2eb-kube-api-access-qpqjk" (OuterVolumeSpecName: "kube-api-access-qpqjk") pod "5b193bd2-b536-4056-92f1-94c9836ab2eb" (UID: "5b193bd2-b536-4056-92f1-94c9836ab2eb"). InnerVolumeSpecName "kube-api-access-qpqjk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.685877 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "5b193bd2-b536-4056-92f1-94c9836ab2eb" (UID: "5b193bd2-b536-4056-92f1-94c9836ab2eb"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.685891 4824 scope.go:117] "RemoveContainer" containerID="ce7aa078800e6f299cfe03770fa7e3b64fa3742fe0c0f383c87d84cf48569ddf" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.690627 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4mrwp"] Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.705399 4824 scope.go:117] "RemoveContainer" containerID="c345850891744a0240781277e9a89a59f6899c4c944995f62e44650f807440a1" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.723060 4824 scope.go:117] "RemoveContainer" containerID="51f3c41d86ddc44f7eca118029876ad353317f753d0ad173ed4d30c2a15f10e6" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.738739 4824 scope.go:117] "RemoveContainer" containerID="6aa9fc1f1bff192c69bd59d32a05d4a72d09641d9ce4b265bc4517147a7e15dd" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.750550 4824 scope.go:117] "RemoveContainer" containerID="4b4a3bc50eb80777188206adab774e8acad8b2e4952a8415f91b1c29ba9a0ffd" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.761276 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/6d025429-c91d-4528-84ba-498965f3883a-log-socket\") pod \"ovnkube-node-65ngd\" (UID: \"6d025429-c91d-4528-84ba-498965f3883a\") " pod="openshift-ovn-kubernetes/ovnkube-node-65ngd" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.761359 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6d025429-c91d-4528-84ba-498965f3883a-ovn-node-metrics-cert\") pod \"ovnkube-node-65ngd\" (UID: \"6d025429-c91d-4528-84ba-498965f3883a\") " pod="openshift-ovn-kubernetes/ovnkube-node-65ngd" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.761405 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6d025429-c91d-4528-84ba-498965f3883a-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-65ngd\" (UID: \"6d025429-c91d-4528-84ba-498965f3883a\") " pod="openshift-ovn-kubernetes/ovnkube-node-65ngd" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.761458 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/6d025429-c91d-4528-84ba-498965f3883a-systemd-units\") pod \"ovnkube-node-65ngd\" (UID: \"6d025429-c91d-4528-84ba-498965f3883a\") " pod="openshift-ovn-kubernetes/ovnkube-node-65ngd" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.761541 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/6d025429-c91d-4528-84ba-498965f3883a-host-kubelet\") pod \"ovnkube-node-65ngd\" (UID: \"6d025429-c91d-4528-84ba-498965f3883a\") " pod="openshift-ovn-kubernetes/ovnkube-node-65ngd" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.761570 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/6d025429-c91d-4528-84ba-498965f3883a-run-ovn\") pod \"ovnkube-node-65ngd\" (UID: \"6d025429-c91d-4528-84ba-498965f3883a\") " pod="openshift-ovn-kubernetes/ovnkube-node-65ngd" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.761616 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zwrnk\" (UniqueName: \"kubernetes.io/projected/6d025429-c91d-4528-84ba-498965f3883a-kube-api-access-zwrnk\") pod \"ovnkube-node-65ngd\" (UID: \"6d025429-c91d-4528-84ba-498965f3883a\") " pod="openshift-ovn-kubernetes/ovnkube-node-65ngd" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.761648 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6d025429-c91d-4528-84ba-498965f3883a-etc-openvswitch\") pod \"ovnkube-node-65ngd\" (UID: \"6d025429-c91d-4528-84ba-498965f3883a\") " pod="openshift-ovn-kubernetes/ovnkube-node-65ngd" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.761697 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/6d025429-c91d-4528-84ba-498965f3883a-host-cni-bin\") pod \"ovnkube-node-65ngd\" (UID: \"6d025429-c91d-4528-84ba-498965f3883a\") " pod="openshift-ovn-kubernetes/ovnkube-node-65ngd" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.761724 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/6d025429-c91d-4528-84ba-498965f3883a-node-log\") pod \"ovnkube-node-65ngd\" (UID: \"6d025429-c91d-4528-84ba-498965f3883a\") " pod="openshift-ovn-kubernetes/ovnkube-node-65ngd" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.761745 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6d025429-c91d-4528-84ba-498965f3883a-host-run-ovn-kubernetes\") pod \"ovnkube-node-65ngd\" (UID: \"6d025429-c91d-4528-84ba-498965f3883a\") " pod="openshift-ovn-kubernetes/ovnkube-node-65ngd" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.761768 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6d025429-c91d-4528-84ba-498965f3883a-run-openvswitch\") pod \"ovnkube-node-65ngd\" (UID: \"6d025429-c91d-4528-84ba-498965f3883a\") " pod="openshift-ovn-kubernetes/ovnkube-node-65ngd" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.761794 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/6d025429-c91d-4528-84ba-498965f3883a-host-slash\") pod \"ovnkube-node-65ngd\" (UID: \"6d025429-c91d-4528-84ba-498965f3883a\") " pod="openshift-ovn-kubernetes/ovnkube-node-65ngd" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.761835 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/6d025429-c91d-4528-84ba-498965f3883a-host-run-netns\") pod \"ovnkube-node-65ngd\" (UID: \"6d025429-c91d-4528-84ba-498965f3883a\") " pod="openshift-ovn-kubernetes/ovnkube-node-65ngd" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.761858 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6d025429-c91d-4528-84ba-498965f3883a-ovnkube-config\") pod \"ovnkube-node-65ngd\" (UID: \"6d025429-c91d-4528-84ba-498965f3883a\") " pod="openshift-ovn-kubernetes/ovnkube-node-65ngd" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.761882 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/6d025429-c91d-4528-84ba-498965f3883a-host-cni-netd\") pod \"ovnkube-node-65ngd\" (UID: \"6d025429-c91d-4528-84ba-498965f3883a\") " pod="openshift-ovn-kubernetes/ovnkube-node-65ngd" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.761902 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6d025429-c91d-4528-84ba-498965f3883a-ovnkube-script-lib\") pod \"ovnkube-node-65ngd\" (UID: \"6d025429-c91d-4528-84ba-498965f3883a\") " pod="openshift-ovn-kubernetes/ovnkube-node-65ngd" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.761923 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/6d025429-c91d-4528-84ba-498965f3883a-run-systemd\") pod \"ovnkube-node-65ngd\" (UID: \"6d025429-c91d-4528-84ba-498965f3883a\") " pod="openshift-ovn-kubernetes/ovnkube-node-65ngd" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.761966 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6d025429-c91d-4528-84ba-498965f3883a-var-lib-openvswitch\") pod \"ovnkube-node-65ngd\" (UID: \"6d025429-c91d-4528-84ba-498965f3883a\") " pod="openshift-ovn-kubernetes/ovnkube-node-65ngd" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.762009 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6d025429-c91d-4528-84ba-498965f3883a-env-overrides\") pod \"ovnkube-node-65ngd\" (UID: \"6d025429-c91d-4528-84ba-498965f3883a\") " pod="openshift-ovn-kubernetes/ovnkube-node-65ngd" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.762075 4824 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.762095 4824 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-host-slash\") on node \"crc\" DevicePath \"\"" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.762125 4824 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-log-socket\") on node \"crc\" DevicePath \"\"" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.762138 4824 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-systemd-units\") on node \"crc\" DevicePath \"\"" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.762150 4824 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5b193bd2-b536-4056-92f1-94c9836ab2eb-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.762163 4824 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.762174 4824 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/5b193bd2-b536-4056-92f1-94c9836ab2eb-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.762186 4824 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-host-cni-netd\") on node \"crc\" DevicePath \"\"" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.762199 4824 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-host-run-netns\") on node \"crc\" DevicePath \"\"" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.762212 4824 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.762225 4824 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5b193bd2-b536-4056-92f1-94c9836ab2eb-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.762237 4824 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.762249 4824 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-host-cni-bin\") on node \"crc\" DevicePath \"\"" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.762260 4824 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-host-kubelet\") on node \"crc\" DevicePath \"\"" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.762272 4824 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-node-log\") on node \"crc\" DevicePath \"\"" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.762283 4824 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-run-systemd\") on node \"crc\" DevicePath \"\"" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.762294 4824 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5b193bd2-b536-4056-92f1-94c9836ab2eb-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.762307 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qpqjk\" (UniqueName: \"kubernetes.io/projected/5b193bd2-b536-4056-92f1-94c9836ab2eb-kube-api-access-qpqjk\") on node \"crc\" DevicePath \"\"" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.762320 4824 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.762332 4824 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5b193bd2-b536-4056-92f1-94c9836ab2eb-run-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.763694 4824 scope.go:117] "RemoveContainer" containerID="4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.777032 4824 scope.go:117] "RemoveContainer" containerID="8bb86ae130147feda9a5af751733489c3e0cc293e47e555879ae1fe0392c640e" Dec 11 02:14:07 crc kubenswrapper[4824]: E1211 02:14:07.777344 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8bb86ae130147feda9a5af751733489c3e0cc293e47e555879ae1fe0392c640e\": container with ID starting with 8bb86ae130147feda9a5af751733489c3e0cc293e47e555879ae1fe0392c640e not found: ID does not exist" containerID="8bb86ae130147feda9a5af751733489c3e0cc293e47e555879ae1fe0392c640e" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.777371 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8bb86ae130147feda9a5af751733489c3e0cc293e47e555879ae1fe0392c640e"} err="failed to get container status \"8bb86ae130147feda9a5af751733489c3e0cc293e47e555879ae1fe0392c640e\": rpc error: code = NotFound desc = could not find container \"8bb86ae130147feda9a5af751733489c3e0cc293e47e555879ae1fe0392c640e\": container with ID starting with 8bb86ae130147feda9a5af751733489c3e0cc293e47e555879ae1fe0392c640e not found: ID does not exist" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.777392 4824 scope.go:117] "RemoveContainer" containerID="154d82fd6acee749bfd37220547e52647dc7f9d86a73c28ce7d8577c9cadef6e" Dec 11 02:14:07 crc kubenswrapper[4824]: E1211 02:14:07.777669 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"154d82fd6acee749bfd37220547e52647dc7f9d86a73c28ce7d8577c9cadef6e\": container with ID starting with 154d82fd6acee749bfd37220547e52647dc7f9d86a73c28ce7d8577c9cadef6e not found: ID does not exist" containerID="154d82fd6acee749bfd37220547e52647dc7f9d86a73c28ce7d8577c9cadef6e" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.777695 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"154d82fd6acee749bfd37220547e52647dc7f9d86a73c28ce7d8577c9cadef6e"} err="failed to get container status \"154d82fd6acee749bfd37220547e52647dc7f9d86a73c28ce7d8577c9cadef6e\": rpc error: code = NotFound desc = could not find container \"154d82fd6acee749bfd37220547e52647dc7f9d86a73c28ce7d8577c9cadef6e\": container with ID starting with 154d82fd6acee749bfd37220547e52647dc7f9d86a73c28ce7d8577c9cadef6e not found: ID does not exist" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.777710 4824 scope.go:117] "RemoveContainer" containerID="9ead58694813806d307bbbc6a7b3cf41730d1c059b231a3002684bf09c10c7f4" Dec 11 02:14:07 crc kubenswrapper[4824]: E1211 02:14:07.777960 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ead58694813806d307bbbc6a7b3cf41730d1c059b231a3002684bf09c10c7f4\": container with ID starting with 9ead58694813806d307bbbc6a7b3cf41730d1c059b231a3002684bf09c10c7f4 not found: ID does not exist" containerID="9ead58694813806d307bbbc6a7b3cf41730d1c059b231a3002684bf09c10c7f4" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.778002 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ead58694813806d307bbbc6a7b3cf41730d1c059b231a3002684bf09c10c7f4"} err="failed to get container status \"9ead58694813806d307bbbc6a7b3cf41730d1c059b231a3002684bf09c10c7f4\": rpc error: code = NotFound desc = could not find container \"9ead58694813806d307bbbc6a7b3cf41730d1c059b231a3002684bf09c10c7f4\": container with ID starting with 9ead58694813806d307bbbc6a7b3cf41730d1c059b231a3002684bf09c10c7f4 not found: ID does not exist" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.778030 4824 scope.go:117] "RemoveContainer" containerID="4ed2ae40b1fe28111efc76d9d72345edecb5f3ea34522b238f3730812b29f2bd" Dec 11 02:14:07 crc kubenswrapper[4824]: E1211 02:14:07.778403 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4ed2ae40b1fe28111efc76d9d72345edecb5f3ea34522b238f3730812b29f2bd\": container with ID starting with 4ed2ae40b1fe28111efc76d9d72345edecb5f3ea34522b238f3730812b29f2bd not found: ID does not exist" containerID="4ed2ae40b1fe28111efc76d9d72345edecb5f3ea34522b238f3730812b29f2bd" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.778427 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ed2ae40b1fe28111efc76d9d72345edecb5f3ea34522b238f3730812b29f2bd"} err="failed to get container status \"4ed2ae40b1fe28111efc76d9d72345edecb5f3ea34522b238f3730812b29f2bd\": rpc error: code = NotFound desc = could not find container \"4ed2ae40b1fe28111efc76d9d72345edecb5f3ea34522b238f3730812b29f2bd\": container with ID starting with 4ed2ae40b1fe28111efc76d9d72345edecb5f3ea34522b238f3730812b29f2bd not found: ID does not exist" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.778443 4824 scope.go:117] "RemoveContainer" containerID="ce7aa078800e6f299cfe03770fa7e3b64fa3742fe0c0f383c87d84cf48569ddf" Dec 11 02:14:07 crc kubenswrapper[4824]: E1211 02:14:07.778672 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce7aa078800e6f299cfe03770fa7e3b64fa3742fe0c0f383c87d84cf48569ddf\": container with ID starting with ce7aa078800e6f299cfe03770fa7e3b64fa3742fe0c0f383c87d84cf48569ddf not found: ID does not exist" containerID="ce7aa078800e6f299cfe03770fa7e3b64fa3742fe0c0f383c87d84cf48569ddf" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.778697 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce7aa078800e6f299cfe03770fa7e3b64fa3742fe0c0f383c87d84cf48569ddf"} err="failed to get container status \"ce7aa078800e6f299cfe03770fa7e3b64fa3742fe0c0f383c87d84cf48569ddf\": rpc error: code = NotFound desc = could not find container \"ce7aa078800e6f299cfe03770fa7e3b64fa3742fe0c0f383c87d84cf48569ddf\": container with ID starting with ce7aa078800e6f299cfe03770fa7e3b64fa3742fe0c0f383c87d84cf48569ddf not found: ID does not exist" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.778712 4824 scope.go:117] "RemoveContainer" containerID="c345850891744a0240781277e9a89a59f6899c4c944995f62e44650f807440a1" Dec 11 02:14:07 crc kubenswrapper[4824]: E1211 02:14:07.778982 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c345850891744a0240781277e9a89a59f6899c4c944995f62e44650f807440a1\": container with ID starting with c345850891744a0240781277e9a89a59f6899c4c944995f62e44650f807440a1 not found: ID does not exist" containerID="c345850891744a0240781277e9a89a59f6899c4c944995f62e44650f807440a1" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.779005 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c345850891744a0240781277e9a89a59f6899c4c944995f62e44650f807440a1"} err="failed to get container status \"c345850891744a0240781277e9a89a59f6899c4c944995f62e44650f807440a1\": rpc error: code = NotFound desc = could not find container \"c345850891744a0240781277e9a89a59f6899c4c944995f62e44650f807440a1\": container with ID starting with c345850891744a0240781277e9a89a59f6899c4c944995f62e44650f807440a1 not found: ID does not exist" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.779019 4824 scope.go:117] "RemoveContainer" containerID="51f3c41d86ddc44f7eca118029876ad353317f753d0ad173ed4d30c2a15f10e6" Dec 11 02:14:07 crc kubenswrapper[4824]: E1211 02:14:07.779206 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"51f3c41d86ddc44f7eca118029876ad353317f753d0ad173ed4d30c2a15f10e6\": container with ID starting with 51f3c41d86ddc44f7eca118029876ad353317f753d0ad173ed4d30c2a15f10e6 not found: ID does not exist" containerID="51f3c41d86ddc44f7eca118029876ad353317f753d0ad173ed4d30c2a15f10e6" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.779227 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"51f3c41d86ddc44f7eca118029876ad353317f753d0ad173ed4d30c2a15f10e6"} err="failed to get container status \"51f3c41d86ddc44f7eca118029876ad353317f753d0ad173ed4d30c2a15f10e6\": rpc error: code = NotFound desc = could not find container \"51f3c41d86ddc44f7eca118029876ad353317f753d0ad173ed4d30c2a15f10e6\": container with ID starting with 51f3c41d86ddc44f7eca118029876ad353317f753d0ad173ed4d30c2a15f10e6 not found: ID does not exist" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.779240 4824 scope.go:117] "RemoveContainer" containerID="6aa9fc1f1bff192c69bd59d32a05d4a72d09641d9ce4b265bc4517147a7e15dd" Dec 11 02:14:07 crc kubenswrapper[4824]: E1211 02:14:07.779431 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6aa9fc1f1bff192c69bd59d32a05d4a72d09641d9ce4b265bc4517147a7e15dd\": container with ID starting with 6aa9fc1f1bff192c69bd59d32a05d4a72d09641d9ce4b265bc4517147a7e15dd not found: ID does not exist" containerID="6aa9fc1f1bff192c69bd59d32a05d4a72d09641d9ce4b265bc4517147a7e15dd" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.779449 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6aa9fc1f1bff192c69bd59d32a05d4a72d09641d9ce4b265bc4517147a7e15dd"} err="failed to get container status \"6aa9fc1f1bff192c69bd59d32a05d4a72d09641d9ce4b265bc4517147a7e15dd\": rpc error: code = NotFound desc = could not find container \"6aa9fc1f1bff192c69bd59d32a05d4a72d09641d9ce4b265bc4517147a7e15dd\": container with ID starting with 6aa9fc1f1bff192c69bd59d32a05d4a72d09641d9ce4b265bc4517147a7e15dd not found: ID does not exist" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.779462 4824 scope.go:117] "RemoveContainer" containerID="4b4a3bc50eb80777188206adab774e8acad8b2e4952a8415f91b1c29ba9a0ffd" Dec 11 02:14:07 crc kubenswrapper[4824]: E1211 02:14:07.779932 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4b4a3bc50eb80777188206adab774e8acad8b2e4952a8415f91b1c29ba9a0ffd\": container with ID starting with 4b4a3bc50eb80777188206adab774e8acad8b2e4952a8415f91b1c29ba9a0ffd not found: ID does not exist" containerID="4b4a3bc50eb80777188206adab774e8acad8b2e4952a8415f91b1c29ba9a0ffd" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.779950 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b4a3bc50eb80777188206adab774e8acad8b2e4952a8415f91b1c29ba9a0ffd"} err="failed to get container status \"4b4a3bc50eb80777188206adab774e8acad8b2e4952a8415f91b1c29ba9a0ffd\": rpc error: code = NotFound desc = could not find container \"4b4a3bc50eb80777188206adab774e8acad8b2e4952a8415f91b1c29ba9a0ffd\": container with ID starting with 4b4a3bc50eb80777188206adab774e8acad8b2e4952a8415f91b1c29ba9a0ffd not found: ID does not exist" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.779962 4824 scope.go:117] "RemoveContainer" containerID="4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df" Dec 11 02:14:07 crc kubenswrapper[4824]: E1211 02:14:07.780205 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df\": container with ID starting with 4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df not found: ID does not exist" containerID="4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.780224 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df"} err="failed to get container status \"4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df\": rpc error: code = NotFound desc = could not find container \"4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df\": container with ID starting with 4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df not found: ID does not exist" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.780237 4824 scope.go:117] "RemoveContainer" containerID="8bb86ae130147feda9a5af751733489c3e0cc293e47e555879ae1fe0392c640e" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.780495 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8bb86ae130147feda9a5af751733489c3e0cc293e47e555879ae1fe0392c640e"} err="failed to get container status \"8bb86ae130147feda9a5af751733489c3e0cc293e47e555879ae1fe0392c640e\": rpc error: code = NotFound desc = could not find container \"8bb86ae130147feda9a5af751733489c3e0cc293e47e555879ae1fe0392c640e\": container with ID starting with 8bb86ae130147feda9a5af751733489c3e0cc293e47e555879ae1fe0392c640e not found: ID does not exist" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.780513 4824 scope.go:117] "RemoveContainer" containerID="154d82fd6acee749bfd37220547e52647dc7f9d86a73c28ce7d8577c9cadef6e" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.780684 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"154d82fd6acee749bfd37220547e52647dc7f9d86a73c28ce7d8577c9cadef6e"} err="failed to get container status \"154d82fd6acee749bfd37220547e52647dc7f9d86a73c28ce7d8577c9cadef6e\": rpc error: code = NotFound desc = could not find container \"154d82fd6acee749bfd37220547e52647dc7f9d86a73c28ce7d8577c9cadef6e\": container with ID starting with 154d82fd6acee749bfd37220547e52647dc7f9d86a73c28ce7d8577c9cadef6e not found: ID does not exist" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.780699 4824 scope.go:117] "RemoveContainer" containerID="9ead58694813806d307bbbc6a7b3cf41730d1c059b231a3002684bf09c10c7f4" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.780937 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ead58694813806d307bbbc6a7b3cf41730d1c059b231a3002684bf09c10c7f4"} err="failed to get container status \"9ead58694813806d307bbbc6a7b3cf41730d1c059b231a3002684bf09c10c7f4\": rpc error: code = NotFound desc = could not find container \"9ead58694813806d307bbbc6a7b3cf41730d1c059b231a3002684bf09c10c7f4\": container with ID starting with 9ead58694813806d307bbbc6a7b3cf41730d1c059b231a3002684bf09c10c7f4 not found: ID does not exist" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.780952 4824 scope.go:117] "RemoveContainer" containerID="4ed2ae40b1fe28111efc76d9d72345edecb5f3ea34522b238f3730812b29f2bd" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.781337 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ed2ae40b1fe28111efc76d9d72345edecb5f3ea34522b238f3730812b29f2bd"} err="failed to get container status \"4ed2ae40b1fe28111efc76d9d72345edecb5f3ea34522b238f3730812b29f2bd\": rpc error: code = NotFound desc = could not find container \"4ed2ae40b1fe28111efc76d9d72345edecb5f3ea34522b238f3730812b29f2bd\": container with ID starting with 4ed2ae40b1fe28111efc76d9d72345edecb5f3ea34522b238f3730812b29f2bd not found: ID does not exist" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.781356 4824 scope.go:117] "RemoveContainer" containerID="ce7aa078800e6f299cfe03770fa7e3b64fa3742fe0c0f383c87d84cf48569ddf" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.781589 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce7aa078800e6f299cfe03770fa7e3b64fa3742fe0c0f383c87d84cf48569ddf"} err="failed to get container status \"ce7aa078800e6f299cfe03770fa7e3b64fa3742fe0c0f383c87d84cf48569ddf\": rpc error: code = NotFound desc = could not find container \"ce7aa078800e6f299cfe03770fa7e3b64fa3742fe0c0f383c87d84cf48569ddf\": container with ID starting with ce7aa078800e6f299cfe03770fa7e3b64fa3742fe0c0f383c87d84cf48569ddf not found: ID does not exist" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.781606 4824 scope.go:117] "RemoveContainer" containerID="c345850891744a0240781277e9a89a59f6899c4c944995f62e44650f807440a1" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.781824 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c345850891744a0240781277e9a89a59f6899c4c944995f62e44650f807440a1"} err="failed to get container status \"c345850891744a0240781277e9a89a59f6899c4c944995f62e44650f807440a1\": rpc error: code = NotFound desc = could not find container \"c345850891744a0240781277e9a89a59f6899c4c944995f62e44650f807440a1\": container with ID starting with c345850891744a0240781277e9a89a59f6899c4c944995f62e44650f807440a1 not found: ID does not exist" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.781844 4824 scope.go:117] "RemoveContainer" containerID="51f3c41d86ddc44f7eca118029876ad353317f753d0ad173ed4d30c2a15f10e6" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.782097 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"51f3c41d86ddc44f7eca118029876ad353317f753d0ad173ed4d30c2a15f10e6"} err="failed to get container status \"51f3c41d86ddc44f7eca118029876ad353317f753d0ad173ed4d30c2a15f10e6\": rpc error: code = NotFound desc = could not find container \"51f3c41d86ddc44f7eca118029876ad353317f753d0ad173ed4d30c2a15f10e6\": container with ID starting with 51f3c41d86ddc44f7eca118029876ad353317f753d0ad173ed4d30c2a15f10e6 not found: ID does not exist" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.782133 4824 scope.go:117] "RemoveContainer" containerID="6aa9fc1f1bff192c69bd59d32a05d4a72d09641d9ce4b265bc4517147a7e15dd" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.782360 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6aa9fc1f1bff192c69bd59d32a05d4a72d09641d9ce4b265bc4517147a7e15dd"} err="failed to get container status \"6aa9fc1f1bff192c69bd59d32a05d4a72d09641d9ce4b265bc4517147a7e15dd\": rpc error: code = NotFound desc = could not find container \"6aa9fc1f1bff192c69bd59d32a05d4a72d09641d9ce4b265bc4517147a7e15dd\": container with ID starting with 6aa9fc1f1bff192c69bd59d32a05d4a72d09641d9ce4b265bc4517147a7e15dd not found: ID does not exist" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.782377 4824 scope.go:117] "RemoveContainer" containerID="4b4a3bc50eb80777188206adab774e8acad8b2e4952a8415f91b1c29ba9a0ffd" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.782627 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b4a3bc50eb80777188206adab774e8acad8b2e4952a8415f91b1c29ba9a0ffd"} err="failed to get container status \"4b4a3bc50eb80777188206adab774e8acad8b2e4952a8415f91b1c29ba9a0ffd\": rpc error: code = NotFound desc = could not find container \"4b4a3bc50eb80777188206adab774e8acad8b2e4952a8415f91b1c29ba9a0ffd\": container with ID starting with 4b4a3bc50eb80777188206adab774e8acad8b2e4952a8415f91b1c29ba9a0ffd not found: ID does not exist" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.782643 4824 scope.go:117] "RemoveContainer" containerID="4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.782860 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df"} err="failed to get container status \"4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df\": rpc error: code = NotFound desc = could not find container \"4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df\": container with ID starting with 4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df not found: ID does not exist" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.782882 4824 scope.go:117] "RemoveContainer" containerID="8bb86ae130147feda9a5af751733489c3e0cc293e47e555879ae1fe0392c640e" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.783155 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8bb86ae130147feda9a5af751733489c3e0cc293e47e555879ae1fe0392c640e"} err="failed to get container status \"8bb86ae130147feda9a5af751733489c3e0cc293e47e555879ae1fe0392c640e\": rpc error: code = NotFound desc = could not find container \"8bb86ae130147feda9a5af751733489c3e0cc293e47e555879ae1fe0392c640e\": container with ID starting with 8bb86ae130147feda9a5af751733489c3e0cc293e47e555879ae1fe0392c640e not found: ID does not exist" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.783171 4824 scope.go:117] "RemoveContainer" containerID="154d82fd6acee749bfd37220547e52647dc7f9d86a73c28ce7d8577c9cadef6e" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.783372 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"154d82fd6acee749bfd37220547e52647dc7f9d86a73c28ce7d8577c9cadef6e"} err="failed to get container status \"154d82fd6acee749bfd37220547e52647dc7f9d86a73c28ce7d8577c9cadef6e\": rpc error: code = NotFound desc = could not find container \"154d82fd6acee749bfd37220547e52647dc7f9d86a73c28ce7d8577c9cadef6e\": container with ID starting with 154d82fd6acee749bfd37220547e52647dc7f9d86a73c28ce7d8577c9cadef6e not found: ID does not exist" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.783390 4824 scope.go:117] "RemoveContainer" containerID="9ead58694813806d307bbbc6a7b3cf41730d1c059b231a3002684bf09c10c7f4" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.783648 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ead58694813806d307bbbc6a7b3cf41730d1c059b231a3002684bf09c10c7f4"} err="failed to get container status \"9ead58694813806d307bbbc6a7b3cf41730d1c059b231a3002684bf09c10c7f4\": rpc error: code = NotFound desc = could not find container \"9ead58694813806d307bbbc6a7b3cf41730d1c059b231a3002684bf09c10c7f4\": container with ID starting with 9ead58694813806d307bbbc6a7b3cf41730d1c059b231a3002684bf09c10c7f4 not found: ID does not exist" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.783666 4824 scope.go:117] "RemoveContainer" containerID="4ed2ae40b1fe28111efc76d9d72345edecb5f3ea34522b238f3730812b29f2bd" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.783864 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ed2ae40b1fe28111efc76d9d72345edecb5f3ea34522b238f3730812b29f2bd"} err="failed to get container status \"4ed2ae40b1fe28111efc76d9d72345edecb5f3ea34522b238f3730812b29f2bd\": rpc error: code = NotFound desc = could not find container \"4ed2ae40b1fe28111efc76d9d72345edecb5f3ea34522b238f3730812b29f2bd\": container with ID starting with 4ed2ae40b1fe28111efc76d9d72345edecb5f3ea34522b238f3730812b29f2bd not found: ID does not exist" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.783885 4824 scope.go:117] "RemoveContainer" containerID="ce7aa078800e6f299cfe03770fa7e3b64fa3742fe0c0f383c87d84cf48569ddf" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.784076 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce7aa078800e6f299cfe03770fa7e3b64fa3742fe0c0f383c87d84cf48569ddf"} err="failed to get container status \"ce7aa078800e6f299cfe03770fa7e3b64fa3742fe0c0f383c87d84cf48569ddf\": rpc error: code = NotFound desc = could not find container \"ce7aa078800e6f299cfe03770fa7e3b64fa3742fe0c0f383c87d84cf48569ddf\": container with ID starting with ce7aa078800e6f299cfe03770fa7e3b64fa3742fe0c0f383c87d84cf48569ddf not found: ID does not exist" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.784094 4824 scope.go:117] "RemoveContainer" containerID="c345850891744a0240781277e9a89a59f6899c4c944995f62e44650f807440a1" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.784355 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c345850891744a0240781277e9a89a59f6899c4c944995f62e44650f807440a1"} err="failed to get container status \"c345850891744a0240781277e9a89a59f6899c4c944995f62e44650f807440a1\": rpc error: code = NotFound desc = could not find container \"c345850891744a0240781277e9a89a59f6899c4c944995f62e44650f807440a1\": container with ID starting with c345850891744a0240781277e9a89a59f6899c4c944995f62e44650f807440a1 not found: ID does not exist" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.784368 4824 scope.go:117] "RemoveContainer" containerID="51f3c41d86ddc44f7eca118029876ad353317f753d0ad173ed4d30c2a15f10e6" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.784883 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"51f3c41d86ddc44f7eca118029876ad353317f753d0ad173ed4d30c2a15f10e6"} err="failed to get container status \"51f3c41d86ddc44f7eca118029876ad353317f753d0ad173ed4d30c2a15f10e6\": rpc error: code = NotFound desc = could not find container \"51f3c41d86ddc44f7eca118029876ad353317f753d0ad173ed4d30c2a15f10e6\": container with ID starting with 51f3c41d86ddc44f7eca118029876ad353317f753d0ad173ed4d30c2a15f10e6 not found: ID does not exist" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.784900 4824 scope.go:117] "RemoveContainer" containerID="6aa9fc1f1bff192c69bd59d32a05d4a72d09641d9ce4b265bc4517147a7e15dd" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.785162 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6aa9fc1f1bff192c69bd59d32a05d4a72d09641d9ce4b265bc4517147a7e15dd"} err="failed to get container status \"6aa9fc1f1bff192c69bd59d32a05d4a72d09641d9ce4b265bc4517147a7e15dd\": rpc error: code = NotFound desc = could not find container \"6aa9fc1f1bff192c69bd59d32a05d4a72d09641d9ce4b265bc4517147a7e15dd\": container with ID starting with 6aa9fc1f1bff192c69bd59d32a05d4a72d09641d9ce4b265bc4517147a7e15dd not found: ID does not exist" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.785176 4824 scope.go:117] "RemoveContainer" containerID="4b4a3bc50eb80777188206adab774e8acad8b2e4952a8415f91b1c29ba9a0ffd" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.785702 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b4a3bc50eb80777188206adab774e8acad8b2e4952a8415f91b1c29ba9a0ffd"} err="failed to get container status \"4b4a3bc50eb80777188206adab774e8acad8b2e4952a8415f91b1c29ba9a0ffd\": rpc error: code = NotFound desc = could not find container \"4b4a3bc50eb80777188206adab774e8acad8b2e4952a8415f91b1c29ba9a0ffd\": container with ID starting with 4b4a3bc50eb80777188206adab774e8acad8b2e4952a8415f91b1c29ba9a0ffd not found: ID does not exist" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.785723 4824 scope.go:117] "RemoveContainer" containerID="4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.786089 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df"} err="failed to get container status \"4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df\": rpc error: code = NotFound desc = could not find container \"4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df\": container with ID starting with 4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df not found: ID does not exist" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.786130 4824 scope.go:117] "RemoveContainer" containerID="8bb86ae130147feda9a5af751733489c3e0cc293e47e555879ae1fe0392c640e" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.786465 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8bb86ae130147feda9a5af751733489c3e0cc293e47e555879ae1fe0392c640e"} err="failed to get container status \"8bb86ae130147feda9a5af751733489c3e0cc293e47e555879ae1fe0392c640e\": rpc error: code = NotFound desc = could not find container \"8bb86ae130147feda9a5af751733489c3e0cc293e47e555879ae1fe0392c640e\": container with ID starting with 8bb86ae130147feda9a5af751733489c3e0cc293e47e555879ae1fe0392c640e not found: ID does not exist" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.786481 4824 scope.go:117] "RemoveContainer" containerID="154d82fd6acee749bfd37220547e52647dc7f9d86a73c28ce7d8577c9cadef6e" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.786744 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"154d82fd6acee749bfd37220547e52647dc7f9d86a73c28ce7d8577c9cadef6e"} err="failed to get container status \"154d82fd6acee749bfd37220547e52647dc7f9d86a73c28ce7d8577c9cadef6e\": rpc error: code = NotFound desc = could not find container \"154d82fd6acee749bfd37220547e52647dc7f9d86a73c28ce7d8577c9cadef6e\": container with ID starting with 154d82fd6acee749bfd37220547e52647dc7f9d86a73c28ce7d8577c9cadef6e not found: ID does not exist" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.786761 4824 scope.go:117] "RemoveContainer" containerID="9ead58694813806d307bbbc6a7b3cf41730d1c059b231a3002684bf09c10c7f4" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.787190 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ead58694813806d307bbbc6a7b3cf41730d1c059b231a3002684bf09c10c7f4"} err="failed to get container status \"9ead58694813806d307bbbc6a7b3cf41730d1c059b231a3002684bf09c10c7f4\": rpc error: code = NotFound desc = could not find container \"9ead58694813806d307bbbc6a7b3cf41730d1c059b231a3002684bf09c10c7f4\": container with ID starting with 9ead58694813806d307bbbc6a7b3cf41730d1c059b231a3002684bf09c10c7f4 not found: ID does not exist" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.787205 4824 scope.go:117] "RemoveContainer" containerID="4ed2ae40b1fe28111efc76d9d72345edecb5f3ea34522b238f3730812b29f2bd" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.787524 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ed2ae40b1fe28111efc76d9d72345edecb5f3ea34522b238f3730812b29f2bd"} err="failed to get container status \"4ed2ae40b1fe28111efc76d9d72345edecb5f3ea34522b238f3730812b29f2bd\": rpc error: code = NotFound desc = could not find container \"4ed2ae40b1fe28111efc76d9d72345edecb5f3ea34522b238f3730812b29f2bd\": container with ID starting with 4ed2ae40b1fe28111efc76d9d72345edecb5f3ea34522b238f3730812b29f2bd not found: ID does not exist" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.787882 4824 scope.go:117] "RemoveContainer" containerID="ce7aa078800e6f299cfe03770fa7e3b64fa3742fe0c0f383c87d84cf48569ddf" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.788123 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce7aa078800e6f299cfe03770fa7e3b64fa3742fe0c0f383c87d84cf48569ddf"} err="failed to get container status \"ce7aa078800e6f299cfe03770fa7e3b64fa3742fe0c0f383c87d84cf48569ddf\": rpc error: code = NotFound desc = could not find container \"ce7aa078800e6f299cfe03770fa7e3b64fa3742fe0c0f383c87d84cf48569ddf\": container with ID starting with ce7aa078800e6f299cfe03770fa7e3b64fa3742fe0c0f383c87d84cf48569ddf not found: ID does not exist" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.788139 4824 scope.go:117] "RemoveContainer" containerID="c345850891744a0240781277e9a89a59f6899c4c944995f62e44650f807440a1" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.788435 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c345850891744a0240781277e9a89a59f6899c4c944995f62e44650f807440a1"} err="failed to get container status \"c345850891744a0240781277e9a89a59f6899c4c944995f62e44650f807440a1\": rpc error: code = NotFound desc = could not find container \"c345850891744a0240781277e9a89a59f6899c4c944995f62e44650f807440a1\": container with ID starting with c345850891744a0240781277e9a89a59f6899c4c944995f62e44650f807440a1 not found: ID does not exist" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.788451 4824 scope.go:117] "RemoveContainer" containerID="51f3c41d86ddc44f7eca118029876ad353317f753d0ad173ed4d30c2a15f10e6" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.788681 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"51f3c41d86ddc44f7eca118029876ad353317f753d0ad173ed4d30c2a15f10e6"} err="failed to get container status \"51f3c41d86ddc44f7eca118029876ad353317f753d0ad173ed4d30c2a15f10e6\": rpc error: code = NotFound desc = could not find container \"51f3c41d86ddc44f7eca118029876ad353317f753d0ad173ed4d30c2a15f10e6\": container with ID starting with 51f3c41d86ddc44f7eca118029876ad353317f753d0ad173ed4d30c2a15f10e6 not found: ID does not exist" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.788700 4824 scope.go:117] "RemoveContainer" containerID="6aa9fc1f1bff192c69bd59d32a05d4a72d09641d9ce4b265bc4517147a7e15dd" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.788890 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6aa9fc1f1bff192c69bd59d32a05d4a72d09641d9ce4b265bc4517147a7e15dd"} err="failed to get container status \"6aa9fc1f1bff192c69bd59d32a05d4a72d09641d9ce4b265bc4517147a7e15dd\": rpc error: code = NotFound desc = could not find container \"6aa9fc1f1bff192c69bd59d32a05d4a72d09641d9ce4b265bc4517147a7e15dd\": container with ID starting with 6aa9fc1f1bff192c69bd59d32a05d4a72d09641d9ce4b265bc4517147a7e15dd not found: ID does not exist" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.788910 4824 scope.go:117] "RemoveContainer" containerID="4b4a3bc50eb80777188206adab774e8acad8b2e4952a8415f91b1c29ba9a0ffd" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.789233 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b4a3bc50eb80777188206adab774e8acad8b2e4952a8415f91b1c29ba9a0ffd"} err="failed to get container status \"4b4a3bc50eb80777188206adab774e8acad8b2e4952a8415f91b1c29ba9a0ffd\": rpc error: code = NotFound desc = could not find container \"4b4a3bc50eb80777188206adab774e8acad8b2e4952a8415f91b1c29ba9a0ffd\": container with ID starting with 4b4a3bc50eb80777188206adab774e8acad8b2e4952a8415f91b1c29ba9a0ffd not found: ID does not exist" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.789252 4824 scope.go:117] "RemoveContainer" containerID="4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.789552 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df"} err="failed to get container status \"4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df\": rpc error: code = NotFound desc = could not find container \"4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df\": container with ID starting with 4603cf1937408f9491fbd4cb159b2d15c8e07f91e2ba77c8f0aca69b2ddc84df not found: ID does not exist" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.789575 4824 scope.go:117] "RemoveContainer" containerID="8bb86ae130147feda9a5af751733489c3e0cc293e47e555879ae1fe0392c640e" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.789828 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8bb86ae130147feda9a5af751733489c3e0cc293e47e555879ae1fe0392c640e"} err="failed to get container status \"8bb86ae130147feda9a5af751733489c3e0cc293e47e555879ae1fe0392c640e\": rpc error: code = NotFound desc = could not find container \"8bb86ae130147feda9a5af751733489c3e0cc293e47e555879ae1fe0392c640e\": container with ID starting with 8bb86ae130147feda9a5af751733489c3e0cc293e47e555879ae1fe0392c640e not found: ID does not exist" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.863492 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6d025429-c91d-4528-84ba-498965f3883a-var-lib-openvswitch\") pod \"ovnkube-node-65ngd\" (UID: \"6d025429-c91d-4528-84ba-498965f3883a\") " pod="openshift-ovn-kubernetes/ovnkube-node-65ngd" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.863539 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6d025429-c91d-4528-84ba-498965f3883a-env-overrides\") pod \"ovnkube-node-65ngd\" (UID: \"6d025429-c91d-4528-84ba-498965f3883a\") " pod="openshift-ovn-kubernetes/ovnkube-node-65ngd" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.863584 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/6d025429-c91d-4528-84ba-498965f3883a-log-socket\") pod \"ovnkube-node-65ngd\" (UID: \"6d025429-c91d-4528-84ba-498965f3883a\") " pod="openshift-ovn-kubernetes/ovnkube-node-65ngd" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.863610 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6d025429-c91d-4528-84ba-498965f3883a-ovn-node-metrics-cert\") pod \"ovnkube-node-65ngd\" (UID: \"6d025429-c91d-4528-84ba-498965f3883a\") " pod="openshift-ovn-kubernetes/ovnkube-node-65ngd" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.863634 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6d025429-c91d-4528-84ba-498965f3883a-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-65ngd\" (UID: \"6d025429-c91d-4528-84ba-498965f3883a\") " pod="openshift-ovn-kubernetes/ovnkube-node-65ngd" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.863645 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6d025429-c91d-4528-84ba-498965f3883a-var-lib-openvswitch\") pod \"ovnkube-node-65ngd\" (UID: \"6d025429-c91d-4528-84ba-498965f3883a\") " pod="openshift-ovn-kubernetes/ovnkube-node-65ngd" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.863708 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/6d025429-c91d-4528-84ba-498965f3883a-systemd-units\") pod \"ovnkube-node-65ngd\" (UID: \"6d025429-c91d-4528-84ba-498965f3883a\") " pod="openshift-ovn-kubernetes/ovnkube-node-65ngd" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.863666 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/6d025429-c91d-4528-84ba-498965f3883a-systemd-units\") pod \"ovnkube-node-65ngd\" (UID: \"6d025429-c91d-4528-84ba-498965f3883a\") " pod="openshift-ovn-kubernetes/ovnkube-node-65ngd" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.863724 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/6d025429-c91d-4528-84ba-498965f3883a-log-socket\") pod \"ovnkube-node-65ngd\" (UID: \"6d025429-c91d-4528-84ba-498965f3883a\") " pod="openshift-ovn-kubernetes/ovnkube-node-65ngd" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.863783 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/6d025429-c91d-4528-84ba-498965f3883a-host-kubelet\") pod \"ovnkube-node-65ngd\" (UID: \"6d025429-c91d-4528-84ba-498965f3883a\") " pod="openshift-ovn-kubernetes/ovnkube-node-65ngd" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.863818 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/6d025429-c91d-4528-84ba-498965f3883a-run-ovn\") pod \"ovnkube-node-65ngd\" (UID: \"6d025429-c91d-4528-84ba-498965f3883a\") " pod="openshift-ovn-kubernetes/ovnkube-node-65ngd" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.863847 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6d025429-c91d-4528-84ba-498965f3883a-etc-openvswitch\") pod \"ovnkube-node-65ngd\" (UID: \"6d025429-c91d-4528-84ba-498965f3883a\") " pod="openshift-ovn-kubernetes/ovnkube-node-65ngd" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.863863 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zwrnk\" (UniqueName: \"kubernetes.io/projected/6d025429-c91d-4528-84ba-498965f3883a-kube-api-access-zwrnk\") pod \"ovnkube-node-65ngd\" (UID: \"6d025429-c91d-4528-84ba-498965f3883a\") " pod="openshift-ovn-kubernetes/ovnkube-node-65ngd" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.863854 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6d025429-c91d-4528-84ba-498965f3883a-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-65ngd\" (UID: \"6d025429-c91d-4528-84ba-498965f3883a\") " pod="openshift-ovn-kubernetes/ovnkube-node-65ngd" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.863927 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6d025429-c91d-4528-84ba-498965f3883a-etc-openvswitch\") pod \"ovnkube-node-65ngd\" (UID: \"6d025429-c91d-4528-84ba-498965f3883a\") " pod="openshift-ovn-kubernetes/ovnkube-node-65ngd" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.863913 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/6d025429-c91d-4528-84ba-498965f3883a-host-cni-bin\") pod \"ovnkube-node-65ngd\" (UID: \"6d025429-c91d-4528-84ba-498965f3883a\") " pod="openshift-ovn-kubernetes/ovnkube-node-65ngd" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.863896 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/6d025429-c91d-4528-84ba-498965f3883a-host-cni-bin\") pod \"ovnkube-node-65ngd\" (UID: \"6d025429-c91d-4528-84ba-498965f3883a\") " pod="openshift-ovn-kubernetes/ovnkube-node-65ngd" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.863972 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/6d025429-c91d-4528-84ba-498965f3883a-host-kubelet\") pod \"ovnkube-node-65ngd\" (UID: \"6d025429-c91d-4528-84ba-498965f3883a\") " pod="openshift-ovn-kubernetes/ovnkube-node-65ngd" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.863977 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/6d025429-c91d-4528-84ba-498965f3883a-run-ovn\") pod \"ovnkube-node-65ngd\" (UID: \"6d025429-c91d-4528-84ba-498965f3883a\") " pod="openshift-ovn-kubernetes/ovnkube-node-65ngd" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.864068 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/6d025429-c91d-4528-84ba-498965f3883a-node-log\") pod \"ovnkube-node-65ngd\" (UID: \"6d025429-c91d-4528-84ba-498965f3883a\") " pod="openshift-ovn-kubernetes/ovnkube-node-65ngd" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.864140 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6d025429-c91d-4528-84ba-498965f3883a-host-run-ovn-kubernetes\") pod \"ovnkube-node-65ngd\" (UID: \"6d025429-c91d-4528-84ba-498965f3883a\") " pod="openshift-ovn-kubernetes/ovnkube-node-65ngd" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.864184 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6d025429-c91d-4528-84ba-498965f3883a-run-openvswitch\") pod \"ovnkube-node-65ngd\" (UID: \"6d025429-c91d-4528-84ba-498965f3883a\") " pod="openshift-ovn-kubernetes/ovnkube-node-65ngd" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.864222 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/6d025429-c91d-4528-84ba-498965f3883a-node-log\") pod \"ovnkube-node-65ngd\" (UID: \"6d025429-c91d-4528-84ba-498965f3883a\") " pod="openshift-ovn-kubernetes/ovnkube-node-65ngd" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.864226 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6d025429-c91d-4528-84ba-498965f3883a-host-run-ovn-kubernetes\") pod \"ovnkube-node-65ngd\" (UID: \"6d025429-c91d-4528-84ba-498965f3883a\") " pod="openshift-ovn-kubernetes/ovnkube-node-65ngd" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.864258 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/6d025429-c91d-4528-84ba-498965f3883a-host-slash\") pod \"ovnkube-node-65ngd\" (UID: \"6d025429-c91d-4528-84ba-498965f3883a\") " pod="openshift-ovn-kubernetes/ovnkube-node-65ngd" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.864318 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/6d025429-c91d-4528-84ba-498965f3883a-host-run-netns\") pod \"ovnkube-node-65ngd\" (UID: \"6d025429-c91d-4528-84ba-498965f3883a\") " pod="openshift-ovn-kubernetes/ovnkube-node-65ngd" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.864354 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6d025429-c91d-4528-84ba-498965f3883a-ovnkube-config\") pod \"ovnkube-node-65ngd\" (UID: \"6d025429-c91d-4528-84ba-498965f3883a\") " pod="openshift-ovn-kubernetes/ovnkube-node-65ngd" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.864338 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6d025429-c91d-4528-84ba-498965f3883a-run-openvswitch\") pod \"ovnkube-node-65ngd\" (UID: \"6d025429-c91d-4528-84ba-498965f3883a\") " pod="openshift-ovn-kubernetes/ovnkube-node-65ngd" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.864395 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/6d025429-c91d-4528-84ba-498965f3883a-host-cni-netd\") pod \"ovnkube-node-65ngd\" (UID: \"6d025429-c91d-4528-84ba-498965f3883a\") " pod="openshift-ovn-kubernetes/ovnkube-node-65ngd" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.864399 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/6d025429-c91d-4528-84ba-498965f3883a-host-slash\") pod \"ovnkube-node-65ngd\" (UID: \"6d025429-c91d-4528-84ba-498965f3883a\") " pod="openshift-ovn-kubernetes/ovnkube-node-65ngd" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.864429 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/6d025429-c91d-4528-84ba-498965f3883a-run-systemd\") pod \"ovnkube-node-65ngd\" (UID: \"6d025429-c91d-4528-84ba-498965f3883a\") " pod="openshift-ovn-kubernetes/ovnkube-node-65ngd" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.864440 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/6d025429-c91d-4528-84ba-498965f3883a-host-cni-netd\") pod \"ovnkube-node-65ngd\" (UID: \"6d025429-c91d-4528-84ba-498965f3883a\") " pod="openshift-ovn-kubernetes/ovnkube-node-65ngd" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.864457 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6d025429-c91d-4528-84ba-498965f3883a-env-overrides\") pod \"ovnkube-node-65ngd\" (UID: \"6d025429-c91d-4528-84ba-498965f3883a\") " pod="openshift-ovn-kubernetes/ovnkube-node-65ngd" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.864467 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6d025429-c91d-4528-84ba-498965f3883a-ovnkube-script-lib\") pod \"ovnkube-node-65ngd\" (UID: \"6d025429-c91d-4528-84ba-498965f3883a\") " pod="openshift-ovn-kubernetes/ovnkube-node-65ngd" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.864399 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/6d025429-c91d-4528-84ba-498965f3883a-host-run-netns\") pod \"ovnkube-node-65ngd\" (UID: \"6d025429-c91d-4528-84ba-498965f3883a\") " pod="openshift-ovn-kubernetes/ovnkube-node-65ngd" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.864497 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/6d025429-c91d-4528-84ba-498965f3883a-run-systemd\") pod \"ovnkube-node-65ngd\" (UID: \"6d025429-c91d-4528-84ba-498965f3883a\") " pod="openshift-ovn-kubernetes/ovnkube-node-65ngd" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.864923 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6d025429-c91d-4528-84ba-498965f3883a-ovnkube-config\") pod \"ovnkube-node-65ngd\" (UID: \"6d025429-c91d-4528-84ba-498965f3883a\") " pod="openshift-ovn-kubernetes/ovnkube-node-65ngd" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.865700 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6d025429-c91d-4528-84ba-498965f3883a-ovnkube-script-lib\") pod \"ovnkube-node-65ngd\" (UID: \"6d025429-c91d-4528-84ba-498965f3883a\") " pod="openshift-ovn-kubernetes/ovnkube-node-65ngd" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.868786 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6d025429-c91d-4528-84ba-498965f3883a-ovn-node-metrics-cert\") pod \"ovnkube-node-65ngd\" (UID: \"6d025429-c91d-4528-84ba-498965f3883a\") " pod="openshift-ovn-kubernetes/ovnkube-node-65ngd" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.888602 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zwrnk\" (UniqueName: \"kubernetes.io/projected/6d025429-c91d-4528-84ba-498965f3883a-kube-api-access-zwrnk\") pod \"ovnkube-node-65ngd\" (UID: \"6d025429-c91d-4528-84ba-498965f3883a\") " pod="openshift-ovn-kubernetes/ovnkube-node-65ngd" Dec 11 02:14:07 crc kubenswrapper[4824]: I1211 02:14:07.959487 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-65ngd" Dec 11 02:14:07 crc kubenswrapper[4824]: W1211 02:14:07.987210 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6d025429_c91d_4528_84ba_498965f3883a.slice/crio-9ac1f3d660d78f5954d3197ce7a44346ce515ac9cee820eb3c5e4fa736d12bc5 WatchSource:0}: Error finding container 9ac1f3d660d78f5954d3197ce7a44346ce515ac9cee820eb3c5e4fa736d12bc5: Status 404 returned error can't find the container with id 9ac1f3d660d78f5954d3197ce7a44346ce515ac9cee820eb3c5e4fa736d12bc5 Dec 11 02:14:08 crc kubenswrapper[4824]: I1211 02:14:08.503261 4824 generic.go:334] "Generic (PLEG): container finished" podID="6d025429-c91d-4528-84ba-498965f3883a" containerID="bc15a86ad88003f44a2f6b01b46a3afa665eeca1365fbd0f8e9549edddc6ee8f" exitCode=0 Dec 11 02:14:08 crc kubenswrapper[4824]: I1211 02:14:08.503358 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-65ngd" event={"ID":"6d025429-c91d-4528-84ba-498965f3883a","Type":"ContainerDied","Data":"bc15a86ad88003f44a2f6b01b46a3afa665eeca1365fbd0f8e9549edddc6ee8f"} Dec 11 02:14:08 crc kubenswrapper[4824]: I1211 02:14:08.503791 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-65ngd" event={"ID":"6d025429-c91d-4528-84ba-498965f3883a","Type":"ContainerStarted","Data":"9ac1f3d660d78f5954d3197ce7a44346ce515ac9cee820eb3c5e4fa736d12bc5"} Dec 11 02:14:08 crc kubenswrapper[4824]: I1211 02:14:08.505743 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-f6pgc" Dec 11 02:14:08 crc kubenswrapper[4824]: I1211 02:14:08.509490 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6bkc2_8022f696-d0e7-437b-848a-3cd25bd1f364/kube-multus/2.log" Dec 11 02:14:08 crc kubenswrapper[4824]: I1211 02:14:08.510347 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6bkc2_8022f696-d0e7-437b-848a-3cd25bd1f364/kube-multus/1.log" Dec 11 02:14:08 crc kubenswrapper[4824]: I1211 02:14:08.510478 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-6bkc2" event={"ID":"8022f696-d0e7-437b-848a-3cd25bd1f364","Type":"ContainerStarted","Data":"4920c7463098ce8e5ba9a84330cf5227210370fe67c507b853842f45c39c0797"} Dec 11 02:14:08 crc kubenswrapper[4824]: I1211 02:14:08.515626 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" event={"ID":"44b156e4-64a4-4d45-aa5e-9b10a862faed","Type":"ContainerStarted","Data":"65dbaed5184dd7e582c2d25d2c2754654e16b28021f7e2281401ebdcef00a13e"} Dec 11 02:14:08 crc kubenswrapper[4824]: I1211 02:14:08.583473 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-f6pgc"] Dec 11 02:14:08 crc kubenswrapper[4824]: I1211 02:14:08.595582 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-f6pgc"] Dec 11 02:14:08 crc kubenswrapper[4824]: I1211 02:14:08.644235 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b193bd2-b536-4056-92f1-94c9836ab2eb" path="/var/lib/kubelet/pods/5b193bd2-b536-4056-92f1-94c9836ab2eb/volumes" Dec 11 02:14:09 crc kubenswrapper[4824]: I1211 02:14:09.529014 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-65ngd" event={"ID":"6d025429-c91d-4528-84ba-498965f3883a","Type":"ContainerStarted","Data":"c0d3d15ea883de421605be89555bf94f7641e20f10d1193bf2d6a43a32d99bb1"} Dec 11 02:14:09 crc kubenswrapper[4824]: I1211 02:14:09.530046 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-65ngd" event={"ID":"6d025429-c91d-4528-84ba-498965f3883a","Type":"ContainerStarted","Data":"9efc040ebdce85b42049fac2474690d6280dab6257b0b76d1dccfec2512bcab2"} Dec 11 02:14:09 crc kubenswrapper[4824]: I1211 02:14:09.530076 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-65ngd" event={"ID":"6d025429-c91d-4528-84ba-498965f3883a","Type":"ContainerStarted","Data":"425c0f64638e2122d626b4b62152f1f9e92ccc8222e8618bb23d7fb52abd5377"} Dec 11 02:14:09 crc kubenswrapper[4824]: I1211 02:14:09.530095 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-65ngd" event={"ID":"6d025429-c91d-4528-84ba-498965f3883a","Type":"ContainerStarted","Data":"1c6a3ec30b04dba8cc116566625d7c44d981b9d2167349b051756c1770f9d149"} Dec 11 02:14:09 crc kubenswrapper[4824]: I1211 02:14:09.530137 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-65ngd" event={"ID":"6d025429-c91d-4528-84ba-498965f3883a","Type":"ContainerStarted","Data":"6031bb22310d604187f5671f316d06b9b4eb90669b07c2470fdf7e717339a623"} Dec 11 02:14:09 crc kubenswrapper[4824]: I1211 02:14:09.530155 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-65ngd" event={"ID":"6d025429-c91d-4528-84ba-498965f3883a","Type":"ContainerStarted","Data":"b35c6e7788e365f732b3031dbb964bb836e00f80f96c357227084f221218ad8e"} Dec 11 02:14:09 crc kubenswrapper[4824]: I1211 02:14:09.530359 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-4mrwp" podUID="11965819-0e2c-4107-a14c-2c7c2653701e" containerName="registry-server" containerID="cri-o://43d294562be066e780dc4992ad6c2822a7f104ac79ca294155feabd9d73fec4f" gracePeriod=2 Dec 11 02:14:09 crc kubenswrapper[4824]: I1211 02:14:09.731442 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4mrwp" Dec 11 02:14:09 crc kubenswrapper[4824]: I1211 02:14:09.896974 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jqssf\" (UniqueName: \"kubernetes.io/projected/11965819-0e2c-4107-a14c-2c7c2653701e-kube-api-access-jqssf\") pod \"11965819-0e2c-4107-a14c-2c7c2653701e\" (UID: \"11965819-0e2c-4107-a14c-2c7c2653701e\") " Dec 11 02:14:09 crc kubenswrapper[4824]: I1211 02:14:09.897081 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/11965819-0e2c-4107-a14c-2c7c2653701e-catalog-content\") pod \"11965819-0e2c-4107-a14c-2c7c2653701e\" (UID: \"11965819-0e2c-4107-a14c-2c7c2653701e\") " Dec 11 02:14:09 crc kubenswrapper[4824]: I1211 02:14:09.897175 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/11965819-0e2c-4107-a14c-2c7c2653701e-utilities\") pod \"11965819-0e2c-4107-a14c-2c7c2653701e\" (UID: \"11965819-0e2c-4107-a14c-2c7c2653701e\") " Dec 11 02:14:09 crc kubenswrapper[4824]: I1211 02:14:09.898159 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/11965819-0e2c-4107-a14c-2c7c2653701e-utilities" (OuterVolumeSpecName: "utilities") pod "11965819-0e2c-4107-a14c-2c7c2653701e" (UID: "11965819-0e2c-4107-a14c-2c7c2653701e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:14:09 crc kubenswrapper[4824]: I1211 02:14:09.906941 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11965819-0e2c-4107-a14c-2c7c2653701e-kube-api-access-jqssf" (OuterVolumeSpecName: "kube-api-access-jqssf") pod "11965819-0e2c-4107-a14c-2c7c2653701e" (UID: "11965819-0e2c-4107-a14c-2c7c2653701e"). InnerVolumeSpecName "kube-api-access-jqssf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:14:09 crc kubenswrapper[4824]: I1211 02:14:09.917857 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/11965819-0e2c-4107-a14c-2c7c2653701e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "11965819-0e2c-4107-a14c-2c7c2653701e" (UID: "11965819-0e2c-4107-a14c-2c7c2653701e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:14:09 crc kubenswrapper[4824]: I1211 02:14:09.998869 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/11965819-0e2c-4107-a14c-2c7c2653701e-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 02:14:09 crc kubenswrapper[4824]: I1211 02:14:09.998904 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jqssf\" (UniqueName: \"kubernetes.io/projected/11965819-0e2c-4107-a14c-2c7c2653701e-kube-api-access-jqssf\") on node \"crc\" DevicePath \"\"" Dec 11 02:14:09 crc kubenswrapper[4824]: I1211 02:14:09.998916 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/11965819-0e2c-4107-a14c-2c7c2653701e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 02:14:10 crc kubenswrapper[4824]: I1211 02:14:10.537724 4824 generic.go:334] "Generic (PLEG): container finished" podID="11965819-0e2c-4107-a14c-2c7c2653701e" containerID="43d294562be066e780dc4992ad6c2822a7f104ac79ca294155feabd9d73fec4f" exitCode=0 Dec 11 02:14:10 crc kubenswrapper[4824]: I1211 02:14:10.537785 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4mrwp" Dec 11 02:14:10 crc kubenswrapper[4824]: I1211 02:14:10.537839 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4mrwp" event={"ID":"11965819-0e2c-4107-a14c-2c7c2653701e","Type":"ContainerDied","Data":"43d294562be066e780dc4992ad6c2822a7f104ac79ca294155feabd9d73fec4f"} Dec 11 02:14:10 crc kubenswrapper[4824]: I1211 02:14:10.538236 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4mrwp" event={"ID":"11965819-0e2c-4107-a14c-2c7c2653701e","Type":"ContainerDied","Data":"2c6371b30959fdcb41e9ff7a99ef960977653369f596d2087a68fdbc9064ee9f"} Dec 11 02:14:10 crc kubenswrapper[4824]: I1211 02:14:10.538272 4824 scope.go:117] "RemoveContainer" containerID="43d294562be066e780dc4992ad6c2822a7f104ac79ca294155feabd9d73fec4f" Dec 11 02:14:10 crc kubenswrapper[4824]: I1211 02:14:10.561358 4824 scope.go:117] "RemoveContainer" containerID="c26153691c06d47721e32a065fe6744a4638a9a48a77897c876e81428f70b088" Dec 11 02:14:10 crc kubenswrapper[4824]: I1211 02:14:10.574628 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4mrwp"] Dec 11 02:14:10 crc kubenswrapper[4824]: I1211 02:14:10.587980 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-4mrwp"] Dec 11 02:14:10 crc kubenswrapper[4824]: I1211 02:14:10.596337 4824 scope.go:117] "RemoveContainer" containerID="75fec8e1f16bd2bed3219ed3ec1e9d024bdd535d556fd0f186ab3ef4e3fe1a25" Dec 11 02:14:10 crc kubenswrapper[4824]: I1211 02:14:10.619726 4824 scope.go:117] "RemoveContainer" containerID="43d294562be066e780dc4992ad6c2822a7f104ac79ca294155feabd9d73fec4f" Dec 11 02:14:10 crc kubenswrapper[4824]: E1211 02:14:10.620357 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"43d294562be066e780dc4992ad6c2822a7f104ac79ca294155feabd9d73fec4f\": container with ID starting with 43d294562be066e780dc4992ad6c2822a7f104ac79ca294155feabd9d73fec4f not found: ID does not exist" containerID="43d294562be066e780dc4992ad6c2822a7f104ac79ca294155feabd9d73fec4f" Dec 11 02:14:10 crc kubenswrapper[4824]: I1211 02:14:10.620414 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"43d294562be066e780dc4992ad6c2822a7f104ac79ca294155feabd9d73fec4f"} err="failed to get container status \"43d294562be066e780dc4992ad6c2822a7f104ac79ca294155feabd9d73fec4f\": rpc error: code = NotFound desc = could not find container \"43d294562be066e780dc4992ad6c2822a7f104ac79ca294155feabd9d73fec4f\": container with ID starting with 43d294562be066e780dc4992ad6c2822a7f104ac79ca294155feabd9d73fec4f not found: ID does not exist" Dec 11 02:14:10 crc kubenswrapper[4824]: I1211 02:14:10.620452 4824 scope.go:117] "RemoveContainer" containerID="c26153691c06d47721e32a065fe6744a4638a9a48a77897c876e81428f70b088" Dec 11 02:14:10 crc kubenswrapper[4824]: E1211 02:14:10.621020 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c26153691c06d47721e32a065fe6744a4638a9a48a77897c876e81428f70b088\": container with ID starting with c26153691c06d47721e32a065fe6744a4638a9a48a77897c876e81428f70b088 not found: ID does not exist" containerID="c26153691c06d47721e32a065fe6744a4638a9a48a77897c876e81428f70b088" Dec 11 02:14:10 crc kubenswrapper[4824]: I1211 02:14:10.621082 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c26153691c06d47721e32a065fe6744a4638a9a48a77897c876e81428f70b088"} err="failed to get container status \"c26153691c06d47721e32a065fe6744a4638a9a48a77897c876e81428f70b088\": rpc error: code = NotFound desc = could not find container \"c26153691c06d47721e32a065fe6744a4638a9a48a77897c876e81428f70b088\": container with ID starting with c26153691c06d47721e32a065fe6744a4638a9a48a77897c876e81428f70b088 not found: ID does not exist" Dec 11 02:14:10 crc kubenswrapper[4824]: I1211 02:14:10.621147 4824 scope.go:117] "RemoveContainer" containerID="75fec8e1f16bd2bed3219ed3ec1e9d024bdd535d556fd0f186ab3ef4e3fe1a25" Dec 11 02:14:10 crc kubenswrapper[4824]: E1211 02:14:10.621676 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"75fec8e1f16bd2bed3219ed3ec1e9d024bdd535d556fd0f186ab3ef4e3fe1a25\": container with ID starting with 75fec8e1f16bd2bed3219ed3ec1e9d024bdd535d556fd0f186ab3ef4e3fe1a25 not found: ID does not exist" containerID="75fec8e1f16bd2bed3219ed3ec1e9d024bdd535d556fd0f186ab3ef4e3fe1a25" Dec 11 02:14:10 crc kubenswrapper[4824]: I1211 02:14:10.621727 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"75fec8e1f16bd2bed3219ed3ec1e9d024bdd535d556fd0f186ab3ef4e3fe1a25"} err="failed to get container status \"75fec8e1f16bd2bed3219ed3ec1e9d024bdd535d556fd0f186ab3ef4e3fe1a25\": rpc error: code = NotFound desc = could not find container \"75fec8e1f16bd2bed3219ed3ec1e9d024bdd535d556fd0f186ab3ef4e3fe1a25\": container with ID starting with 75fec8e1f16bd2bed3219ed3ec1e9d024bdd535d556fd0f186ab3ef4e3fe1a25 not found: ID does not exist" Dec 11 02:14:10 crc kubenswrapper[4824]: I1211 02:14:10.644039 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="11965819-0e2c-4107-a14c-2c7c2653701e" path="/var/lib/kubelet/pods/11965819-0e2c-4107-a14c-2c7c2653701e/volumes" Dec 11 02:14:13 crc kubenswrapper[4824]: I1211 02:14:13.565172 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-65ngd" event={"ID":"6d025429-c91d-4528-84ba-498965f3883a","Type":"ContainerStarted","Data":"66d0ad48d250f72da1552335ccc5c6cd6852dfa37487077179dc0cb9a93c5059"} Dec 11 02:14:16 crc kubenswrapper[4824]: I1211 02:14:16.590229 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-65ngd" event={"ID":"6d025429-c91d-4528-84ba-498965f3883a","Type":"ContainerStarted","Data":"cb0ba3663f2a4f6bcdc0e18afde041be3c91d811697b0d06d3cd4984aa437d5c"} Dec 11 02:14:16 crc kubenswrapper[4824]: I1211 02:14:16.590610 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-65ngd" Dec 11 02:14:16 crc kubenswrapper[4824]: I1211 02:14:16.590634 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-65ngd" Dec 11 02:14:16 crc kubenswrapper[4824]: I1211 02:14:16.590652 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-65ngd" Dec 11 02:14:16 crc kubenswrapper[4824]: I1211 02:14:16.616916 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-65ngd" Dec 11 02:14:16 crc kubenswrapper[4824]: I1211 02:14:16.627467 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-65ngd" podStartSLOduration=9.627448514 podStartE2EDuration="9.627448514s" podCreationTimestamp="2025-12-11 02:14:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:14:16.624993656 +0000 UTC m=+798.314031035" watchObservedRunningTime="2025-12-11 02:14:16.627448514 +0000 UTC m=+798.316485903" Dec 11 02:14:16 crc kubenswrapper[4824]: I1211 02:14:16.630764 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-65ngd" Dec 11 02:14:37 crc kubenswrapper[4824]: I1211 02:14:37.995607 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-65ngd" Dec 11 02:14:47 crc kubenswrapper[4824]: I1211 02:14:47.359466 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f97jkp"] Dec 11 02:14:47 crc kubenswrapper[4824]: E1211 02:14:47.360091 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11965819-0e2c-4107-a14c-2c7c2653701e" containerName="extract-content" Dec 11 02:14:47 crc kubenswrapper[4824]: I1211 02:14:47.360102 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="11965819-0e2c-4107-a14c-2c7c2653701e" containerName="extract-content" Dec 11 02:14:47 crc kubenswrapper[4824]: E1211 02:14:47.360135 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11965819-0e2c-4107-a14c-2c7c2653701e" containerName="extract-utilities" Dec 11 02:14:47 crc kubenswrapper[4824]: I1211 02:14:47.360142 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="11965819-0e2c-4107-a14c-2c7c2653701e" containerName="extract-utilities" Dec 11 02:14:47 crc kubenswrapper[4824]: E1211 02:14:47.360157 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11965819-0e2c-4107-a14c-2c7c2653701e" containerName="registry-server" Dec 11 02:14:47 crc kubenswrapper[4824]: I1211 02:14:47.360163 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="11965819-0e2c-4107-a14c-2c7c2653701e" containerName="registry-server" Dec 11 02:14:47 crc kubenswrapper[4824]: I1211 02:14:47.360253 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="11965819-0e2c-4107-a14c-2c7c2653701e" containerName="registry-server" Dec 11 02:14:47 crc kubenswrapper[4824]: I1211 02:14:47.360885 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f97jkp" Dec 11 02:14:47 crc kubenswrapper[4824]: I1211 02:14:47.363186 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 11 02:14:47 crc kubenswrapper[4824]: I1211 02:14:47.372454 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f97jkp"] Dec 11 02:14:47 crc kubenswrapper[4824]: I1211 02:14:47.463143 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bb38965f-b3bb-4d2e-8430-8c29298730e4-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f97jkp\" (UID: \"bb38965f-b3bb-4d2e-8430-8c29298730e4\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f97jkp" Dec 11 02:14:47 crc kubenswrapper[4824]: I1211 02:14:47.463210 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-krrwj\" (UniqueName: \"kubernetes.io/projected/bb38965f-b3bb-4d2e-8430-8c29298730e4-kube-api-access-krrwj\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f97jkp\" (UID: \"bb38965f-b3bb-4d2e-8430-8c29298730e4\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f97jkp" Dec 11 02:14:47 crc kubenswrapper[4824]: I1211 02:14:47.463252 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bb38965f-b3bb-4d2e-8430-8c29298730e4-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f97jkp\" (UID: \"bb38965f-b3bb-4d2e-8430-8c29298730e4\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f97jkp" Dec 11 02:14:47 crc kubenswrapper[4824]: I1211 02:14:47.564761 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-krrwj\" (UniqueName: \"kubernetes.io/projected/bb38965f-b3bb-4d2e-8430-8c29298730e4-kube-api-access-krrwj\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f97jkp\" (UID: \"bb38965f-b3bb-4d2e-8430-8c29298730e4\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f97jkp" Dec 11 02:14:47 crc kubenswrapper[4824]: I1211 02:14:47.564867 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bb38965f-b3bb-4d2e-8430-8c29298730e4-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f97jkp\" (UID: \"bb38965f-b3bb-4d2e-8430-8c29298730e4\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f97jkp" Dec 11 02:14:47 crc kubenswrapper[4824]: I1211 02:14:47.564974 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bb38965f-b3bb-4d2e-8430-8c29298730e4-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f97jkp\" (UID: \"bb38965f-b3bb-4d2e-8430-8c29298730e4\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f97jkp" Dec 11 02:14:47 crc kubenswrapper[4824]: I1211 02:14:47.565826 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bb38965f-b3bb-4d2e-8430-8c29298730e4-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f97jkp\" (UID: \"bb38965f-b3bb-4d2e-8430-8c29298730e4\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f97jkp" Dec 11 02:14:47 crc kubenswrapper[4824]: I1211 02:14:47.565920 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bb38965f-b3bb-4d2e-8430-8c29298730e4-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f97jkp\" (UID: \"bb38965f-b3bb-4d2e-8430-8c29298730e4\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f97jkp" Dec 11 02:14:47 crc kubenswrapper[4824]: I1211 02:14:47.601863 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-krrwj\" (UniqueName: \"kubernetes.io/projected/bb38965f-b3bb-4d2e-8430-8c29298730e4-kube-api-access-krrwj\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f97jkp\" (UID: \"bb38965f-b3bb-4d2e-8430-8c29298730e4\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f97jkp" Dec 11 02:14:47 crc kubenswrapper[4824]: I1211 02:14:47.682505 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f97jkp" Dec 11 02:14:48 crc kubenswrapper[4824]: I1211 02:14:48.027487 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f97jkp"] Dec 11 02:14:48 crc kubenswrapper[4824]: W1211 02:14:48.040078 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbb38965f_b3bb_4d2e_8430_8c29298730e4.slice/crio-a257963bb90690d20a5d3d2d27a8a2f8f6eef7617d0c6e22ef02a6ed1598117f WatchSource:0}: Error finding container a257963bb90690d20a5d3d2d27a8a2f8f6eef7617d0c6e22ef02a6ed1598117f: Status 404 returned error can't find the container with id a257963bb90690d20a5d3d2d27a8a2f8f6eef7617d0c6e22ef02a6ed1598117f Dec 11 02:14:48 crc kubenswrapper[4824]: I1211 02:14:48.791048 4824 generic.go:334] "Generic (PLEG): container finished" podID="bb38965f-b3bb-4d2e-8430-8c29298730e4" containerID="185bcf229c0725db95f0c5e449220a2b3e6a99e39a96e98201d9182c41ffce1e" exitCode=0 Dec 11 02:14:48 crc kubenswrapper[4824]: I1211 02:14:48.791148 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f97jkp" event={"ID":"bb38965f-b3bb-4d2e-8430-8c29298730e4","Type":"ContainerDied","Data":"185bcf229c0725db95f0c5e449220a2b3e6a99e39a96e98201d9182c41ffce1e"} Dec 11 02:14:48 crc kubenswrapper[4824]: I1211 02:14:48.791591 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f97jkp" event={"ID":"bb38965f-b3bb-4d2e-8430-8c29298730e4","Type":"ContainerStarted","Data":"a257963bb90690d20a5d3d2d27a8a2f8f6eef7617d0c6e22ef02a6ed1598117f"} Dec 11 02:14:49 crc kubenswrapper[4824]: I1211 02:14:49.548422 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-md4vc"] Dec 11 02:14:49 crc kubenswrapper[4824]: I1211 02:14:49.549540 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-md4vc" Dec 11 02:14:49 crc kubenswrapper[4824]: I1211 02:14:49.565265 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-md4vc"] Dec 11 02:14:49 crc kubenswrapper[4824]: I1211 02:14:49.593828 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5fhzw\" (UniqueName: \"kubernetes.io/projected/31523d77-e1de-4227-a1f3-be75363b3321-kube-api-access-5fhzw\") pod \"redhat-operators-md4vc\" (UID: \"31523d77-e1de-4227-a1f3-be75363b3321\") " pod="openshift-marketplace/redhat-operators-md4vc" Dec 11 02:14:49 crc kubenswrapper[4824]: I1211 02:14:49.593879 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/31523d77-e1de-4227-a1f3-be75363b3321-catalog-content\") pod \"redhat-operators-md4vc\" (UID: \"31523d77-e1de-4227-a1f3-be75363b3321\") " pod="openshift-marketplace/redhat-operators-md4vc" Dec 11 02:14:49 crc kubenswrapper[4824]: I1211 02:14:49.593963 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/31523d77-e1de-4227-a1f3-be75363b3321-utilities\") pod \"redhat-operators-md4vc\" (UID: \"31523d77-e1de-4227-a1f3-be75363b3321\") " pod="openshift-marketplace/redhat-operators-md4vc" Dec 11 02:14:49 crc kubenswrapper[4824]: I1211 02:14:49.695362 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/31523d77-e1de-4227-a1f3-be75363b3321-utilities\") pod \"redhat-operators-md4vc\" (UID: \"31523d77-e1de-4227-a1f3-be75363b3321\") " pod="openshift-marketplace/redhat-operators-md4vc" Dec 11 02:14:49 crc kubenswrapper[4824]: I1211 02:14:49.695789 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5fhzw\" (UniqueName: \"kubernetes.io/projected/31523d77-e1de-4227-a1f3-be75363b3321-kube-api-access-5fhzw\") pod \"redhat-operators-md4vc\" (UID: \"31523d77-e1de-4227-a1f3-be75363b3321\") " pod="openshift-marketplace/redhat-operators-md4vc" Dec 11 02:14:49 crc kubenswrapper[4824]: I1211 02:14:49.695981 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/31523d77-e1de-4227-a1f3-be75363b3321-catalog-content\") pod \"redhat-operators-md4vc\" (UID: \"31523d77-e1de-4227-a1f3-be75363b3321\") " pod="openshift-marketplace/redhat-operators-md4vc" Dec 11 02:14:49 crc kubenswrapper[4824]: I1211 02:14:49.695880 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/31523d77-e1de-4227-a1f3-be75363b3321-utilities\") pod \"redhat-operators-md4vc\" (UID: \"31523d77-e1de-4227-a1f3-be75363b3321\") " pod="openshift-marketplace/redhat-operators-md4vc" Dec 11 02:14:49 crc kubenswrapper[4824]: I1211 02:14:49.696321 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/31523d77-e1de-4227-a1f3-be75363b3321-catalog-content\") pod \"redhat-operators-md4vc\" (UID: \"31523d77-e1de-4227-a1f3-be75363b3321\") " pod="openshift-marketplace/redhat-operators-md4vc" Dec 11 02:14:49 crc kubenswrapper[4824]: I1211 02:14:49.724523 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5fhzw\" (UniqueName: \"kubernetes.io/projected/31523d77-e1de-4227-a1f3-be75363b3321-kube-api-access-5fhzw\") pod \"redhat-operators-md4vc\" (UID: \"31523d77-e1de-4227-a1f3-be75363b3321\") " pod="openshift-marketplace/redhat-operators-md4vc" Dec 11 02:14:49 crc kubenswrapper[4824]: I1211 02:14:49.877813 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-md4vc" Dec 11 02:14:50 crc kubenswrapper[4824]: I1211 02:14:50.313712 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-md4vc"] Dec 11 02:14:50 crc kubenswrapper[4824]: W1211 02:14:50.316282 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod31523d77_e1de_4227_a1f3_be75363b3321.slice/crio-fddaef2efa18822eabe257fe340a93757cfd04a49d3589d5a0fc79b319029fca WatchSource:0}: Error finding container fddaef2efa18822eabe257fe340a93757cfd04a49d3589d5a0fc79b319029fca: Status 404 returned error can't find the container with id fddaef2efa18822eabe257fe340a93757cfd04a49d3589d5a0fc79b319029fca Dec 11 02:14:50 crc kubenswrapper[4824]: I1211 02:14:50.805630 4824 generic.go:334] "Generic (PLEG): container finished" podID="31523d77-e1de-4227-a1f3-be75363b3321" containerID="ba63082c996f39244eb9d5eb4fea9097d12f6218c34bca7a452de9a2aba18b9e" exitCode=0 Dec 11 02:14:50 crc kubenswrapper[4824]: I1211 02:14:50.805689 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-md4vc" event={"ID":"31523d77-e1de-4227-a1f3-be75363b3321","Type":"ContainerDied","Data":"ba63082c996f39244eb9d5eb4fea9097d12f6218c34bca7a452de9a2aba18b9e"} Dec 11 02:14:50 crc kubenswrapper[4824]: I1211 02:14:50.806062 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-md4vc" event={"ID":"31523d77-e1de-4227-a1f3-be75363b3321","Type":"ContainerStarted","Data":"fddaef2efa18822eabe257fe340a93757cfd04a49d3589d5a0fc79b319029fca"} Dec 11 02:14:50 crc kubenswrapper[4824]: I1211 02:14:50.808084 4824 generic.go:334] "Generic (PLEG): container finished" podID="bb38965f-b3bb-4d2e-8430-8c29298730e4" containerID="7e68d1b52f2336bb4c3d2806d6f171b4d3d397e67653d7e3e87f857c7087e0ef" exitCode=0 Dec 11 02:14:50 crc kubenswrapper[4824]: I1211 02:14:50.808156 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f97jkp" event={"ID":"bb38965f-b3bb-4d2e-8430-8c29298730e4","Type":"ContainerDied","Data":"7e68d1b52f2336bb4c3d2806d6f171b4d3d397e67653d7e3e87f857c7087e0ef"} Dec 11 02:14:51 crc kubenswrapper[4824]: I1211 02:14:51.818675 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-md4vc" event={"ID":"31523d77-e1de-4227-a1f3-be75363b3321","Type":"ContainerStarted","Data":"2e6a3777349dd8a674956f11d6f63ccfade0316c78f1cc5e69a920cdd1b5ed7e"} Dec 11 02:14:51 crc kubenswrapper[4824]: I1211 02:14:51.825254 4824 generic.go:334] "Generic (PLEG): container finished" podID="bb38965f-b3bb-4d2e-8430-8c29298730e4" containerID="b33b4091895eacc87c7455d95e91633a1cd614ac4809ac1daf13c8b9c4e92c03" exitCode=0 Dec 11 02:14:51 crc kubenswrapper[4824]: I1211 02:14:51.825347 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f97jkp" event={"ID":"bb38965f-b3bb-4d2e-8430-8c29298730e4","Type":"ContainerDied","Data":"b33b4091895eacc87c7455d95e91633a1cd614ac4809ac1daf13c8b9c4e92c03"} Dec 11 02:14:52 crc kubenswrapper[4824]: I1211 02:14:52.836185 4824 generic.go:334] "Generic (PLEG): container finished" podID="31523d77-e1de-4227-a1f3-be75363b3321" containerID="2e6a3777349dd8a674956f11d6f63ccfade0316c78f1cc5e69a920cdd1b5ed7e" exitCode=0 Dec 11 02:14:52 crc kubenswrapper[4824]: I1211 02:14:52.836459 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-md4vc" event={"ID":"31523d77-e1de-4227-a1f3-be75363b3321","Type":"ContainerDied","Data":"2e6a3777349dd8a674956f11d6f63ccfade0316c78f1cc5e69a920cdd1b5ed7e"} Dec 11 02:14:53 crc kubenswrapper[4824]: I1211 02:14:53.091077 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f97jkp" Dec 11 02:14:53 crc kubenswrapper[4824]: I1211 02:14:53.237335 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bb38965f-b3bb-4d2e-8430-8c29298730e4-util\") pod \"bb38965f-b3bb-4d2e-8430-8c29298730e4\" (UID: \"bb38965f-b3bb-4d2e-8430-8c29298730e4\") " Dec 11 02:14:53 crc kubenswrapper[4824]: I1211 02:14:53.237466 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bb38965f-b3bb-4d2e-8430-8c29298730e4-bundle\") pod \"bb38965f-b3bb-4d2e-8430-8c29298730e4\" (UID: \"bb38965f-b3bb-4d2e-8430-8c29298730e4\") " Dec 11 02:14:53 crc kubenswrapper[4824]: I1211 02:14:53.237530 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-krrwj\" (UniqueName: \"kubernetes.io/projected/bb38965f-b3bb-4d2e-8430-8c29298730e4-kube-api-access-krrwj\") pod \"bb38965f-b3bb-4d2e-8430-8c29298730e4\" (UID: \"bb38965f-b3bb-4d2e-8430-8c29298730e4\") " Dec 11 02:14:53 crc kubenswrapper[4824]: I1211 02:14:53.238722 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb38965f-b3bb-4d2e-8430-8c29298730e4-bundle" (OuterVolumeSpecName: "bundle") pod "bb38965f-b3bb-4d2e-8430-8c29298730e4" (UID: "bb38965f-b3bb-4d2e-8430-8c29298730e4"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:14:53 crc kubenswrapper[4824]: I1211 02:14:53.245663 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb38965f-b3bb-4d2e-8430-8c29298730e4-kube-api-access-krrwj" (OuterVolumeSpecName: "kube-api-access-krrwj") pod "bb38965f-b3bb-4d2e-8430-8c29298730e4" (UID: "bb38965f-b3bb-4d2e-8430-8c29298730e4"). InnerVolumeSpecName "kube-api-access-krrwj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:14:53 crc kubenswrapper[4824]: I1211 02:14:53.266550 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb38965f-b3bb-4d2e-8430-8c29298730e4-util" (OuterVolumeSpecName: "util") pod "bb38965f-b3bb-4d2e-8430-8c29298730e4" (UID: "bb38965f-b3bb-4d2e-8430-8c29298730e4"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:14:53 crc kubenswrapper[4824]: I1211 02:14:53.338978 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-krrwj\" (UniqueName: \"kubernetes.io/projected/bb38965f-b3bb-4d2e-8430-8c29298730e4-kube-api-access-krrwj\") on node \"crc\" DevicePath \"\"" Dec 11 02:14:53 crc kubenswrapper[4824]: I1211 02:14:53.339011 4824 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bb38965f-b3bb-4d2e-8430-8c29298730e4-util\") on node \"crc\" DevicePath \"\"" Dec 11 02:14:53 crc kubenswrapper[4824]: I1211 02:14:53.339020 4824 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bb38965f-b3bb-4d2e-8430-8c29298730e4-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 02:14:53 crc kubenswrapper[4824]: I1211 02:14:53.849223 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f97jkp" Dec 11 02:14:53 crc kubenswrapper[4824]: I1211 02:14:53.849230 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f97jkp" event={"ID":"bb38965f-b3bb-4d2e-8430-8c29298730e4","Type":"ContainerDied","Data":"a257963bb90690d20a5d3d2d27a8a2f8f6eef7617d0c6e22ef02a6ed1598117f"} Dec 11 02:14:53 crc kubenswrapper[4824]: I1211 02:14:53.849759 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a257963bb90690d20a5d3d2d27a8a2f8f6eef7617d0c6e22ef02a6ed1598117f" Dec 11 02:14:53 crc kubenswrapper[4824]: I1211 02:14:53.853511 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-md4vc" event={"ID":"31523d77-e1de-4227-a1f3-be75363b3321","Type":"ContainerStarted","Data":"2b87c45b76a95a83c96828c78b73a2ef481c25cda3daf2d3b7a53a27946188dd"} Dec 11 02:14:53 crc kubenswrapper[4824]: I1211 02:14:53.879035 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-md4vc" podStartSLOduration=2.169745457 podStartE2EDuration="4.879009132s" podCreationTimestamp="2025-12-11 02:14:49 +0000 UTC" firstStartedPulling="2025-12-11 02:14:50.807311172 +0000 UTC m=+832.496348551" lastFinishedPulling="2025-12-11 02:14:53.516574817 +0000 UTC m=+835.205612226" observedRunningTime="2025-12-11 02:14:53.878422936 +0000 UTC m=+835.567460345" watchObservedRunningTime="2025-12-11 02:14:53.879009132 +0000 UTC m=+835.568046551" Dec 11 02:14:55 crc kubenswrapper[4824]: I1211 02:14:55.097291 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-fd644"] Dec 11 02:14:55 crc kubenswrapper[4824]: E1211 02:14:55.097695 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb38965f-b3bb-4d2e-8430-8c29298730e4" containerName="pull" Dec 11 02:14:55 crc kubenswrapper[4824]: I1211 02:14:55.097728 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb38965f-b3bb-4d2e-8430-8c29298730e4" containerName="pull" Dec 11 02:14:55 crc kubenswrapper[4824]: E1211 02:14:55.097757 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb38965f-b3bb-4d2e-8430-8c29298730e4" containerName="extract" Dec 11 02:14:55 crc kubenswrapper[4824]: I1211 02:14:55.097770 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb38965f-b3bb-4d2e-8430-8c29298730e4" containerName="extract" Dec 11 02:14:55 crc kubenswrapper[4824]: E1211 02:14:55.097795 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb38965f-b3bb-4d2e-8430-8c29298730e4" containerName="util" Dec 11 02:14:55 crc kubenswrapper[4824]: I1211 02:14:55.097809 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb38965f-b3bb-4d2e-8430-8c29298730e4" containerName="util" Dec 11 02:14:55 crc kubenswrapper[4824]: I1211 02:14:55.098023 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb38965f-b3bb-4d2e-8430-8c29298730e4" containerName="extract" Dec 11 02:14:55 crc kubenswrapper[4824]: I1211 02:14:55.098644 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-fd644" Dec 11 02:14:55 crc kubenswrapper[4824]: I1211 02:14:55.101179 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Dec 11 02:14:55 crc kubenswrapper[4824]: I1211 02:14:55.101351 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Dec 11 02:14:55 crc kubenswrapper[4824]: I1211 02:14:55.102196 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-j6w22" Dec 11 02:14:55 crc kubenswrapper[4824]: I1211 02:14:55.108649 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-fd644"] Dec 11 02:14:55 crc kubenswrapper[4824]: I1211 02:14:55.264352 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nsqdh\" (UniqueName: \"kubernetes.io/projected/4ad81e5d-ce6e-4e3c-ae63-7a87ce2a4def-kube-api-access-nsqdh\") pod \"nmstate-operator-5b5b58f5c8-fd644\" (UID: \"4ad81e5d-ce6e-4e3c-ae63-7a87ce2a4def\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-fd644" Dec 11 02:14:55 crc kubenswrapper[4824]: I1211 02:14:55.365195 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nsqdh\" (UniqueName: \"kubernetes.io/projected/4ad81e5d-ce6e-4e3c-ae63-7a87ce2a4def-kube-api-access-nsqdh\") pod \"nmstate-operator-5b5b58f5c8-fd644\" (UID: \"4ad81e5d-ce6e-4e3c-ae63-7a87ce2a4def\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-fd644" Dec 11 02:14:55 crc kubenswrapper[4824]: I1211 02:14:55.387271 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nsqdh\" (UniqueName: \"kubernetes.io/projected/4ad81e5d-ce6e-4e3c-ae63-7a87ce2a4def-kube-api-access-nsqdh\") pod \"nmstate-operator-5b5b58f5c8-fd644\" (UID: \"4ad81e5d-ce6e-4e3c-ae63-7a87ce2a4def\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-fd644" Dec 11 02:14:55 crc kubenswrapper[4824]: I1211 02:14:55.411928 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-fd644" Dec 11 02:14:55 crc kubenswrapper[4824]: I1211 02:14:55.713679 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-fd644"] Dec 11 02:14:55 crc kubenswrapper[4824]: W1211 02:14:55.739550 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4ad81e5d_ce6e_4e3c_ae63_7a87ce2a4def.slice/crio-37000e5b0567f0c34955d29df96c70a31de6ad835dfe53381383c4da0351c87e WatchSource:0}: Error finding container 37000e5b0567f0c34955d29df96c70a31de6ad835dfe53381383c4da0351c87e: Status 404 returned error can't find the container with id 37000e5b0567f0c34955d29df96c70a31de6ad835dfe53381383c4da0351c87e Dec 11 02:14:55 crc kubenswrapper[4824]: I1211 02:14:55.865999 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-fd644" event={"ID":"4ad81e5d-ce6e-4e3c-ae63-7a87ce2a4def","Type":"ContainerStarted","Data":"37000e5b0567f0c34955d29df96c70a31de6ad835dfe53381383c4da0351c87e"} Dec 11 02:14:58 crc kubenswrapper[4824]: I1211 02:14:58.907765 4824 scope.go:117] "RemoveContainer" containerID="f40719f59881e077e7edbab3adb9efd101359bd461e5434c365e119e8cdbddb2" Dec 11 02:14:59 crc kubenswrapper[4824]: I1211 02:14:59.879364 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-md4vc" Dec 11 02:14:59 crc kubenswrapper[4824]: I1211 02:14:59.879740 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-md4vc" Dec 11 02:14:59 crc kubenswrapper[4824]: I1211 02:14:59.892952 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6bkc2_8022f696-d0e7-437b-848a-3cd25bd1f364/kube-multus/2.log" Dec 11 02:14:59 crc kubenswrapper[4824]: I1211 02:14:59.895398 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-fd644" event={"ID":"4ad81e5d-ce6e-4e3c-ae63-7a87ce2a4def","Type":"ContainerStarted","Data":"f1e149e7891cafb44cadf693729e29d13517902949632cc8bb0b420a2e35423f"} Dec 11 02:14:59 crc kubenswrapper[4824]: I1211 02:14:59.932218 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-fd644" podStartSLOduration=1.087892381 podStartE2EDuration="4.932189778s" podCreationTimestamp="2025-12-11 02:14:55 +0000 UTC" firstStartedPulling="2025-12-11 02:14:55.74196369 +0000 UTC m=+837.431001069" lastFinishedPulling="2025-12-11 02:14:59.586261087 +0000 UTC m=+841.275298466" observedRunningTime="2025-12-11 02:14:59.921264367 +0000 UTC m=+841.610301816" watchObservedRunningTime="2025-12-11 02:14:59.932189778 +0000 UTC m=+841.621227227" Dec 11 02:15:00 crc kubenswrapper[4824]: I1211 02:15:00.184187 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423655-kzzdn"] Dec 11 02:15:00 crc kubenswrapper[4824]: I1211 02:15:00.185570 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423655-kzzdn" Dec 11 02:15:00 crc kubenswrapper[4824]: I1211 02:15:00.189694 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 11 02:15:00 crc kubenswrapper[4824]: I1211 02:15:00.190008 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 11 02:15:00 crc kubenswrapper[4824]: I1211 02:15:00.196285 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423655-kzzdn"] Dec 11 02:15:00 crc kubenswrapper[4824]: I1211 02:15:00.357531 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4f5cf3cd-83ba-4408-b280-6bbcdd085ab1-secret-volume\") pod \"collect-profiles-29423655-kzzdn\" (UID: \"4f5cf3cd-83ba-4408-b280-6bbcdd085ab1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423655-kzzdn" Dec 11 02:15:00 crc kubenswrapper[4824]: I1211 02:15:00.357596 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4f5cf3cd-83ba-4408-b280-6bbcdd085ab1-config-volume\") pod \"collect-profiles-29423655-kzzdn\" (UID: \"4f5cf3cd-83ba-4408-b280-6bbcdd085ab1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423655-kzzdn" Dec 11 02:15:00 crc kubenswrapper[4824]: I1211 02:15:00.357688 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wf2sq\" (UniqueName: \"kubernetes.io/projected/4f5cf3cd-83ba-4408-b280-6bbcdd085ab1-kube-api-access-wf2sq\") pod \"collect-profiles-29423655-kzzdn\" (UID: \"4f5cf3cd-83ba-4408-b280-6bbcdd085ab1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423655-kzzdn" Dec 11 02:15:00 crc kubenswrapper[4824]: I1211 02:15:00.458968 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4f5cf3cd-83ba-4408-b280-6bbcdd085ab1-secret-volume\") pod \"collect-profiles-29423655-kzzdn\" (UID: \"4f5cf3cd-83ba-4408-b280-6bbcdd085ab1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423655-kzzdn" Dec 11 02:15:00 crc kubenswrapper[4824]: I1211 02:15:00.459036 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4f5cf3cd-83ba-4408-b280-6bbcdd085ab1-config-volume\") pod \"collect-profiles-29423655-kzzdn\" (UID: \"4f5cf3cd-83ba-4408-b280-6bbcdd085ab1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423655-kzzdn" Dec 11 02:15:00 crc kubenswrapper[4824]: I1211 02:15:00.459087 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wf2sq\" (UniqueName: \"kubernetes.io/projected/4f5cf3cd-83ba-4408-b280-6bbcdd085ab1-kube-api-access-wf2sq\") pod \"collect-profiles-29423655-kzzdn\" (UID: \"4f5cf3cd-83ba-4408-b280-6bbcdd085ab1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423655-kzzdn" Dec 11 02:15:00 crc kubenswrapper[4824]: I1211 02:15:00.460509 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4f5cf3cd-83ba-4408-b280-6bbcdd085ab1-config-volume\") pod \"collect-profiles-29423655-kzzdn\" (UID: \"4f5cf3cd-83ba-4408-b280-6bbcdd085ab1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423655-kzzdn" Dec 11 02:15:00 crc kubenswrapper[4824]: I1211 02:15:00.465797 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4f5cf3cd-83ba-4408-b280-6bbcdd085ab1-secret-volume\") pod \"collect-profiles-29423655-kzzdn\" (UID: \"4f5cf3cd-83ba-4408-b280-6bbcdd085ab1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423655-kzzdn" Dec 11 02:15:00 crc kubenswrapper[4824]: I1211 02:15:00.489577 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wf2sq\" (UniqueName: \"kubernetes.io/projected/4f5cf3cd-83ba-4408-b280-6bbcdd085ab1-kube-api-access-wf2sq\") pod \"collect-profiles-29423655-kzzdn\" (UID: \"4f5cf3cd-83ba-4408-b280-6bbcdd085ab1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423655-kzzdn" Dec 11 02:15:00 crc kubenswrapper[4824]: I1211 02:15:00.505602 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423655-kzzdn" Dec 11 02:15:00 crc kubenswrapper[4824]: I1211 02:15:00.822286 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423655-kzzdn"] Dec 11 02:15:00 crc kubenswrapper[4824]: W1211 02:15:00.828913 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4f5cf3cd_83ba_4408_b280_6bbcdd085ab1.slice/crio-d745853b805eba719591562e1c1991d6e06a92156eb1c19041cdddc0136fe051 WatchSource:0}: Error finding container d745853b805eba719591562e1c1991d6e06a92156eb1c19041cdddc0136fe051: Status 404 returned error can't find the container with id d745853b805eba719591562e1c1991d6e06a92156eb1c19041cdddc0136fe051 Dec 11 02:15:00 crc kubenswrapper[4824]: I1211 02:15:00.901218 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423655-kzzdn" event={"ID":"4f5cf3cd-83ba-4408-b280-6bbcdd085ab1","Type":"ContainerStarted","Data":"d745853b805eba719591562e1c1991d6e06a92156eb1c19041cdddc0136fe051"} Dec 11 02:15:00 crc kubenswrapper[4824]: I1211 02:15:00.952875 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-md4vc" podUID="31523d77-e1de-4227-a1f3-be75363b3321" containerName="registry-server" probeResult="failure" output=< Dec 11 02:15:00 crc kubenswrapper[4824]: timeout: failed to connect service ":50051" within 1s Dec 11 02:15:00 crc kubenswrapper[4824]: > Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.000266 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-qpjg9"] Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.001057 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-qpjg9" Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.008646 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-58hzg" Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.012583 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-8v6lh"] Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.013335 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-8v6lh" Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.016326 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.022657 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-qpjg9"] Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.037173 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-b4h4m"] Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.037799 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-b4h4m" Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.043074 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-8v6lh"] Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.069264 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m967f\" (UniqueName: \"kubernetes.io/projected/2b946a34-6de7-4afe-8876-18b15e09ea98-kube-api-access-m967f\") pod \"nmstate-webhook-5f6d4c5ccb-8v6lh\" (UID: \"2b946a34-6de7-4afe-8876-18b15e09ea98\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-8v6lh" Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.069315 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x7b5f\" (UniqueName: \"kubernetes.io/projected/731b5801-36ca-4f1d-8a34-4ad8554a0c3b-kube-api-access-x7b5f\") pod \"nmstate-metrics-7f946cbc9-qpjg9\" (UID: \"731b5801-36ca-4f1d-8a34-4ad8554a0c3b\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-qpjg9" Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.069344 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/2b946a34-6de7-4afe-8876-18b15e09ea98-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-8v6lh\" (UID: \"2b946a34-6de7-4afe-8876-18b15e09ea98\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-8v6lh" Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.069434 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/22b4f565-3eff-4432-8171-56fa15cf7fcb-nmstate-lock\") pod \"nmstate-handler-b4h4m\" (UID: \"22b4f565-3eff-4432-8171-56fa15cf7fcb\") " pod="openshift-nmstate/nmstate-handler-b4h4m" Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.069499 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/22b4f565-3eff-4432-8171-56fa15cf7fcb-ovs-socket\") pod \"nmstate-handler-b4h4m\" (UID: \"22b4f565-3eff-4432-8171-56fa15cf7fcb\") " pod="openshift-nmstate/nmstate-handler-b4h4m" Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.069564 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7rl49\" (UniqueName: \"kubernetes.io/projected/22b4f565-3eff-4432-8171-56fa15cf7fcb-kube-api-access-7rl49\") pod \"nmstate-handler-b4h4m\" (UID: \"22b4f565-3eff-4432-8171-56fa15cf7fcb\") " pod="openshift-nmstate/nmstate-handler-b4h4m" Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.069592 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/22b4f565-3eff-4432-8171-56fa15cf7fcb-dbus-socket\") pod \"nmstate-handler-b4h4m\" (UID: \"22b4f565-3eff-4432-8171-56fa15cf7fcb\") " pod="openshift-nmstate/nmstate-handler-b4h4m" Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.166396 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-6hcp5"] Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.167431 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-6hcp5" Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.169124 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.169614 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.169999 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-szzxj" Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.172587 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/22b4f565-3eff-4432-8171-56fa15cf7fcb-dbus-socket\") pod \"nmstate-handler-b4h4m\" (UID: \"22b4f565-3eff-4432-8171-56fa15cf7fcb\") " pod="openshift-nmstate/nmstate-handler-b4h4m" Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.172652 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m967f\" (UniqueName: \"kubernetes.io/projected/2b946a34-6de7-4afe-8876-18b15e09ea98-kube-api-access-m967f\") pod \"nmstate-webhook-5f6d4c5ccb-8v6lh\" (UID: \"2b946a34-6de7-4afe-8876-18b15e09ea98\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-8v6lh" Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.172676 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x7b5f\" (UniqueName: \"kubernetes.io/projected/731b5801-36ca-4f1d-8a34-4ad8554a0c3b-kube-api-access-x7b5f\") pod \"nmstate-metrics-7f946cbc9-qpjg9\" (UID: \"731b5801-36ca-4f1d-8a34-4ad8554a0c3b\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-qpjg9" Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.172700 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/2b946a34-6de7-4afe-8876-18b15e09ea98-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-8v6lh\" (UID: \"2b946a34-6de7-4afe-8876-18b15e09ea98\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-8v6lh" Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.172720 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/22b4f565-3eff-4432-8171-56fa15cf7fcb-nmstate-lock\") pod \"nmstate-handler-b4h4m\" (UID: \"22b4f565-3eff-4432-8171-56fa15cf7fcb\") " pod="openshift-nmstate/nmstate-handler-b4h4m" Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.172758 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/22b4f565-3eff-4432-8171-56fa15cf7fcb-nmstate-lock\") pod \"nmstate-handler-b4h4m\" (UID: \"22b4f565-3eff-4432-8171-56fa15cf7fcb\") " pod="openshift-nmstate/nmstate-handler-b4h4m" Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.172955 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/22b4f565-3eff-4432-8171-56fa15cf7fcb-dbus-socket\") pod \"nmstate-handler-b4h4m\" (UID: \"22b4f565-3eff-4432-8171-56fa15cf7fcb\") " pod="openshift-nmstate/nmstate-handler-b4h4m" Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.173185 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/22b4f565-3eff-4432-8171-56fa15cf7fcb-ovs-socket\") pod \"nmstate-handler-b4h4m\" (UID: \"22b4f565-3eff-4432-8171-56fa15cf7fcb\") " pod="openshift-nmstate/nmstate-handler-b4h4m" Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.173261 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7rl49\" (UniqueName: \"kubernetes.io/projected/22b4f565-3eff-4432-8171-56fa15cf7fcb-kube-api-access-7rl49\") pod \"nmstate-handler-b4h4m\" (UID: \"22b4f565-3eff-4432-8171-56fa15cf7fcb\") " pod="openshift-nmstate/nmstate-handler-b4h4m" Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.173587 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/22b4f565-3eff-4432-8171-56fa15cf7fcb-ovs-socket\") pod \"nmstate-handler-b4h4m\" (UID: \"22b4f565-3eff-4432-8171-56fa15cf7fcb\") " pod="openshift-nmstate/nmstate-handler-b4h4m" Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.178257 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-6hcp5"] Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.179639 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/2b946a34-6de7-4afe-8876-18b15e09ea98-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-8v6lh\" (UID: \"2b946a34-6de7-4afe-8876-18b15e09ea98\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-8v6lh" Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.193328 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x7b5f\" (UniqueName: \"kubernetes.io/projected/731b5801-36ca-4f1d-8a34-4ad8554a0c3b-kube-api-access-x7b5f\") pod \"nmstate-metrics-7f946cbc9-qpjg9\" (UID: \"731b5801-36ca-4f1d-8a34-4ad8554a0c3b\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-qpjg9" Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.195460 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7rl49\" (UniqueName: \"kubernetes.io/projected/22b4f565-3eff-4432-8171-56fa15cf7fcb-kube-api-access-7rl49\") pod \"nmstate-handler-b4h4m\" (UID: \"22b4f565-3eff-4432-8171-56fa15cf7fcb\") " pod="openshift-nmstate/nmstate-handler-b4h4m" Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.201188 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m967f\" (UniqueName: \"kubernetes.io/projected/2b946a34-6de7-4afe-8876-18b15e09ea98-kube-api-access-m967f\") pod \"nmstate-webhook-5f6d4c5ccb-8v6lh\" (UID: \"2b946a34-6de7-4afe-8876-18b15e09ea98\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-8v6lh" Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.274148 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/ffe586ab-061d-47f6-a3e8-220bf9ffebe2-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-6hcp5\" (UID: \"ffe586ab-061d-47f6-a3e8-220bf9ffebe2\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-6hcp5" Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.274211 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/ffe586ab-061d-47f6-a3e8-220bf9ffebe2-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-6hcp5\" (UID: \"ffe586ab-061d-47f6-a3e8-220bf9ffebe2\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-6hcp5" Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.274316 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4k6hf\" (UniqueName: \"kubernetes.io/projected/ffe586ab-061d-47f6-a3e8-220bf9ffebe2-kube-api-access-4k6hf\") pod \"nmstate-console-plugin-7fbb5f6569-6hcp5\" (UID: \"ffe586ab-061d-47f6-a3e8-220bf9ffebe2\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-6hcp5" Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.317825 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-qpjg9" Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.319478 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-75bc6c8444-5tv6n"] Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.320260 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-75bc6c8444-5tv6n" Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.334370 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-8v6lh" Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.341146 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-75bc6c8444-5tv6n"] Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.356397 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-b4h4m" Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.375203 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/63454e54-67b7-4c08-8693-db7fd93a5fd9-console-serving-cert\") pod \"console-75bc6c8444-5tv6n\" (UID: \"63454e54-67b7-4c08-8693-db7fd93a5fd9\") " pod="openshift-console/console-75bc6c8444-5tv6n" Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.375245 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/63454e54-67b7-4c08-8693-db7fd93a5fd9-console-oauth-config\") pod \"console-75bc6c8444-5tv6n\" (UID: \"63454e54-67b7-4c08-8693-db7fd93a5fd9\") " pod="openshift-console/console-75bc6c8444-5tv6n" Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.375278 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/ffe586ab-061d-47f6-a3e8-220bf9ffebe2-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-6hcp5\" (UID: \"ffe586ab-061d-47f6-a3e8-220bf9ffebe2\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-6hcp5" Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.375336 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/ffe586ab-061d-47f6-a3e8-220bf9ffebe2-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-6hcp5\" (UID: \"ffe586ab-061d-47f6-a3e8-220bf9ffebe2\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-6hcp5" Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.375395 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4k6hf\" (UniqueName: \"kubernetes.io/projected/ffe586ab-061d-47f6-a3e8-220bf9ffebe2-kube-api-access-4k6hf\") pod \"nmstate-console-plugin-7fbb5f6569-6hcp5\" (UID: \"ffe586ab-061d-47f6-a3e8-220bf9ffebe2\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-6hcp5" Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.375433 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/63454e54-67b7-4c08-8693-db7fd93a5fd9-service-ca\") pod \"console-75bc6c8444-5tv6n\" (UID: \"63454e54-67b7-4c08-8693-db7fd93a5fd9\") " pod="openshift-console/console-75bc6c8444-5tv6n" Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.375631 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4ttk2\" (UniqueName: \"kubernetes.io/projected/63454e54-67b7-4c08-8693-db7fd93a5fd9-kube-api-access-4ttk2\") pod \"console-75bc6c8444-5tv6n\" (UID: \"63454e54-67b7-4c08-8693-db7fd93a5fd9\") " pod="openshift-console/console-75bc6c8444-5tv6n" Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.375659 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/63454e54-67b7-4c08-8693-db7fd93a5fd9-console-config\") pod \"console-75bc6c8444-5tv6n\" (UID: \"63454e54-67b7-4c08-8693-db7fd93a5fd9\") " pod="openshift-console/console-75bc6c8444-5tv6n" Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.375681 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/63454e54-67b7-4c08-8693-db7fd93a5fd9-oauth-serving-cert\") pod \"console-75bc6c8444-5tv6n\" (UID: \"63454e54-67b7-4c08-8693-db7fd93a5fd9\") " pod="openshift-console/console-75bc6c8444-5tv6n" Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.375713 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/63454e54-67b7-4c08-8693-db7fd93a5fd9-trusted-ca-bundle\") pod \"console-75bc6c8444-5tv6n\" (UID: \"63454e54-67b7-4c08-8693-db7fd93a5fd9\") " pod="openshift-console/console-75bc6c8444-5tv6n" Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.376461 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/ffe586ab-061d-47f6-a3e8-220bf9ffebe2-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-6hcp5\" (UID: \"ffe586ab-061d-47f6-a3e8-220bf9ffebe2\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-6hcp5" Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.379323 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/ffe586ab-061d-47f6-a3e8-220bf9ffebe2-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-6hcp5\" (UID: \"ffe586ab-061d-47f6-a3e8-220bf9ffebe2\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-6hcp5" Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.391439 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4k6hf\" (UniqueName: \"kubernetes.io/projected/ffe586ab-061d-47f6-a3e8-220bf9ffebe2-kube-api-access-4k6hf\") pod \"nmstate-console-plugin-7fbb5f6569-6hcp5\" (UID: \"ffe586ab-061d-47f6-a3e8-220bf9ffebe2\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-6hcp5" Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.477594 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/63454e54-67b7-4c08-8693-db7fd93a5fd9-service-ca\") pod \"console-75bc6c8444-5tv6n\" (UID: \"63454e54-67b7-4c08-8693-db7fd93a5fd9\") " pod="openshift-console/console-75bc6c8444-5tv6n" Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.477837 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4ttk2\" (UniqueName: \"kubernetes.io/projected/63454e54-67b7-4c08-8693-db7fd93a5fd9-kube-api-access-4ttk2\") pod \"console-75bc6c8444-5tv6n\" (UID: \"63454e54-67b7-4c08-8693-db7fd93a5fd9\") " pod="openshift-console/console-75bc6c8444-5tv6n" Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.477860 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/63454e54-67b7-4c08-8693-db7fd93a5fd9-console-config\") pod \"console-75bc6c8444-5tv6n\" (UID: \"63454e54-67b7-4c08-8693-db7fd93a5fd9\") " pod="openshift-console/console-75bc6c8444-5tv6n" Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.477877 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/63454e54-67b7-4c08-8693-db7fd93a5fd9-oauth-serving-cert\") pod \"console-75bc6c8444-5tv6n\" (UID: \"63454e54-67b7-4c08-8693-db7fd93a5fd9\") " pod="openshift-console/console-75bc6c8444-5tv6n" Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.477899 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/63454e54-67b7-4c08-8693-db7fd93a5fd9-trusted-ca-bundle\") pod \"console-75bc6c8444-5tv6n\" (UID: \"63454e54-67b7-4c08-8693-db7fd93a5fd9\") " pod="openshift-console/console-75bc6c8444-5tv6n" Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.477922 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/63454e54-67b7-4c08-8693-db7fd93a5fd9-console-oauth-config\") pod \"console-75bc6c8444-5tv6n\" (UID: \"63454e54-67b7-4c08-8693-db7fd93a5fd9\") " pod="openshift-console/console-75bc6c8444-5tv6n" Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.477937 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/63454e54-67b7-4c08-8693-db7fd93a5fd9-console-serving-cert\") pod \"console-75bc6c8444-5tv6n\" (UID: \"63454e54-67b7-4c08-8693-db7fd93a5fd9\") " pod="openshift-console/console-75bc6c8444-5tv6n" Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.478663 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/63454e54-67b7-4c08-8693-db7fd93a5fd9-service-ca\") pod \"console-75bc6c8444-5tv6n\" (UID: \"63454e54-67b7-4c08-8693-db7fd93a5fd9\") " pod="openshift-console/console-75bc6c8444-5tv6n" Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.478818 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/63454e54-67b7-4c08-8693-db7fd93a5fd9-oauth-serving-cert\") pod \"console-75bc6c8444-5tv6n\" (UID: \"63454e54-67b7-4c08-8693-db7fd93a5fd9\") " pod="openshift-console/console-75bc6c8444-5tv6n" Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.479891 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/63454e54-67b7-4c08-8693-db7fd93a5fd9-trusted-ca-bundle\") pod \"console-75bc6c8444-5tv6n\" (UID: \"63454e54-67b7-4c08-8693-db7fd93a5fd9\") " pod="openshift-console/console-75bc6c8444-5tv6n" Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.482042 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-6hcp5" Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.482179 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/63454e54-67b7-4c08-8693-db7fd93a5fd9-console-config\") pod \"console-75bc6c8444-5tv6n\" (UID: \"63454e54-67b7-4c08-8693-db7fd93a5fd9\") " pod="openshift-console/console-75bc6c8444-5tv6n" Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.484672 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/63454e54-67b7-4c08-8693-db7fd93a5fd9-console-serving-cert\") pod \"console-75bc6c8444-5tv6n\" (UID: \"63454e54-67b7-4c08-8693-db7fd93a5fd9\") " pod="openshift-console/console-75bc6c8444-5tv6n" Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.484893 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/63454e54-67b7-4c08-8693-db7fd93a5fd9-console-oauth-config\") pod \"console-75bc6c8444-5tv6n\" (UID: \"63454e54-67b7-4c08-8693-db7fd93a5fd9\") " pod="openshift-console/console-75bc6c8444-5tv6n" Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.498354 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4ttk2\" (UniqueName: \"kubernetes.io/projected/63454e54-67b7-4c08-8693-db7fd93a5fd9-kube-api-access-4ttk2\") pod \"console-75bc6c8444-5tv6n\" (UID: \"63454e54-67b7-4c08-8693-db7fd93a5fd9\") " pod="openshift-console/console-75bc6c8444-5tv6n" Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.583824 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-8v6lh"] Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.631877 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-75bc6c8444-5tv6n" Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.660503 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-6hcp5"] Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.721533 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-qpjg9"] Dec 11 02:15:01 crc kubenswrapper[4824]: W1211 02:15:01.730252 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod731b5801_36ca_4f1d_8a34_4ad8554a0c3b.slice/crio-f34adb46de8a12cf68bad9fd080a791ef20dcf5c0e90a91e665930b1fef059ce WatchSource:0}: Error finding container f34adb46de8a12cf68bad9fd080a791ef20dcf5c0e90a91e665930b1fef059ce: Status 404 returned error can't find the container with id f34adb46de8a12cf68bad9fd080a791ef20dcf5c0e90a91e665930b1fef059ce Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.818370 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-75bc6c8444-5tv6n"] Dec 11 02:15:01 crc kubenswrapper[4824]: W1211 02:15:01.829622 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod63454e54_67b7_4c08_8693_db7fd93a5fd9.slice/crio-5e038c0bf4d66a61d30c2aaf178ecaba64702c745dea99bd5ccbfaed8ec7651e WatchSource:0}: Error finding container 5e038c0bf4d66a61d30c2aaf178ecaba64702c745dea99bd5ccbfaed8ec7651e: Status 404 returned error can't find the container with id 5e038c0bf4d66a61d30c2aaf178ecaba64702c745dea99bd5ccbfaed8ec7651e Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.908575 4824 generic.go:334] "Generic (PLEG): container finished" podID="4f5cf3cd-83ba-4408-b280-6bbcdd085ab1" containerID="3e921b7d8cabc3f5640f6ce4cff6e9867224ae1ceb66ae93c4747189d2f3e643" exitCode=0 Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.908674 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423655-kzzdn" event={"ID":"4f5cf3cd-83ba-4408-b280-6bbcdd085ab1","Type":"ContainerDied","Data":"3e921b7d8cabc3f5640f6ce4cff6e9867224ae1ceb66ae93c4747189d2f3e643"} Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.911209 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-b4h4m" event={"ID":"22b4f565-3eff-4432-8171-56fa15cf7fcb","Type":"ContainerStarted","Data":"8788de4bdd41604ee5dd90d91bf9017b8e2ca1e3817fa8464214a2835c4d45e3"} Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.912619 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-8v6lh" event={"ID":"2b946a34-6de7-4afe-8876-18b15e09ea98","Type":"ContainerStarted","Data":"98f2ef38ad61acee5a466542baf647e5a5085b0eb03b5baca344dee5cc45c785"} Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.915372 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-qpjg9" event={"ID":"731b5801-36ca-4f1d-8a34-4ad8554a0c3b","Type":"ContainerStarted","Data":"f34adb46de8a12cf68bad9fd080a791ef20dcf5c0e90a91e665930b1fef059ce"} Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.916733 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-75bc6c8444-5tv6n" event={"ID":"63454e54-67b7-4c08-8693-db7fd93a5fd9","Type":"ContainerStarted","Data":"5e038c0bf4d66a61d30c2aaf178ecaba64702c745dea99bd5ccbfaed8ec7651e"} Dec 11 02:15:01 crc kubenswrapper[4824]: I1211 02:15:01.917802 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-6hcp5" event={"ID":"ffe586ab-061d-47f6-a3e8-220bf9ffebe2","Type":"ContainerStarted","Data":"4c862cbea24b0086b76d67280cb1afbe1a1e40fe8ca2864de377182517e0e1be"} Dec 11 02:15:02 crc kubenswrapper[4824]: I1211 02:15:02.927910 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-75bc6c8444-5tv6n" event={"ID":"63454e54-67b7-4c08-8693-db7fd93a5fd9","Type":"ContainerStarted","Data":"a7564522d4f721e092e43cfabee5bc2b3b2dde0ac6834429fcf4ea8cb51ee2f5"} Dec 11 02:15:02 crc kubenswrapper[4824]: I1211 02:15:02.948522 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-75bc6c8444-5tv6n" podStartSLOduration=1.9484954239999999 podStartE2EDuration="1.948495424s" podCreationTimestamp="2025-12-11 02:15:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:15:02.946659143 +0000 UTC m=+844.635696612" watchObservedRunningTime="2025-12-11 02:15:02.948495424 +0000 UTC m=+844.637532833" Dec 11 02:15:03 crc kubenswrapper[4824]: I1211 02:15:03.229342 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423655-kzzdn" Dec 11 02:15:03 crc kubenswrapper[4824]: I1211 02:15:03.407876 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wf2sq\" (UniqueName: \"kubernetes.io/projected/4f5cf3cd-83ba-4408-b280-6bbcdd085ab1-kube-api-access-wf2sq\") pod \"4f5cf3cd-83ba-4408-b280-6bbcdd085ab1\" (UID: \"4f5cf3cd-83ba-4408-b280-6bbcdd085ab1\") " Dec 11 02:15:03 crc kubenswrapper[4824]: I1211 02:15:03.407993 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4f5cf3cd-83ba-4408-b280-6bbcdd085ab1-config-volume\") pod \"4f5cf3cd-83ba-4408-b280-6bbcdd085ab1\" (UID: \"4f5cf3cd-83ba-4408-b280-6bbcdd085ab1\") " Dec 11 02:15:03 crc kubenswrapper[4824]: I1211 02:15:03.408102 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4f5cf3cd-83ba-4408-b280-6bbcdd085ab1-secret-volume\") pod \"4f5cf3cd-83ba-4408-b280-6bbcdd085ab1\" (UID: \"4f5cf3cd-83ba-4408-b280-6bbcdd085ab1\") " Dec 11 02:15:03 crc kubenswrapper[4824]: I1211 02:15:03.408929 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4f5cf3cd-83ba-4408-b280-6bbcdd085ab1-config-volume" (OuterVolumeSpecName: "config-volume") pod "4f5cf3cd-83ba-4408-b280-6bbcdd085ab1" (UID: "4f5cf3cd-83ba-4408-b280-6bbcdd085ab1"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:15:03 crc kubenswrapper[4824]: I1211 02:15:03.414048 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f5cf3cd-83ba-4408-b280-6bbcdd085ab1-kube-api-access-wf2sq" (OuterVolumeSpecName: "kube-api-access-wf2sq") pod "4f5cf3cd-83ba-4408-b280-6bbcdd085ab1" (UID: "4f5cf3cd-83ba-4408-b280-6bbcdd085ab1"). InnerVolumeSpecName "kube-api-access-wf2sq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:15:03 crc kubenswrapper[4824]: I1211 02:15:03.414249 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f5cf3cd-83ba-4408-b280-6bbcdd085ab1-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "4f5cf3cd-83ba-4408-b280-6bbcdd085ab1" (UID: "4f5cf3cd-83ba-4408-b280-6bbcdd085ab1"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:15:03 crc kubenswrapper[4824]: I1211 02:15:03.509674 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wf2sq\" (UniqueName: \"kubernetes.io/projected/4f5cf3cd-83ba-4408-b280-6bbcdd085ab1-kube-api-access-wf2sq\") on node \"crc\" DevicePath \"\"" Dec 11 02:15:03 crc kubenswrapper[4824]: I1211 02:15:03.509719 4824 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4f5cf3cd-83ba-4408-b280-6bbcdd085ab1-config-volume\") on node \"crc\" DevicePath \"\"" Dec 11 02:15:03 crc kubenswrapper[4824]: I1211 02:15:03.509732 4824 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4f5cf3cd-83ba-4408-b280-6bbcdd085ab1-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 11 02:15:03 crc kubenswrapper[4824]: I1211 02:15:03.934983 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423655-kzzdn" Dec 11 02:15:03 crc kubenswrapper[4824]: I1211 02:15:03.934981 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423655-kzzdn" event={"ID":"4f5cf3cd-83ba-4408-b280-6bbcdd085ab1","Type":"ContainerDied","Data":"d745853b805eba719591562e1c1991d6e06a92156eb1c19041cdddc0136fe051"} Dec 11 02:15:03 crc kubenswrapper[4824]: I1211 02:15:03.935130 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d745853b805eba719591562e1c1991d6e06a92156eb1c19041cdddc0136fe051" Dec 11 02:15:06 crc kubenswrapper[4824]: I1211 02:15:06.957207 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-qpjg9" event={"ID":"731b5801-36ca-4f1d-8a34-4ad8554a0c3b","Type":"ContainerStarted","Data":"876a23199ec5fbe3c89f1b0233866a40a493149e92f0626de8efe7c19179d5f5"} Dec 11 02:15:06 crc kubenswrapper[4824]: I1211 02:15:06.960791 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-6hcp5" event={"ID":"ffe586ab-061d-47f6-a3e8-220bf9ffebe2","Type":"ContainerStarted","Data":"b84efc8b2620800b42823e361a45520f3314dbe75bc87ac1974845cc1c98a315"} Dec 11 02:15:06 crc kubenswrapper[4824]: I1211 02:15:06.966395 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-b4h4m" event={"ID":"22b4f565-3eff-4432-8171-56fa15cf7fcb","Type":"ContainerStarted","Data":"273972dde4138778e4112f36477cc4c0409ef71d779678ccdc9e7d080cf8440f"} Dec 11 02:15:06 crc kubenswrapper[4824]: I1211 02:15:06.966462 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-b4h4m" Dec 11 02:15:06 crc kubenswrapper[4824]: I1211 02:15:06.968797 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-8v6lh" event={"ID":"2b946a34-6de7-4afe-8876-18b15e09ea98","Type":"ContainerStarted","Data":"7d3055d656c4b4ee00f8e08959c26bddeb70f4522a6ff90b5b274619079f45be"} Dec 11 02:15:06 crc kubenswrapper[4824]: I1211 02:15:06.968947 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-8v6lh" Dec 11 02:15:06 crc kubenswrapper[4824]: I1211 02:15:06.994541 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-6hcp5" podStartSLOduration=1.12045729 podStartE2EDuration="5.994514449s" podCreationTimestamp="2025-12-11 02:15:01 +0000 UTC" firstStartedPulling="2025-12-11 02:15:01.664635181 +0000 UTC m=+843.353672560" lastFinishedPulling="2025-12-11 02:15:06.53869231 +0000 UTC m=+848.227729719" observedRunningTime="2025-12-11 02:15:06.987973459 +0000 UTC m=+848.677010848" watchObservedRunningTime="2025-12-11 02:15:06.994514449 +0000 UTC m=+848.683551838" Dec 11 02:15:07 crc kubenswrapper[4824]: I1211 02:15:07.018803 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-b4h4m" podStartSLOduration=0.836251429 podStartE2EDuration="6.018785558s" podCreationTimestamp="2025-12-11 02:15:01 +0000 UTC" firstStartedPulling="2025-12-11 02:15:01.376827871 +0000 UTC m=+843.065865250" lastFinishedPulling="2025-12-11 02:15:06.55936196 +0000 UTC m=+848.248399379" observedRunningTime="2025-12-11 02:15:07.015682263 +0000 UTC m=+848.704719652" watchObservedRunningTime="2025-12-11 02:15:07.018785558 +0000 UTC m=+848.707822947" Dec 11 02:15:07 crc kubenswrapper[4824]: I1211 02:15:07.040640 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-8v6lh" podStartSLOduration=2.042851762 podStartE2EDuration="7.04062076s" podCreationTimestamp="2025-12-11 02:15:00 +0000 UTC" firstStartedPulling="2025-12-11 02:15:01.595379233 +0000 UTC m=+843.284416612" lastFinishedPulling="2025-12-11 02:15:06.593148191 +0000 UTC m=+848.282185610" observedRunningTime="2025-12-11 02:15:07.036245019 +0000 UTC m=+848.725282428" watchObservedRunningTime="2025-12-11 02:15:07.04062076 +0000 UTC m=+848.729658149" Dec 11 02:15:09 crc kubenswrapper[4824]: I1211 02:15:09.948557 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-md4vc" Dec 11 02:15:09 crc kubenswrapper[4824]: I1211 02:15:09.994661 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-qpjg9" event={"ID":"731b5801-36ca-4f1d-8a34-4ad8554a0c3b","Type":"ContainerStarted","Data":"0576f8b0eba353625c374d2668955f352acce40a4305f5f62b313e7b15a5dcff"} Dec 11 02:15:10 crc kubenswrapper[4824]: I1211 02:15:10.021784 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-qpjg9" podStartSLOduration=2.203975812 podStartE2EDuration="10.021756065s" podCreationTimestamp="2025-12-11 02:15:00 +0000 UTC" firstStartedPulling="2025-12-11 02:15:01.73281899 +0000 UTC m=+843.421856369" lastFinishedPulling="2025-12-11 02:15:09.550599233 +0000 UTC m=+851.239636622" observedRunningTime="2025-12-11 02:15:10.01470021 +0000 UTC m=+851.703737629" watchObservedRunningTime="2025-12-11 02:15:10.021756065 +0000 UTC m=+851.710793484" Dec 11 02:15:10 crc kubenswrapper[4824]: I1211 02:15:10.047721 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-md4vc" Dec 11 02:15:10 crc kubenswrapper[4824]: I1211 02:15:10.196136 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-md4vc"] Dec 11 02:15:11 crc kubenswrapper[4824]: I1211 02:15:11.005418 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-md4vc" podUID="31523d77-e1de-4227-a1f3-be75363b3321" containerName="registry-server" containerID="cri-o://2b87c45b76a95a83c96828c78b73a2ef481c25cda3daf2d3b7a53a27946188dd" gracePeriod=2 Dec 11 02:15:11 crc kubenswrapper[4824]: I1211 02:15:11.386708 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-b4h4m" Dec 11 02:15:11 crc kubenswrapper[4824]: I1211 02:15:11.424056 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-md4vc" Dec 11 02:15:11 crc kubenswrapper[4824]: I1211 02:15:11.521782 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/31523d77-e1de-4227-a1f3-be75363b3321-utilities\") pod \"31523d77-e1de-4227-a1f3-be75363b3321\" (UID: \"31523d77-e1de-4227-a1f3-be75363b3321\") " Dec 11 02:15:11 crc kubenswrapper[4824]: I1211 02:15:11.521859 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5fhzw\" (UniqueName: \"kubernetes.io/projected/31523d77-e1de-4227-a1f3-be75363b3321-kube-api-access-5fhzw\") pod \"31523d77-e1de-4227-a1f3-be75363b3321\" (UID: \"31523d77-e1de-4227-a1f3-be75363b3321\") " Dec 11 02:15:11 crc kubenswrapper[4824]: I1211 02:15:11.521947 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/31523d77-e1de-4227-a1f3-be75363b3321-catalog-content\") pod \"31523d77-e1de-4227-a1f3-be75363b3321\" (UID: \"31523d77-e1de-4227-a1f3-be75363b3321\") " Dec 11 02:15:11 crc kubenswrapper[4824]: I1211 02:15:11.525746 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/31523d77-e1de-4227-a1f3-be75363b3321-utilities" (OuterVolumeSpecName: "utilities") pod "31523d77-e1de-4227-a1f3-be75363b3321" (UID: "31523d77-e1de-4227-a1f3-be75363b3321"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:15:11 crc kubenswrapper[4824]: I1211 02:15:11.531143 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31523d77-e1de-4227-a1f3-be75363b3321-kube-api-access-5fhzw" (OuterVolumeSpecName: "kube-api-access-5fhzw") pod "31523d77-e1de-4227-a1f3-be75363b3321" (UID: "31523d77-e1de-4227-a1f3-be75363b3321"). InnerVolumeSpecName "kube-api-access-5fhzw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:15:11 crc kubenswrapper[4824]: I1211 02:15:11.624072 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/31523d77-e1de-4227-a1f3-be75363b3321-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 02:15:11 crc kubenswrapper[4824]: I1211 02:15:11.624188 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5fhzw\" (UniqueName: \"kubernetes.io/projected/31523d77-e1de-4227-a1f3-be75363b3321-kube-api-access-5fhzw\") on node \"crc\" DevicePath \"\"" Dec 11 02:15:11 crc kubenswrapper[4824]: I1211 02:15:11.632812 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-75bc6c8444-5tv6n" Dec 11 02:15:11 crc kubenswrapper[4824]: I1211 02:15:11.632897 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-75bc6c8444-5tv6n" Dec 11 02:15:11 crc kubenswrapper[4824]: I1211 02:15:11.639437 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-75bc6c8444-5tv6n" Dec 11 02:15:11 crc kubenswrapper[4824]: I1211 02:15:11.685748 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/31523d77-e1de-4227-a1f3-be75363b3321-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "31523d77-e1de-4227-a1f3-be75363b3321" (UID: "31523d77-e1de-4227-a1f3-be75363b3321"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:15:11 crc kubenswrapper[4824]: I1211 02:15:11.725537 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/31523d77-e1de-4227-a1f3-be75363b3321-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 02:15:12 crc kubenswrapper[4824]: I1211 02:15:12.015310 4824 generic.go:334] "Generic (PLEG): container finished" podID="31523d77-e1de-4227-a1f3-be75363b3321" containerID="2b87c45b76a95a83c96828c78b73a2ef481c25cda3daf2d3b7a53a27946188dd" exitCode=0 Dec 11 02:15:12 crc kubenswrapper[4824]: I1211 02:15:12.015381 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-md4vc" event={"ID":"31523d77-e1de-4227-a1f3-be75363b3321","Type":"ContainerDied","Data":"2b87c45b76a95a83c96828c78b73a2ef481c25cda3daf2d3b7a53a27946188dd"} Dec 11 02:15:12 crc kubenswrapper[4824]: I1211 02:15:12.015423 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-md4vc" Dec 11 02:15:12 crc kubenswrapper[4824]: I1211 02:15:12.015488 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-md4vc" event={"ID":"31523d77-e1de-4227-a1f3-be75363b3321","Type":"ContainerDied","Data":"fddaef2efa18822eabe257fe340a93757cfd04a49d3589d5a0fc79b319029fca"} Dec 11 02:15:12 crc kubenswrapper[4824]: I1211 02:15:12.015551 4824 scope.go:117] "RemoveContainer" containerID="2b87c45b76a95a83c96828c78b73a2ef481c25cda3daf2d3b7a53a27946188dd" Dec 11 02:15:12 crc kubenswrapper[4824]: I1211 02:15:12.025461 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-75bc6c8444-5tv6n" Dec 11 02:15:12 crc kubenswrapper[4824]: I1211 02:15:12.036705 4824 scope.go:117] "RemoveContainer" containerID="2e6a3777349dd8a674956f11d6f63ccfade0316c78f1cc5e69a920cdd1b5ed7e" Dec 11 02:15:12 crc kubenswrapper[4824]: I1211 02:15:12.095039 4824 scope.go:117] "RemoveContainer" containerID="ba63082c996f39244eb9d5eb4fea9097d12f6218c34bca7a452de9a2aba18b9e" Dec 11 02:15:12 crc kubenswrapper[4824]: I1211 02:15:12.113739 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-md4vc"] Dec 11 02:15:12 crc kubenswrapper[4824]: I1211 02:15:12.120822 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-md4vc"] Dec 11 02:15:12 crc kubenswrapper[4824]: I1211 02:15:12.125904 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-k6kls"] Dec 11 02:15:12 crc kubenswrapper[4824]: I1211 02:15:12.154334 4824 scope.go:117] "RemoveContainer" containerID="2b87c45b76a95a83c96828c78b73a2ef481c25cda3daf2d3b7a53a27946188dd" Dec 11 02:15:12 crc kubenswrapper[4824]: E1211 02:15:12.158416 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b87c45b76a95a83c96828c78b73a2ef481c25cda3daf2d3b7a53a27946188dd\": container with ID starting with 2b87c45b76a95a83c96828c78b73a2ef481c25cda3daf2d3b7a53a27946188dd not found: ID does not exist" containerID="2b87c45b76a95a83c96828c78b73a2ef481c25cda3daf2d3b7a53a27946188dd" Dec 11 02:15:12 crc kubenswrapper[4824]: I1211 02:15:12.158472 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b87c45b76a95a83c96828c78b73a2ef481c25cda3daf2d3b7a53a27946188dd"} err="failed to get container status \"2b87c45b76a95a83c96828c78b73a2ef481c25cda3daf2d3b7a53a27946188dd\": rpc error: code = NotFound desc = could not find container \"2b87c45b76a95a83c96828c78b73a2ef481c25cda3daf2d3b7a53a27946188dd\": container with ID starting with 2b87c45b76a95a83c96828c78b73a2ef481c25cda3daf2d3b7a53a27946188dd not found: ID does not exist" Dec 11 02:15:12 crc kubenswrapper[4824]: I1211 02:15:12.158511 4824 scope.go:117] "RemoveContainer" containerID="2e6a3777349dd8a674956f11d6f63ccfade0316c78f1cc5e69a920cdd1b5ed7e" Dec 11 02:15:12 crc kubenswrapper[4824]: E1211 02:15:12.158888 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e6a3777349dd8a674956f11d6f63ccfade0316c78f1cc5e69a920cdd1b5ed7e\": container with ID starting with 2e6a3777349dd8a674956f11d6f63ccfade0316c78f1cc5e69a920cdd1b5ed7e not found: ID does not exist" containerID="2e6a3777349dd8a674956f11d6f63ccfade0316c78f1cc5e69a920cdd1b5ed7e" Dec 11 02:15:12 crc kubenswrapper[4824]: I1211 02:15:12.158935 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e6a3777349dd8a674956f11d6f63ccfade0316c78f1cc5e69a920cdd1b5ed7e"} err="failed to get container status \"2e6a3777349dd8a674956f11d6f63ccfade0316c78f1cc5e69a920cdd1b5ed7e\": rpc error: code = NotFound desc = could not find container \"2e6a3777349dd8a674956f11d6f63ccfade0316c78f1cc5e69a920cdd1b5ed7e\": container with ID starting with 2e6a3777349dd8a674956f11d6f63ccfade0316c78f1cc5e69a920cdd1b5ed7e not found: ID does not exist" Dec 11 02:15:12 crc kubenswrapper[4824]: I1211 02:15:12.158975 4824 scope.go:117] "RemoveContainer" containerID="ba63082c996f39244eb9d5eb4fea9097d12f6218c34bca7a452de9a2aba18b9e" Dec 11 02:15:12 crc kubenswrapper[4824]: E1211 02:15:12.159291 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba63082c996f39244eb9d5eb4fea9097d12f6218c34bca7a452de9a2aba18b9e\": container with ID starting with ba63082c996f39244eb9d5eb4fea9097d12f6218c34bca7a452de9a2aba18b9e not found: ID does not exist" containerID="ba63082c996f39244eb9d5eb4fea9097d12f6218c34bca7a452de9a2aba18b9e" Dec 11 02:15:12 crc kubenswrapper[4824]: I1211 02:15:12.159309 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba63082c996f39244eb9d5eb4fea9097d12f6218c34bca7a452de9a2aba18b9e"} err="failed to get container status \"ba63082c996f39244eb9d5eb4fea9097d12f6218c34bca7a452de9a2aba18b9e\": rpc error: code = NotFound desc = could not find container \"ba63082c996f39244eb9d5eb4fea9097d12f6218c34bca7a452de9a2aba18b9e\": container with ID starting with ba63082c996f39244eb9d5eb4fea9097d12f6218c34bca7a452de9a2aba18b9e not found: ID does not exist" Dec 11 02:15:12 crc kubenswrapper[4824]: I1211 02:15:12.646860 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31523d77-e1de-4227-a1f3-be75363b3321" path="/var/lib/kubelet/pods/31523d77-e1de-4227-a1f3-be75363b3321/volumes" Dec 11 02:15:21 crc kubenswrapper[4824]: I1211 02:15:21.344458 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-8v6lh" Dec 11 02:15:32 crc kubenswrapper[4824]: I1211 02:15:32.154077 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-nqwwb"] Dec 11 02:15:32 crc kubenswrapper[4824]: E1211 02:15:32.157330 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f5cf3cd-83ba-4408-b280-6bbcdd085ab1" containerName="collect-profiles" Dec 11 02:15:32 crc kubenswrapper[4824]: I1211 02:15:32.157560 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f5cf3cd-83ba-4408-b280-6bbcdd085ab1" containerName="collect-profiles" Dec 11 02:15:32 crc kubenswrapper[4824]: E1211 02:15:32.157690 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31523d77-e1de-4227-a1f3-be75363b3321" containerName="extract-content" Dec 11 02:15:32 crc kubenswrapper[4824]: I1211 02:15:32.157841 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="31523d77-e1de-4227-a1f3-be75363b3321" containerName="extract-content" Dec 11 02:15:32 crc kubenswrapper[4824]: E1211 02:15:32.157975 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31523d77-e1de-4227-a1f3-be75363b3321" containerName="registry-server" Dec 11 02:15:32 crc kubenswrapper[4824]: I1211 02:15:32.158088 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="31523d77-e1de-4227-a1f3-be75363b3321" containerName="registry-server" Dec 11 02:15:32 crc kubenswrapper[4824]: E1211 02:15:32.158347 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31523d77-e1de-4227-a1f3-be75363b3321" containerName="extract-utilities" Dec 11 02:15:32 crc kubenswrapper[4824]: I1211 02:15:32.158509 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="31523d77-e1de-4227-a1f3-be75363b3321" containerName="extract-utilities" Dec 11 02:15:32 crc kubenswrapper[4824]: I1211 02:15:32.159428 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="31523d77-e1de-4227-a1f3-be75363b3321" containerName="registry-server" Dec 11 02:15:32 crc kubenswrapper[4824]: I1211 02:15:32.161196 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f5cf3cd-83ba-4408-b280-6bbcdd085ab1" containerName="collect-profiles" Dec 11 02:15:32 crc kubenswrapper[4824]: I1211 02:15:32.162940 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nqwwb" Dec 11 02:15:32 crc kubenswrapper[4824]: I1211 02:15:32.168392 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nqwwb"] Dec 11 02:15:32 crc kubenswrapper[4824]: I1211 02:15:32.336230 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pd6pv\" (UniqueName: \"kubernetes.io/projected/a61fd179-f3bd-4a79-ae61-d338e7994602-kube-api-access-pd6pv\") pod \"certified-operators-nqwwb\" (UID: \"a61fd179-f3bd-4a79-ae61-d338e7994602\") " pod="openshift-marketplace/certified-operators-nqwwb" Dec 11 02:15:32 crc kubenswrapper[4824]: I1211 02:15:32.336627 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a61fd179-f3bd-4a79-ae61-d338e7994602-catalog-content\") pod \"certified-operators-nqwwb\" (UID: \"a61fd179-f3bd-4a79-ae61-d338e7994602\") " pod="openshift-marketplace/certified-operators-nqwwb" Dec 11 02:15:32 crc kubenswrapper[4824]: I1211 02:15:32.336697 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a61fd179-f3bd-4a79-ae61-d338e7994602-utilities\") pod \"certified-operators-nqwwb\" (UID: \"a61fd179-f3bd-4a79-ae61-d338e7994602\") " pod="openshift-marketplace/certified-operators-nqwwb" Dec 11 02:15:32 crc kubenswrapper[4824]: I1211 02:15:32.438199 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a61fd179-f3bd-4a79-ae61-d338e7994602-utilities\") pod \"certified-operators-nqwwb\" (UID: \"a61fd179-f3bd-4a79-ae61-d338e7994602\") " pod="openshift-marketplace/certified-operators-nqwwb" Dec 11 02:15:32 crc kubenswrapper[4824]: I1211 02:15:32.438311 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pd6pv\" (UniqueName: \"kubernetes.io/projected/a61fd179-f3bd-4a79-ae61-d338e7994602-kube-api-access-pd6pv\") pod \"certified-operators-nqwwb\" (UID: \"a61fd179-f3bd-4a79-ae61-d338e7994602\") " pod="openshift-marketplace/certified-operators-nqwwb" Dec 11 02:15:32 crc kubenswrapper[4824]: I1211 02:15:32.438343 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a61fd179-f3bd-4a79-ae61-d338e7994602-catalog-content\") pod \"certified-operators-nqwwb\" (UID: \"a61fd179-f3bd-4a79-ae61-d338e7994602\") " pod="openshift-marketplace/certified-operators-nqwwb" Dec 11 02:15:32 crc kubenswrapper[4824]: I1211 02:15:32.438927 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a61fd179-f3bd-4a79-ae61-d338e7994602-catalog-content\") pod \"certified-operators-nqwwb\" (UID: \"a61fd179-f3bd-4a79-ae61-d338e7994602\") " pod="openshift-marketplace/certified-operators-nqwwb" Dec 11 02:15:32 crc kubenswrapper[4824]: I1211 02:15:32.438986 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a61fd179-f3bd-4a79-ae61-d338e7994602-utilities\") pod \"certified-operators-nqwwb\" (UID: \"a61fd179-f3bd-4a79-ae61-d338e7994602\") " pod="openshift-marketplace/certified-operators-nqwwb" Dec 11 02:15:32 crc kubenswrapper[4824]: I1211 02:15:32.459088 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pd6pv\" (UniqueName: \"kubernetes.io/projected/a61fd179-f3bd-4a79-ae61-d338e7994602-kube-api-access-pd6pv\") pod \"certified-operators-nqwwb\" (UID: \"a61fd179-f3bd-4a79-ae61-d338e7994602\") " pod="openshift-marketplace/certified-operators-nqwwb" Dec 11 02:15:32 crc kubenswrapper[4824]: I1211 02:15:32.506941 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nqwwb" Dec 11 02:15:32 crc kubenswrapper[4824]: I1211 02:15:32.786827 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nqwwb"] Dec 11 02:15:33 crc kubenswrapper[4824]: I1211 02:15:33.161061 4824 generic.go:334] "Generic (PLEG): container finished" podID="a61fd179-f3bd-4a79-ae61-d338e7994602" containerID="d69641f2b4eb11fe146e12dd4e7f012fcc4f3dd764037792573bc688974bcc91" exitCode=0 Dec 11 02:15:33 crc kubenswrapper[4824]: I1211 02:15:33.161586 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nqwwb" event={"ID":"a61fd179-f3bd-4a79-ae61-d338e7994602","Type":"ContainerDied","Data":"d69641f2b4eb11fe146e12dd4e7f012fcc4f3dd764037792573bc688974bcc91"} Dec 11 02:15:33 crc kubenswrapper[4824]: I1211 02:15:33.161612 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nqwwb" event={"ID":"a61fd179-f3bd-4a79-ae61-d338e7994602","Type":"ContainerStarted","Data":"bd8c4f84244804a8be2dd3d5434ae6e36f4360801b2871e7d1f023148c639236"} Dec 11 02:15:36 crc kubenswrapper[4824]: I1211 02:15:36.192027 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83x97sf"] Dec 11 02:15:36 crc kubenswrapper[4824]: I1211 02:15:36.193487 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83x97sf" Dec 11 02:15:36 crc kubenswrapper[4824]: I1211 02:15:36.195165 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 11 02:15:36 crc kubenswrapper[4824]: I1211 02:15:36.203332 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83x97sf"] Dec 11 02:15:36 crc kubenswrapper[4824]: I1211 02:15:36.292044 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/55c9ce15-2335-433b-a7ac-cfdf1a3ef599-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83x97sf\" (UID: \"55c9ce15-2335-433b-a7ac-cfdf1a3ef599\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83x97sf" Dec 11 02:15:36 crc kubenswrapper[4824]: I1211 02:15:36.292325 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/55c9ce15-2335-433b-a7ac-cfdf1a3ef599-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83x97sf\" (UID: \"55c9ce15-2335-433b-a7ac-cfdf1a3ef599\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83x97sf" Dec 11 02:15:36 crc kubenswrapper[4824]: I1211 02:15:36.292358 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m9hg8\" (UniqueName: \"kubernetes.io/projected/55c9ce15-2335-433b-a7ac-cfdf1a3ef599-kube-api-access-m9hg8\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83x97sf\" (UID: \"55c9ce15-2335-433b-a7ac-cfdf1a3ef599\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83x97sf" Dec 11 02:15:36 crc kubenswrapper[4824]: I1211 02:15:36.392880 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/55c9ce15-2335-433b-a7ac-cfdf1a3ef599-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83x97sf\" (UID: \"55c9ce15-2335-433b-a7ac-cfdf1a3ef599\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83x97sf" Dec 11 02:15:36 crc kubenswrapper[4824]: I1211 02:15:36.392923 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/55c9ce15-2335-433b-a7ac-cfdf1a3ef599-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83x97sf\" (UID: \"55c9ce15-2335-433b-a7ac-cfdf1a3ef599\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83x97sf" Dec 11 02:15:36 crc kubenswrapper[4824]: I1211 02:15:36.392949 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m9hg8\" (UniqueName: \"kubernetes.io/projected/55c9ce15-2335-433b-a7ac-cfdf1a3ef599-kube-api-access-m9hg8\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83x97sf\" (UID: \"55c9ce15-2335-433b-a7ac-cfdf1a3ef599\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83x97sf" Dec 11 02:15:36 crc kubenswrapper[4824]: I1211 02:15:36.393733 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/55c9ce15-2335-433b-a7ac-cfdf1a3ef599-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83x97sf\" (UID: \"55c9ce15-2335-433b-a7ac-cfdf1a3ef599\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83x97sf" Dec 11 02:15:36 crc kubenswrapper[4824]: I1211 02:15:36.393741 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/55c9ce15-2335-433b-a7ac-cfdf1a3ef599-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83x97sf\" (UID: \"55c9ce15-2335-433b-a7ac-cfdf1a3ef599\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83x97sf" Dec 11 02:15:36 crc kubenswrapper[4824]: I1211 02:15:36.429360 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m9hg8\" (UniqueName: \"kubernetes.io/projected/55c9ce15-2335-433b-a7ac-cfdf1a3ef599-kube-api-access-m9hg8\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83x97sf\" (UID: \"55c9ce15-2335-433b-a7ac-cfdf1a3ef599\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83x97sf" Dec 11 02:15:36 crc kubenswrapper[4824]: I1211 02:15:36.510878 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83x97sf" Dec 11 02:15:37 crc kubenswrapper[4824]: I1211 02:15:37.197415 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-k6kls" podUID="ad510a18-cd72-4cf0-9b0c-dd3540d491fa" containerName="console" containerID="cri-o://c7eb4f774e5a9e63b3626818ebe56f13753bed99b12c3e95972fb1b6876fc303" gracePeriod=15 Dec 11 02:15:37 crc kubenswrapper[4824]: I1211 02:15:37.624600 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83x97sf"] Dec 11 02:15:37 crc kubenswrapper[4824]: W1211 02:15:37.629058 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod55c9ce15_2335_433b_a7ac_cfdf1a3ef599.slice/crio-633cbc56d13de47be960fef94bd670d57cb3203d3590af76f315bd2295c049f7 WatchSource:0}: Error finding container 633cbc56d13de47be960fef94bd670d57cb3203d3590af76f315bd2295c049f7: Status 404 returned error can't find the container with id 633cbc56d13de47be960fef94bd670d57cb3203d3590af76f315bd2295c049f7 Dec 11 02:15:37 crc kubenswrapper[4824]: I1211 02:15:37.775942 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-k6kls_ad510a18-cd72-4cf0-9b0c-dd3540d491fa/console/0.log" Dec 11 02:15:37 crc kubenswrapper[4824]: I1211 02:15:37.776014 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-k6kls" Dec 11 02:15:37 crc kubenswrapper[4824]: I1211 02:15:37.924885 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ad510a18-cd72-4cf0-9b0c-dd3540d491fa-service-ca\") pod \"ad510a18-cd72-4cf0-9b0c-dd3540d491fa\" (UID: \"ad510a18-cd72-4cf0-9b0c-dd3540d491fa\") " Dec 11 02:15:37 crc kubenswrapper[4824]: I1211 02:15:37.924973 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ad510a18-cd72-4cf0-9b0c-dd3540d491fa-console-oauth-config\") pod \"ad510a18-cd72-4cf0-9b0c-dd3540d491fa\" (UID: \"ad510a18-cd72-4cf0-9b0c-dd3540d491fa\") " Dec 11 02:15:37 crc kubenswrapper[4824]: I1211 02:15:37.925011 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ad510a18-cd72-4cf0-9b0c-dd3540d491fa-trusted-ca-bundle\") pod \"ad510a18-cd72-4cf0-9b0c-dd3540d491fa\" (UID: \"ad510a18-cd72-4cf0-9b0c-dd3540d491fa\") " Dec 11 02:15:37 crc kubenswrapper[4824]: I1211 02:15:37.925094 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ad510a18-cd72-4cf0-9b0c-dd3540d491fa-console-serving-cert\") pod \"ad510a18-cd72-4cf0-9b0c-dd3540d491fa\" (UID: \"ad510a18-cd72-4cf0-9b0c-dd3540d491fa\") " Dec 11 02:15:37 crc kubenswrapper[4824]: I1211 02:15:37.925197 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rm2hc\" (UniqueName: \"kubernetes.io/projected/ad510a18-cd72-4cf0-9b0c-dd3540d491fa-kube-api-access-rm2hc\") pod \"ad510a18-cd72-4cf0-9b0c-dd3540d491fa\" (UID: \"ad510a18-cd72-4cf0-9b0c-dd3540d491fa\") " Dec 11 02:15:37 crc kubenswrapper[4824]: I1211 02:15:37.925241 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ad510a18-cd72-4cf0-9b0c-dd3540d491fa-oauth-serving-cert\") pod \"ad510a18-cd72-4cf0-9b0c-dd3540d491fa\" (UID: \"ad510a18-cd72-4cf0-9b0c-dd3540d491fa\") " Dec 11 02:15:37 crc kubenswrapper[4824]: I1211 02:15:37.925305 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ad510a18-cd72-4cf0-9b0c-dd3540d491fa-console-config\") pod \"ad510a18-cd72-4cf0-9b0c-dd3540d491fa\" (UID: \"ad510a18-cd72-4cf0-9b0c-dd3540d491fa\") " Dec 11 02:15:37 crc kubenswrapper[4824]: I1211 02:15:37.927679 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ad510a18-cd72-4cf0-9b0c-dd3540d491fa-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "ad510a18-cd72-4cf0-9b0c-dd3540d491fa" (UID: "ad510a18-cd72-4cf0-9b0c-dd3540d491fa"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:15:37 crc kubenswrapper[4824]: I1211 02:15:37.927786 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ad510a18-cd72-4cf0-9b0c-dd3540d491fa-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "ad510a18-cd72-4cf0-9b0c-dd3540d491fa" (UID: "ad510a18-cd72-4cf0-9b0c-dd3540d491fa"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:15:37 crc kubenswrapper[4824]: I1211 02:15:37.927900 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ad510a18-cd72-4cf0-9b0c-dd3540d491fa-service-ca" (OuterVolumeSpecName: "service-ca") pod "ad510a18-cd72-4cf0-9b0c-dd3540d491fa" (UID: "ad510a18-cd72-4cf0-9b0c-dd3540d491fa"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:15:37 crc kubenswrapper[4824]: I1211 02:15:37.928177 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ad510a18-cd72-4cf0-9b0c-dd3540d491fa-console-config" (OuterVolumeSpecName: "console-config") pod "ad510a18-cd72-4cf0-9b0c-dd3540d491fa" (UID: "ad510a18-cd72-4cf0-9b0c-dd3540d491fa"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:15:37 crc kubenswrapper[4824]: I1211 02:15:37.933015 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad510a18-cd72-4cf0-9b0c-dd3540d491fa-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "ad510a18-cd72-4cf0-9b0c-dd3540d491fa" (UID: "ad510a18-cd72-4cf0-9b0c-dd3540d491fa"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:15:37 crc kubenswrapper[4824]: I1211 02:15:37.933734 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad510a18-cd72-4cf0-9b0c-dd3540d491fa-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "ad510a18-cd72-4cf0-9b0c-dd3540d491fa" (UID: "ad510a18-cd72-4cf0-9b0c-dd3540d491fa"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:15:37 crc kubenswrapper[4824]: I1211 02:15:37.933787 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad510a18-cd72-4cf0-9b0c-dd3540d491fa-kube-api-access-rm2hc" (OuterVolumeSpecName: "kube-api-access-rm2hc") pod "ad510a18-cd72-4cf0-9b0c-dd3540d491fa" (UID: "ad510a18-cd72-4cf0-9b0c-dd3540d491fa"). InnerVolumeSpecName "kube-api-access-rm2hc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:15:38 crc kubenswrapper[4824]: I1211 02:15:38.026747 4824 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ad510a18-cd72-4cf0-9b0c-dd3540d491fa-console-config\") on node \"crc\" DevicePath \"\"" Dec 11 02:15:38 crc kubenswrapper[4824]: I1211 02:15:38.026775 4824 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ad510a18-cd72-4cf0-9b0c-dd3540d491fa-service-ca\") on node \"crc\" DevicePath \"\"" Dec 11 02:15:38 crc kubenswrapper[4824]: I1211 02:15:38.026785 4824 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ad510a18-cd72-4cf0-9b0c-dd3540d491fa-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 11 02:15:38 crc kubenswrapper[4824]: I1211 02:15:38.026793 4824 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ad510a18-cd72-4cf0-9b0c-dd3540d491fa-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 02:15:38 crc kubenswrapper[4824]: I1211 02:15:38.026803 4824 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ad510a18-cd72-4cf0-9b0c-dd3540d491fa-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 02:15:38 crc kubenswrapper[4824]: I1211 02:15:38.026812 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rm2hc\" (UniqueName: \"kubernetes.io/projected/ad510a18-cd72-4cf0-9b0c-dd3540d491fa-kube-api-access-rm2hc\") on node \"crc\" DevicePath \"\"" Dec 11 02:15:38 crc kubenswrapper[4824]: I1211 02:15:38.026822 4824 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ad510a18-cd72-4cf0-9b0c-dd3540d491fa-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 02:15:38 crc kubenswrapper[4824]: I1211 02:15:38.205623 4824 generic.go:334] "Generic (PLEG): container finished" podID="a61fd179-f3bd-4a79-ae61-d338e7994602" containerID="b4463af1689d04f8af4bcc2b1088d14ef3d14ad67f0962f530f1f38d45663b4a" exitCode=0 Dec 11 02:15:38 crc kubenswrapper[4824]: I1211 02:15:38.205765 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nqwwb" event={"ID":"a61fd179-f3bd-4a79-ae61-d338e7994602","Type":"ContainerDied","Data":"b4463af1689d04f8af4bcc2b1088d14ef3d14ad67f0962f530f1f38d45663b4a"} Dec 11 02:15:38 crc kubenswrapper[4824]: I1211 02:15:38.208459 4824 generic.go:334] "Generic (PLEG): container finished" podID="55c9ce15-2335-433b-a7ac-cfdf1a3ef599" containerID="3bd45a506f38add9f5c0c97dfa4b73f5798049dd7e5e860570191d2f1df6f055" exitCode=0 Dec 11 02:15:38 crc kubenswrapper[4824]: I1211 02:15:38.208678 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83x97sf" event={"ID":"55c9ce15-2335-433b-a7ac-cfdf1a3ef599","Type":"ContainerDied","Data":"3bd45a506f38add9f5c0c97dfa4b73f5798049dd7e5e860570191d2f1df6f055"} Dec 11 02:15:38 crc kubenswrapper[4824]: I1211 02:15:38.208759 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83x97sf" event={"ID":"55c9ce15-2335-433b-a7ac-cfdf1a3ef599","Type":"ContainerStarted","Data":"633cbc56d13de47be960fef94bd670d57cb3203d3590af76f315bd2295c049f7"} Dec 11 02:15:38 crc kubenswrapper[4824]: I1211 02:15:38.215744 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-k6kls_ad510a18-cd72-4cf0-9b0c-dd3540d491fa/console/0.log" Dec 11 02:15:38 crc kubenswrapper[4824]: I1211 02:15:38.215808 4824 generic.go:334] "Generic (PLEG): container finished" podID="ad510a18-cd72-4cf0-9b0c-dd3540d491fa" containerID="c7eb4f774e5a9e63b3626818ebe56f13753bed99b12c3e95972fb1b6876fc303" exitCode=2 Dec 11 02:15:38 crc kubenswrapper[4824]: I1211 02:15:38.215858 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-k6kls" event={"ID":"ad510a18-cd72-4cf0-9b0c-dd3540d491fa","Type":"ContainerDied","Data":"c7eb4f774e5a9e63b3626818ebe56f13753bed99b12c3e95972fb1b6876fc303"} Dec 11 02:15:38 crc kubenswrapper[4824]: I1211 02:15:38.215897 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-k6kls" event={"ID":"ad510a18-cd72-4cf0-9b0c-dd3540d491fa","Type":"ContainerDied","Data":"9f6757a7ca4186844eb08113017138bcde22f84e6d25f11852180354ee173046"} Dec 11 02:15:38 crc kubenswrapper[4824]: I1211 02:15:38.215930 4824 scope.go:117] "RemoveContainer" containerID="c7eb4f774e5a9e63b3626818ebe56f13753bed99b12c3e95972fb1b6876fc303" Dec 11 02:15:38 crc kubenswrapper[4824]: I1211 02:15:38.216102 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-k6kls" Dec 11 02:15:38 crc kubenswrapper[4824]: I1211 02:15:38.250739 4824 scope.go:117] "RemoveContainer" containerID="c7eb4f774e5a9e63b3626818ebe56f13753bed99b12c3e95972fb1b6876fc303" Dec 11 02:15:38 crc kubenswrapper[4824]: E1211 02:15:38.251393 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c7eb4f774e5a9e63b3626818ebe56f13753bed99b12c3e95972fb1b6876fc303\": container with ID starting with c7eb4f774e5a9e63b3626818ebe56f13753bed99b12c3e95972fb1b6876fc303 not found: ID does not exist" containerID="c7eb4f774e5a9e63b3626818ebe56f13753bed99b12c3e95972fb1b6876fc303" Dec 11 02:15:38 crc kubenswrapper[4824]: I1211 02:15:38.251462 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c7eb4f774e5a9e63b3626818ebe56f13753bed99b12c3e95972fb1b6876fc303"} err="failed to get container status \"c7eb4f774e5a9e63b3626818ebe56f13753bed99b12c3e95972fb1b6876fc303\": rpc error: code = NotFound desc = could not find container \"c7eb4f774e5a9e63b3626818ebe56f13753bed99b12c3e95972fb1b6876fc303\": container with ID starting with c7eb4f774e5a9e63b3626818ebe56f13753bed99b12c3e95972fb1b6876fc303 not found: ID does not exist" Dec 11 02:15:38 crc kubenswrapper[4824]: I1211 02:15:38.282375 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-k6kls"] Dec 11 02:15:38 crc kubenswrapper[4824]: I1211 02:15:38.290622 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-k6kls"] Dec 11 02:15:38 crc kubenswrapper[4824]: I1211 02:15:38.645216 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad510a18-cd72-4cf0-9b0c-dd3540d491fa" path="/var/lib/kubelet/pods/ad510a18-cd72-4cf0-9b0c-dd3540d491fa/volumes" Dec 11 02:15:39 crc kubenswrapper[4824]: I1211 02:15:39.225625 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nqwwb" event={"ID":"a61fd179-f3bd-4a79-ae61-d338e7994602","Type":"ContainerStarted","Data":"b2d3e2f0bb6ef2a1ac1c44843b1b5085cb811108b1f8a191451cc5ef741e40f7"} Dec 11 02:15:39 crc kubenswrapper[4824]: I1211 02:15:39.264795 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-nqwwb" podStartSLOduration=1.783073451 podStartE2EDuration="7.264769433s" podCreationTimestamp="2025-12-11 02:15:32 +0000 UTC" firstStartedPulling="2025-12-11 02:15:33.16308149 +0000 UTC m=+874.852118869" lastFinishedPulling="2025-12-11 02:15:38.644777442 +0000 UTC m=+880.333814851" observedRunningTime="2025-12-11 02:15:39.257429001 +0000 UTC m=+880.946466420" watchObservedRunningTime="2025-12-11 02:15:39.264769433 +0000 UTC m=+880.953806852" Dec 11 02:15:42 crc kubenswrapper[4824]: I1211 02:15:42.508201 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-nqwwb" Dec 11 02:15:42 crc kubenswrapper[4824]: I1211 02:15:42.508726 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-nqwwb" Dec 11 02:15:42 crc kubenswrapper[4824]: I1211 02:15:42.579000 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-nqwwb" Dec 11 02:15:43 crc kubenswrapper[4824]: I1211 02:15:43.322869 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-nqwwb" Dec 11 02:15:44 crc kubenswrapper[4824]: I1211 02:15:44.558856 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nqwwb"] Dec 11 02:15:44 crc kubenswrapper[4824]: I1211 02:15:44.917651 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7dwps"] Dec 11 02:15:44 crc kubenswrapper[4824]: I1211 02:15:44.918343 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-7dwps" podUID="beb14a1d-6a61-4787-8e2b-ee44942e2828" containerName="registry-server" containerID="cri-o://ca34783218eccde1b6d0a57e038f1523378c0e97d1e5c3b27664aeed3696dac2" gracePeriod=2 Dec 11 02:15:46 crc kubenswrapper[4824]: I1211 02:15:46.283828 4824 generic.go:334] "Generic (PLEG): container finished" podID="beb14a1d-6a61-4787-8e2b-ee44942e2828" containerID="ca34783218eccde1b6d0a57e038f1523378c0e97d1e5c3b27664aeed3696dac2" exitCode=0 Dec 11 02:15:46 crc kubenswrapper[4824]: I1211 02:15:46.283908 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7dwps" event={"ID":"beb14a1d-6a61-4787-8e2b-ee44942e2828","Type":"ContainerDied","Data":"ca34783218eccde1b6d0a57e038f1523378c0e97d1e5c3b27664aeed3696dac2"} Dec 11 02:15:47 crc kubenswrapper[4824]: I1211 02:15:47.994431 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7dwps" Dec 11 02:15:48 crc kubenswrapper[4824]: I1211 02:15:48.178363 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/beb14a1d-6a61-4787-8e2b-ee44942e2828-catalog-content\") pod \"beb14a1d-6a61-4787-8e2b-ee44942e2828\" (UID: \"beb14a1d-6a61-4787-8e2b-ee44942e2828\") " Dec 11 02:15:48 crc kubenswrapper[4824]: I1211 02:15:48.178511 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/beb14a1d-6a61-4787-8e2b-ee44942e2828-utilities\") pod \"beb14a1d-6a61-4787-8e2b-ee44942e2828\" (UID: \"beb14a1d-6a61-4787-8e2b-ee44942e2828\") " Dec 11 02:15:48 crc kubenswrapper[4824]: I1211 02:15:48.178649 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jnfch\" (UniqueName: \"kubernetes.io/projected/beb14a1d-6a61-4787-8e2b-ee44942e2828-kube-api-access-jnfch\") pod \"beb14a1d-6a61-4787-8e2b-ee44942e2828\" (UID: \"beb14a1d-6a61-4787-8e2b-ee44942e2828\") " Dec 11 02:15:48 crc kubenswrapper[4824]: I1211 02:15:48.179895 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/beb14a1d-6a61-4787-8e2b-ee44942e2828-utilities" (OuterVolumeSpecName: "utilities") pod "beb14a1d-6a61-4787-8e2b-ee44942e2828" (UID: "beb14a1d-6a61-4787-8e2b-ee44942e2828"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:15:48 crc kubenswrapper[4824]: I1211 02:15:48.188034 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/beb14a1d-6a61-4787-8e2b-ee44942e2828-kube-api-access-jnfch" (OuterVolumeSpecName: "kube-api-access-jnfch") pod "beb14a1d-6a61-4787-8e2b-ee44942e2828" (UID: "beb14a1d-6a61-4787-8e2b-ee44942e2828"). InnerVolumeSpecName "kube-api-access-jnfch". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:15:48 crc kubenswrapper[4824]: I1211 02:15:48.275456 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/beb14a1d-6a61-4787-8e2b-ee44942e2828-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "beb14a1d-6a61-4787-8e2b-ee44942e2828" (UID: "beb14a1d-6a61-4787-8e2b-ee44942e2828"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:15:48 crc kubenswrapper[4824]: I1211 02:15:48.280212 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/beb14a1d-6a61-4787-8e2b-ee44942e2828-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 02:15:48 crc kubenswrapper[4824]: I1211 02:15:48.280341 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/beb14a1d-6a61-4787-8e2b-ee44942e2828-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 02:15:48 crc kubenswrapper[4824]: I1211 02:15:48.280441 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jnfch\" (UniqueName: \"kubernetes.io/projected/beb14a1d-6a61-4787-8e2b-ee44942e2828-kube-api-access-jnfch\") on node \"crc\" DevicePath \"\"" Dec 11 02:15:48 crc kubenswrapper[4824]: I1211 02:15:48.303848 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7dwps" event={"ID":"beb14a1d-6a61-4787-8e2b-ee44942e2828","Type":"ContainerDied","Data":"2fad8d7713b6adeea66bfeb0cb53bd982b39af5afb63a89207bec2af30f6a337"} Dec 11 02:15:48 crc kubenswrapper[4824]: I1211 02:15:48.303905 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7dwps" Dec 11 02:15:48 crc kubenswrapper[4824]: I1211 02:15:48.304094 4824 scope.go:117] "RemoveContainer" containerID="ca34783218eccde1b6d0a57e038f1523378c0e97d1e5c3b27664aeed3696dac2" Dec 11 02:15:48 crc kubenswrapper[4824]: I1211 02:15:48.339946 4824 scope.go:117] "RemoveContainer" containerID="08ed96f115420fde53df358887000e830716bd70e42ff1a54148fe8dbf8439b0" Dec 11 02:15:48 crc kubenswrapper[4824]: I1211 02:15:48.360674 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7dwps"] Dec 11 02:15:48 crc kubenswrapper[4824]: I1211 02:15:48.369946 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-7dwps"] Dec 11 02:15:48 crc kubenswrapper[4824]: I1211 02:15:48.398202 4824 scope.go:117] "RemoveContainer" containerID="4382dff63b8a8c27a32d6fa7e0519472b18363149789d5febd67109c44d6c190" Dec 11 02:15:48 crc kubenswrapper[4824]: I1211 02:15:48.640912 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="beb14a1d-6a61-4787-8e2b-ee44942e2828" path="/var/lib/kubelet/pods/beb14a1d-6a61-4787-8e2b-ee44942e2828/volumes" Dec 11 02:15:51 crc kubenswrapper[4824]: I1211 02:15:51.330382 4824 generic.go:334] "Generic (PLEG): container finished" podID="55c9ce15-2335-433b-a7ac-cfdf1a3ef599" containerID="1b376326778b12966bd2a2d774da28ad35823f1958bfcd37f7b1f9c8f1dd4472" exitCode=0 Dec 11 02:15:51 crc kubenswrapper[4824]: I1211 02:15:51.330541 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83x97sf" event={"ID":"55c9ce15-2335-433b-a7ac-cfdf1a3ef599","Type":"ContainerDied","Data":"1b376326778b12966bd2a2d774da28ad35823f1958bfcd37f7b1f9c8f1dd4472"} Dec 11 02:15:52 crc kubenswrapper[4824]: I1211 02:15:52.338791 4824 generic.go:334] "Generic (PLEG): container finished" podID="55c9ce15-2335-433b-a7ac-cfdf1a3ef599" containerID="1709100b7a9ae0c97b73289c87374ede9316acd29e4b2329ee35edf5bd9dda0d" exitCode=0 Dec 11 02:15:52 crc kubenswrapper[4824]: I1211 02:15:52.338866 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83x97sf" event={"ID":"55c9ce15-2335-433b-a7ac-cfdf1a3ef599","Type":"ContainerDied","Data":"1709100b7a9ae0c97b73289c87374ede9316acd29e4b2329ee35edf5bd9dda0d"} Dec 11 02:15:53 crc kubenswrapper[4824]: I1211 02:15:53.637324 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83x97sf" Dec 11 02:15:53 crc kubenswrapper[4824]: I1211 02:15:53.762245 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/55c9ce15-2335-433b-a7ac-cfdf1a3ef599-bundle\") pod \"55c9ce15-2335-433b-a7ac-cfdf1a3ef599\" (UID: \"55c9ce15-2335-433b-a7ac-cfdf1a3ef599\") " Dec 11 02:15:53 crc kubenswrapper[4824]: I1211 02:15:53.762331 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m9hg8\" (UniqueName: \"kubernetes.io/projected/55c9ce15-2335-433b-a7ac-cfdf1a3ef599-kube-api-access-m9hg8\") pod \"55c9ce15-2335-433b-a7ac-cfdf1a3ef599\" (UID: \"55c9ce15-2335-433b-a7ac-cfdf1a3ef599\") " Dec 11 02:15:53 crc kubenswrapper[4824]: I1211 02:15:53.762366 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/55c9ce15-2335-433b-a7ac-cfdf1a3ef599-util\") pod \"55c9ce15-2335-433b-a7ac-cfdf1a3ef599\" (UID: \"55c9ce15-2335-433b-a7ac-cfdf1a3ef599\") " Dec 11 02:15:53 crc kubenswrapper[4824]: I1211 02:15:53.765477 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/55c9ce15-2335-433b-a7ac-cfdf1a3ef599-bundle" (OuterVolumeSpecName: "bundle") pod "55c9ce15-2335-433b-a7ac-cfdf1a3ef599" (UID: "55c9ce15-2335-433b-a7ac-cfdf1a3ef599"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:15:53 crc kubenswrapper[4824]: I1211 02:15:53.772176 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55c9ce15-2335-433b-a7ac-cfdf1a3ef599-kube-api-access-m9hg8" (OuterVolumeSpecName: "kube-api-access-m9hg8") pod "55c9ce15-2335-433b-a7ac-cfdf1a3ef599" (UID: "55c9ce15-2335-433b-a7ac-cfdf1a3ef599"). InnerVolumeSpecName "kube-api-access-m9hg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:15:53 crc kubenswrapper[4824]: I1211 02:15:53.786305 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/55c9ce15-2335-433b-a7ac-cfdf1a3ef599-util" (OuterVolumeSpecName: "util") pod "55c9ce15-2335-433b-a7ac-cfdf1a3ef599" (UID: "55c9ce15-2335-433b-a7ac-cfdf1a3ef599"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:15:53 crc kubenswrapper[4824]: I1211 02:15:53.863901 4824 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/55c9ce15-2335-433b-a7ac-cfdf1a3ef599-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 02:15:53 crc kubenswrapper[4824]: I1211 02:15:53.863948 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m9hg8\" (UniqueName: \"kubernetes.io/projected/55c9ce15-2335-433b-a7ac-cfdf1a3ef599-kube-api-access-m9hg8\") on node \"crc\" DevicePath \"\"" Dec 11 02:15:53 crc kubenswrapper[4824]: I1211 02:15:53.863968 4824 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/55c9ce15-2335-433b-a7ac-cfdf1a3ef599-util\") on node \"crc\" DevicePath \"\"" Dec 11 02:15:54 crc kubenswrapper[4824]: I1211 02:15:54.356070 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83x97sf" event={"ID":"55c9ce15-2335-433b-a7ac-cfdf1a3ef599","Type":"ContainerDied","Data":"633cbc56d13de47be960fef94bd670d57cb3203d3590af76f315bd2295c049f7"} Dec 11 02:15:54 crc kubenswrapper[4824]: I1211 02:15:54.356168 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="633cbc56d13de47be960fef94bd670d57cb3203d3590af76f315bd2295c049f7" Dec 11 02:15:54 crc kubenswrapper[4824]: I1211 02:15:54.356218 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83x97sf" Dec 11 02:16:04 crc kubenswrapper[4824]: I1211 02:16:04.477310 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-7d6477f866-c4dqb"] Dec 11 02:16:04 crc kubenswrapper[4824]: E1211 02:16:04.478965 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="beb14a1d-6a61-4787-8e2b-ee44942e2828" containerName="extract-utilities" Dec 11 02:16:04 crc kubenswrapper[4824]: I1211 02:16:04.479056 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="beb14a1d-6a61-4787-8e2b-ee44942e2828" containerName="extract-utilities" Dec 11 02:16:04 crc kubenswrapper[4824]: E1211 02:16:04.479161 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55c9ce15-2335-433b-a7ac-cfdf1a3ef599" containerName="util" Dec 11 02:16:04 crc kubenswrapper[4824]: I1211 02:16:04.479241 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="55c9ce15-2335-433b-a7ac-cfdf1a3ef599" containerName="util" Dec 11 02:16:04 crc kubenswrapper[4824]: E1211 02:16:04.479322 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="beb14a1d-6a61-4787-8e2b-ee44942e2828" containerName="extract-content" Dec 11 02:16:04 crc kubenswrapper[4824]: I1211 02:16:04.479399 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="beb14a1d-6a61-4787-8e2b-ee44942e2828" containerName="extract-content" Dec 11 02:16:04 crc kubenswrapper[4824]: E1211 02:16:04.479480 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="beb14a1d-6a61-4787-8e2b-ee44942e2828" containerName="registry-server" Dec 11 02:16:04 crc kubenswrapper[4824]: I1211 02:16:04.479553 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="beb14a1d-6a61-4787-8e2b-ee44942e2828" containerName="registry-server" Dec 11 02:16:04 crc kubenswrapper[4824]: E1211 02:16:04.479629 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad510a18-cd72-4cf0-9b0c-dd3540d491fa" containerName="console" Dec 11 02:16:04 crc kubenswrapper[4824]: I1211 02:16:04.479700 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad510a18-cd72-4cf0-9b0c-dd3540d491fa" containerName="console" Dec 11 02:16:04 crc kubenswrapper[4824]: E1211 02:16:04.479774 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55c9ce15-2335-433b-a7ac-cfdf1a3ef599" containerName="extract" Dec 11 02:16:04 crc kubenswrapper[4824]: I1211 02:16:04.479849 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="55c9ce15-2335-433b-a7ac-cfdf1a3ef599" containerName="extract" Dec 11 02:16:04 crc kubenswrapper[4824]: E1211 02:16:04.479932 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55c9ce15-2335-433b-a7ac-cfdf1a3ef599" containerName="pull" Dec 11 02:16:04 crc kubenswrapper[4824]: I1211 02:16:04.480007 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="55c9ce15-2335-433b-a7ac-cfdf1a3ef599" containerName="pull" Dec 11 02:16:04 crc kubenswrapper[4824]: I1211 02:16:04.480211 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="55c9ce15-2335-433b-a7ac-cfdf1a3ef599" containerName="extract" Dec 11 02:16:04 crc kubenswrapper[4824]: I1211 02:16:04.480310 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad510a18-cd72-4cf0-9b0c-dd3540d491fa" containerName="console" Dec 11 02:16:04 crc kubenswrapper[4824]: I1211 02:16:04.480387 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="beb14a1d-6a61-4787-8e2b-ee44942e2828" containerName="registry-server" Dec 11 02:16:04 crc kubenswrapper[4824]: I1211 02:16:04.480996 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-7d6477f866-c4dqb" Dec 11 02:16:04 crc kubenswrapper[4824]: I1211 02:16:04.484576 4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Dec 11 02:16:04 crc kubenswrapper[4824]: I1211 02:16:04.484694 4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Dec 11 02:16:04 crc kubenswrapper[4824]: I1211 02:16:04.484661 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Dec 11 02:16:04 crc kubenswrapper[4824]: I1211 02:16:04.485072 4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-gtxwc" Dec 11 02:16:04 crc kubenswrapper[4824]: I1211 02:16:04.485082 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Dec 11 02:16:04 crc kubenswrapper[4824]: I1211 02:16:04.493080 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-7d6477f866-c4dqb"] Dec 11 02:16:04 crc kubenswrapper[4824]: I1211 02:16:04.582689 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b9n8n\" (UniqueName: \"kubernetes.io/projected/47581f0c-dffe-42e8-8211-bb40625aadf1-kube-api-access-b9n8n\") pod \"metallb-operator-controller-manager-7d6477f866-c4dqb\" (UID: \"47581f0c-dffe-42e8-8211-bb40625aadf1\") " pod="metallb-system/metallb-operator-controller-manager-7d6477f866-c4dqb" Dec 11 02:16:04 crc kubenswrapper[4824]: I1211 02:16:04.582784 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/47581f0c-dffe-42e8-8211-bb40625aadf1-webhook-cert\") pod \"metallb-operator-controller-manager-7d6477f866-c4dqb\" (UID: \"47581f0c-dffe-42e8-8211-bb40625aadf1\") " pod="metallb-system/metallb-operator-controller-manager-7d6477f866-c4dqb" Dec 11 02:16:04 crc kubenswrapper[4824]: I1211 02:16:04.582840 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/47581f0c-dffe-42e8-8211-bb40625aadf1-apiservice-cert\") pod \"metallb-operator-controller-manager-7d6477f866-c4dqb\" (UID: \"47581f0c-dffe-42e8-8211-bb40625aadf1\") " pod="metallb-system/metallb-operator-controller-manager-7d6477f866-c4dqb" Dec 11 02:16:04 crc kubenswrapper[4824]: I1211 02:16:04.683776 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/47581f0c-dffe-42e8-8211-bb40625aadf1-webhook-cert\") pod \"metallb-operator-controller-manager-7d6477f866-c4dqb\" (UID: \"47581f0c-dffe-42e8-8211-bb40625aadf1\") " pod="metallb-system/metallb-operator-controller-manager-7d6477f866-c4dqb" Dec 11 02:16:04 crc kubenswrapper[4824]: I1211 02:16:04.683835 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/47581f0c-dffe-42e8-8211-bb40625aadf1-apiservice-cert\") pod \"metallb-operator-controller-manager-7d6477f866-c4dqb\" (UID: \"47581f0c-dffe-42e8-8211-bb40625aadf1\") " pod="metallb-system/metallb-operator-controller-manager-7d6477f866-c4dqb" Dec 11 02:16:04 crc kubenswrapper[4824]: I1211 02:16:04.683873 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b9n8n\" (UniqueName: \"kubernetes.io/projected/47581f0c-dffe-42e8-8211-bb40625aadf1-kube-api-access-b9n8n\") pod \"metallb-operator-controller-manager-7d6477f866-c4dqb\" (UID: \"47581f0c-dffe-42e8-8211-bb40625aadf1\") " pod="metallb-system/metallb-operator-controller-manager-7d6477f866-c4dqb" Dec 11 02:16:04 crc kubenswrapper[4824]: I1211 02:16:04.690179 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/47581f0c-dffe-42e8-8211-bb40625aadf1-apiservice-cert\") pod \"metallb-operator-controller-manager-7d6477f866-c4dqb\" (UID: \"47581f0c-dffe-42e8-8211-bb40625aadf1\") " pod="metallb-system/metallb-operator-controller-manager-7d6477f866-c4dqb" Dec 11 02:16:04 crc kubenswrapper[4824]: I1211 02:16:04.690703 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/47581f0c-dffe-42e8-8211-bb40625aadf1-webhook-cert\") pod \"metallb-operator-controller-manager-7d6477f866-c4dqb\" (UID: \"47581f0c-dffe-42e8-8211-bb40625aadf1\") " pod="metallb-system/metallb-operator-controller-manager-7d6477f866-c4dqb" Dec 11 02:16:04 crc kubenswrapper[4824]: I1211 02:16:04.707050 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-6cb87f7d4c-jt2ns"] Dec 11 02:16:04 crc kubenswrapper[4824]: I1211 02:16:04.707885 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-6cb87f7d4c-jt2ns" Dec 11 02:16:04 crc kubenswrapper[4824]: I1211 02:16:04.709708 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b9n8n\" (UniqueName: \"kubernetes.io/projected/47581f0c-dffe-42e8-8211-bb40625aadf1-kube-api-access-b9n8n\") pod \"metallb-operator-controller-manager-7d6477f866-c4dqb\" (UID: \"47581f0c-dffe-42e8-8211-bb40625aadf1\") " pod="metallb-system/metallb-operator-controller-manager-7d6477f866-c4dqb" Dec 11 02:16:04 crc kubenswrapper[4824]: I1211 02:16:04.710033 4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 11 02:16:04 crc kubenswrapper[4824]: I1211 02:16:04.710120 4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Dec 11 02:16:04 crc kubenswrapper[4824]: I1211 02:16:04.711955 4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-zt7hk" Dec 11 02:16:04 crc kubenswrapper[4824]: I1211 02:16:04.718515 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-6cb87f7d4c-jt2ns"] Dec 11 02:16:04 crc kubenswrapper[4824]: I1211 02:16:04.829001 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-7d6477f866-c4dqb" Dec 11 02:16:04 crc kubenswrapper[4824]: I1211 02:16:04.886221 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0c1d110c-0bd5-410f-af71-7af964dfb17d-webhook-cert\") pod \"metallb-operator-webhook-server-6cb87f7d4c-jt2ns\" (UID: \"0c1d110c-0bd5-410f-af71-7af964dfb17d\") " pod="metallb-system/metallb-operator-webhook-server-6cb87f7d4c-jt2ns" Dec 11 02:16:04 crc kubenswrapper[4824]: I1211 02:16:04.886331 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j599t\" (UniqueName: \"kubernetes.io/projected/0c1d110c-0bd5-410f-af71-7af964dfb17d-kube-api-access-j599t\") pod \"metallb-operator-webhook-server-6cb87f7d4c-jt2ns\" (UID: \"0c1d110c-0bd5-410f-af71-7af964dfb17d\") " pod="metallb-system/metallb-operator-webhook-server-6cb87f7d4c-jt2ns" Dec 11 02:16:04 crc kubenswrapper[4824]: I1211 02:16:04.886360 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0c1d110c-0bd5-410f-af71-7af964dfb17d-apiservice-cert\") pod \"metallb-operator-webhook-server-6cb87f7d4c-jt2ns\" (UID: \"0c1d110c-0bd5-410f-af71-7af964dfb17d\") " pod="metallb-system/metallb-operator-webhook-server-6cb87f7d4c-jt2ns" Dec 11 02:16:04 crc kubenswrapper[4824]: I1211 02:16:04.987817 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j599t\" (UniqueName: \"kubernetes.io/projected/0c1d110c-0bd5-410f-af71-7af964dfb17d-kube-api-access-j599t\") pod \"metallb-operator-webhook-server-6cb87f7d4c-jt2ns\" (UID: \"0c1d110c-0bd5-410f-af71-7af964dfb17d\") " pod="metallb-system/metallb-operator-webhook-server-6cb87f7d4c-jt2ns" Dec 11 02:16:04 crc kubenswrapper[4824]: I1211 02:16:04.988092 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0c1d110c-0bd5-410f-af71-7af964dfb17d-apiservice-cert\") pod \"metallb-operator-webhook-server-6cb87f7d4c-jt2ns\" (UID: \"0c1d110c-0bd5-410f-af71-7af964dfb17d\") " pod="metallb-system/metallb-operator-webhook-server-6cb87f7d4c-jt2ns" Dec 11 02:16:04 crc kubenswrapper[4824]: I1211 02:16:04.988173 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0c1d110c-0bd5-410f-af71-7af964dfb17d-webhook-cert\") pod \"metallb-operator-webhook-server-6cb87f7d4c-jt2ns\" (UID: \"0c1d110c-0bd5-410f-af71-7af964dfb17d\") " pod="metallb-system/metallb-operator-webhook-server-6cb87f7d4c-jt2ns" Dec 11 02:16:05 crc kubenswrapper[4824]: I1211 02:16:05.003991 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0c1d110c-0bd5-410f-af71-7af964dfb17d-apiservice-cert\") pod \"metallb-operator-webhook-server-6cb87f7d4c-jt2ns\" (UID: \"0c1d110c-0bd5-410f-af71-7af964dfb17d\") " pod="metallb-system/metallb-operator-webhook-server-6cb87f7d4c-jt2ns" Dec 11 02:16:05 crc kubenswrapper[4824]: I1211 02:16:05.003995 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0c1d110c-0bd5-410f-af71-7af964dfb17d-webhook-cert\") pod \"metallb-operator-webhook-server-6cb87f7d4c-jt2ns\" (UID: \"0c1d110c-0bd5-410f-af71-7af964dfb17d\") " pod="metallb-system/metallb-operator-webhook-server-6cb87f7d4c-jt2ns" Dec 11 02:16:05 crc kubenswrapper[4824]: I1211 02:16:05.006908 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j599t\" (UniqueName: \"kubernetes.io/projected/0c1d110c-0bd5-410f-af71-7af964dfb17d-kube-api-access-j599t\") pod \"metallb-operator-webhook-server-6cb87f7d4c-jt2ns\" (UID: \"0c1d110c-0bd5-410f-af71-7af964dfb17d\") " pod="metallb-system/metallb-operator-webhook-server-6cb87f7d4c-jt2ns" Dec 11 02:16:05 crc kubenswrapper[4824]: I1211 02:16:05.048276 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-7d6477f866-c4dqb"] Dec 11 02:16:05 crc kubenswrapper[4824]: W1211 02:16:05.054831 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod47581f0c_dffe_42e8_8211_bb40625aadf1.slice/crio-5e2304b01c7b0a4fc56828425aeee7bf353df6b90e80c3aa0c09a523d43e4454 WatchSource:0}: Error finding container 5e2304b01c7b0a4fc56828425aeee7bf353df6b90e80c3aa0c09a523d43e4454: Status 404 returned error can't find the container with id 5e2304b01c7b0a4fc56828425aeee7bf353df6b90e80c3aa0c09a523d43e4454 Dec 11 02:16:05 crc kubenswrapper[4824]: I1211 02:16:05.055247 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-6cb87f7d4c-jt2ns" Dec 11 02:16:05 crc kubenswrapper[4824]: I1211 02:16:05.429008 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-7d6477f866-c4dqb" event={"ID":"47581f0c-dffe-42e8-8211-bb40625aadf1","Type":"ContainerStarted","Data":"5e2304b01c7b0a4fc56828425aeee7bf353df6b90e80c3aa0c09a523d43e4454"} Dec 11 02:16:05 crc kubenswrapper[4824]: I1211 02:16:05.470216 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-6cb87f7d4c-jt2ns"] Dec 11 02:16:05 crc kubenswrapper[4824]: W1211 02:16:05.478880 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0c1d110c_0bd5_410f_af71_7af964dfb17d.slice/crio-5a55dedb01ed2213414f5a9158603cbb861a2efd88a075b19d5a1da9312986f4 WatchSource:0}: Error finding container 5a55dedb01ed2213414f5a9158603cbb861a2efd88a075b19d5a1da9312986f4: Status 404 returned error can't find the container with id 5a55dedb01ed2213414f5a9158603cbb861a2efd88a075b19d5a1da9312986f4 Dec 11 02:16:06 crc kubenswrapper[4824]: I1211 02:16:06.438295 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-6cb87f7d4c-jt2ns" event={"ID":"0c1d110c-0bd5-410f-af71-7af964dfb17d","Type":"ContainerStarted","Data":"5a55dedb01ed2213414f5a9158603cbb861a2efd88a075b19d5a1da9312986f4"} Dec 11 02:16:07 crc kubenswrapper[4824]: I1211 02:16:07.256159 4824 patch_prober.go:28] interesting pod/machine-config-daemon-gx6xt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 02:16:07 crc kubenswrapper[4824]: I1211 02:16:07.256425 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 02:16:08 crc kubenswrapper[4824]: I1211 02:16:08.628702 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-ggtb8"] Dec 11 02:16:08 crc kubenswrapper[4824]: I1211 02:16:08.629678 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ggtb8" Dec 11 02:16:08 crc kubenswrapper[4824]: I1211 02:16:08.660184 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ggtb8"] Dec 11 02:16:08 crc kubenswrapper[4824]: I1211 02:16:08.755242 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7b4718b-2a67-47ab-a213-6977940a8699-utilities\") pod \"community-operators-ggtb8\" (UID: \"b7b4718b-2a67-47ab-a213-6977940a8699\") " pod="openshift-marketplace/community-operators-ggtb8" Dec 11 02:16:08 crc kubenswrapper[4824]: I1211 02:16:08.755678 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kw2xv\" (UniqueName: \"kubernetes.io/projected/b7b4718b-2a67-47ab-a213-6977940a8699-kube-api-access-kw2xv\") pod \"community-operators-ggtb8\" (UID: \"b7b4718b-2a67-47ab-a213-6977940a8699\") " pod="openshift-marketplace/community-operators-ggtb8" Dec 11 02:16:08 crc kubenswrapper[4824]: I1211 02:16:08.755755 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7b4718b-2a67-47ab-a213-6977940a8699-catalog-content\") pod \"community-operators-ggtb8\" (UID: \"b7b4718b-2a67-47ab-a213-6977940a8699\") " pod="openshift-marketplace/community-operators-ggtb8" Dec 11 02:16:08 crc kubenswrapper[4824]: I1211 02:16:08.856943 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kw2xv\" (UniqueName: \"kubernetes.io/projected/b7b4718b-2a67-47ab-a213-6977940a8699-kube-api-access-kw2xv\") pod \"community-operators-ggtb8\" (UID: \"b7b4718b-2a67-47ab-a213-6977940a8699\") " pod="openshift-marketplace/community-operators-ggtb8" Dec 11 02:16:08 crc kubenswrapper[4824]: I1211 02:16:08.857003 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7b4718b-2a67-47ab-a213-6977940a8699-catalog-content\") pod \"community-operators-ggtb8\" (UID: \"b7b4718b-2a67-47ab-a213-6977940a8699\") " pod="openshift-marketplace/community-operators-ggtb8" Dec 11 02:16:08 crc kubenswrapper[4824]: I1211 02:16:08.857048 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7b4718b-2a67-47ab-a213-6977940a8699-utilities\") pod \"community-operators-ggtb8\" (UID: \"b7b4718b-2a67-47ab-a213-6977940a8699\") " pod="openshift-marketplace/community-operators-ggtb8" Dec 11 02:16:08 crc kubenswrapper[4824]: I1211 02:16:08.857468 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7b4718b-2a67-47ab-a213-6977940a8699-utilities\") pod \"community-operators-ggtb8\" (UID: \"b7b4718b-2a67-47ab-a213-6977940a8699\") " pod="openshift-marketplace/community-operators-ggtb8" Dec 11 02:16:08 crc kubenswrapper[4824]: I1211 02:16:08.857559 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7b4718b-2a67-47ab-a213-6977940a8699-catalog-content\") pod \"community-operators-ggtb8\" (UID: \"b7b4718b-2a67-47ab-a213-6977940a8699\") " pod="openshift-marketplace/community-operators-ggtb8" Dec 11 02:16:08 crc kubenswrapper[4824]: I1211 02:16:08.875949 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kw2xv\" (UniqueName: \"kubernetes.io/projected/b7b4718b-2a67-47ab-a213-6977940a8699-kube-api-access-kw2xv\") pod \"community-operators-ggtb8\" (UID: \"b7b4718b-2a67-47ab-a213-6977940a8699\") " pod="openshift-marketplace/community-operators-ggtb8" Dec 11 02:16:08 crc kubenswrapper[4824]: I1211 02:16:08.945444 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ggtb8" Dec 11 02:16:09 crc kubenswrapper[4824]: I1211 02:16:09.429448 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ggtb8"] Dec 11 02:16:09 crc kubenswrapper[4824]: I1211 02:16:09.502444 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ggtb8" event={"ID":"b7b4718b-2a67-47ab-a213-6977940a8699","Type":"ContainerStarted","Data":"a7ca67d072930e4222f7b663a7842df3c3da3ab17465fecb636c3c214778a8e6"} Dec 11 02:16:09 crc kubenswrapper[4824]: I1211 02:16:09.503746 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-7d6477f866-c4dqb" event={"ID":"47581f0c-dffe-42e8-8211-bb40625aadf1","Type":"ContainerStarted","Data":"1447ac775168523a004569a4104e9ca206bd09923b8d6f6dcb0fe600a31fc7c3"} Dec 11 02:16:09 crc kubenswrapper[4824]: I1211 02:16:09.504520 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-7d6477f866-c4dqb" Dec 11 02:16:09 crc kubenswrapper[4824]: I1211 02:16:09.528127 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-7d6477f866-c4dqb" podStartSLOduration=1.366561302 podStartE2EDuration="5.528088314s" podCreationTimestamp="2025-12-11 02:16:04 +0000 UTC" firstStartedPulling="2025-12-11 02:16:05.061750487 +0000 UTC m=+906.750787886" lastFinishedPulling="2025-12-11 02:16:09.223277519 +0000 UTC m=+910.912314898" observedRunningTime="2025-12-11 02:16:09.522135399 +0000 UTC m=+911.211172778" watchObservedRunningTime="2025-12-11 02:16:09.528088314 +0000 UTC m=+911.217125693" Dec 11 02:16:10 crc kubenswrapper[4824]: I1211 02:16:10.512007 4824 generic.go:334] "Generic (PLEG): container finished" podID="b7b4718b-2a67-47ab-a213-6977940a8699" containerID="313289cb0a834701b0526c6a320f396e28b6f9026c4f1d935827f2da43ff1814" exitCode=0 Dec 11 02:16:10 crc kubenswrapper[4824]: I1211 02:16:10.512165 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ggtb8" event={"ID":"b7b4718b-2a67-47ab-a213-6977940a8699","Type":"ContainerDied","Data":"313289cb0a834701b0526c6a320f396e28b6f9026c4f1d935827f2da43ff1814"} Dec 11 02:16:12 crc kubenswrapper[4824]: I1211 02:16:12.528074 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-6cb87f7d4c-jt2ns" event={"ID":"0c1d110c-0bd5-410f-af71-7af964dfb17d","Type":"ContainerStarted","Data":"d393f3b19e8c2887561a8f3ca15c512e177b79855ea2c7d897071ef33fc14113"} Dec 11 02:16:12 crc kubenswrapper[4824]: I1211 02:16:12.528650 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-6cb87f7d4c-jt2ns" Dec 11 02:16:12 crc kubenswrapper[4824]: I1211 02:16:12.559434 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-6cb87f7d4c-jt2ns" podStartSLOduration=2.011430553 podStartE2EDuration="8.559410884s" podCreationTimestamp="2025-12-11 02:16:04 +0000 UTC" firstStartedPulling="2025-12-11 02:16:05.48251619 +0000 UTC m=+907.171553609" lastFinishedPulling="2025-12-11 02:16:12.030496531 +0000 UTC m=+913.719533940" observedRunningTime="2025-12-11 02:16:12.553834739 +0000 UTC m=+914.242872158" watchObservedRunningTime="2025-12-11 02:16:12.559410884 +0000 UTC m=+914.248448293" Dec 11 02:16:13 crc kubenswrapper[4824]: I1211 02:16:13.549743 4824 generic.go:334] "Generic (PLEG): container finished" podID="b7b4718b-2a67-47ab-a213-6977940a8699" containerID="793a7a707c6f5f5da3236b4bb702294ca96907f0f049bcb1436913ab6687ad8d" exitCode=0 Dec 11 02:16:13 crc kubenswrapper[4824]: I1211 02:16:13.549795 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ggtb8" event={"ID":"b7b4718b-2a67-47ab-a213-6977940a8699","Type":"ContainerDied","Data":"793a7a707c6f5f5da3236b4bb702294ca96907f0f049bcb1436913ab6687ad8d"} Dec 11 02:16:14 crc kubenswrapper[4824]: I1211 02:16:14.560377 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ggtb8" event={"ID":"b7b4718b-2a67-47ab-a213-6977940a8699","Type":"ContainerStarted","Data":"cee361e8f4e094e83c1bfc3e79903cf93407273ef927ae1d9d444cd574eb2c9c"} Dec 11 02:16:14 crc kubenswrapper[4824]: I1211 02:16:14.590234 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-ggtb8" podStartSLOduration=4.382333588 podStartE2EDuration="6.590207718s" podCreationTimestamp="2025-12-11 02:16:08 +0000 UTC" firstStartedPulling="2025-12-11 02:16:11.938518489 +0000 UTC m=+913.627555868" lastFinishedPulling="2025-12-11 02:16:14.146392579 +0000 UTC m=+915.835429998" observedRunningTime="2025-12-11 02:16:14.583503444 +0000 UTC m=+916.272540863" watchObservedRunningTime="2025-12-11 02:16:14.590207718 +0000 UTC m=+916.279245137" Dec 11 02:16:18 crc kubenswrapper[4824]: I1211 02:16:18.945862 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-ggtb8" Dec 11 02:16:18 crc kubenswrapper[4824]: I1211 02:16:18.946518 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-ggtb8" Dec 11 02:16:19 crc kubenswrapper[4824]: I1211 02:16:19.035229 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-ggtb8" Dec 11 02:16:19 crc kubenswrapper[4824]: I1211 02:16:19.653033 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-ggtb8" Dec 11 02:16:21 crc kubenswrapper[4824]: I1211 02:16:21.583185 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ggtb8"] Dec 11 02:16:21 crc kubenswrapper[4824]: I1211 02:16:21.601387 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-ggtb8" podUID="b7b4718b-2a67-47ab-a213-6977940a8699" containerName="registry-server" containerID="cri-o://cee361e8f4e094e83c1bfc3e79903cf93407273ef927ae1d9d444cd574eb2c9c" gracePeriod=2 Dec 11 02:16:21 crc kubenswrapper[4824]: I1211 02:16:21.973908 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ggtb8" Dec 11 02:16:22 crc kubenswrapper[4824]: I1211 02:16:22.056124 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7b4718b-2a67-47ab-a213-6977940a8699-catalog-content\") pod \"b7b4718b-2a67-47ab-a213-6977940a8699\" (UID: \"b7b4718b-2a67-47ab-a213-6977940a8699\") " Dec 11 02:16:22 crc kubenswrapper[4824]: I1211 02:16:22.056272 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7b4718b-2a67-47ab-a213-6977940a8699-utilities\") pod \"b7b4718b-2a67-47ab-a213-6977940a8699\" (UID: \"b7b4718b-2a67-47ab-a213-6977940a8699\") " Dec 11 02:16:22 crc kubenswrapper[4824]: I1211 02:16:22.056350 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kw2xv\" (UniqueName: \"kubernetes.io/projected/b7b4718b-2a67-47ab-a213-6977940a8699-kube-api-access-kw2xv\") pod \"b7b4718b-2a67-47ab-a213-6977940a8699\" (UID: \"b7b4718b-2a67-47ab-a213-6977940a8699\") " Dec 11 02:16:22 crc kubenswrapper[4824]: I1211 02:16:22.057089 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b7b4718b-2a67-47ab-a213-6977940a8699-utilities" (OuterVolumeSpecName: "utilities") pod "b7b4718b-2a67-47ab-a213-6977940a8699" (UID: "b7b4718b-2a67-47ab-a213-6977940a8699"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:16:22 crc kubenswrapper[4824]: I1211 02:16:22.063920 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7b4718b-2a67-47ab-a213-6977940a8699-kube-api-access-kw2xv" (OuterVolumeSpecName: "kube-api-access-kw2xv") pod "b7b4718b-2a67-47ab-a213-6977940a8699" (UID: "b7b4718b-2a67-47ab-a213-6977940a8699"). InnerVolumeSpecName "kube-api-access-kw2xv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:16:22 crc kubenswrapper[4824]: I1211 02:16:22.132393 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b7b4718b-2a67-47ab-a213-6977940a8699-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b7b4718b-2a67-47ab-a213-6977940a8699" (UID: "b7b4718b-2a67-47ab-a213-6977940a8699"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:16:22 crc kubenswrapper[4824]: I1211 02:16:22.157350 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kw2xv\" (UniqueName: \"kubernetes.io/projected/b7b4718b-2a67-47ab-a213-6977940a8699-kube-api-access-kw2xv\") on node \"crc\" DevicePath \"\"" Dec 11 02:16:22 crc kubenswrapper[4824]: I1211 02:16:22.157387 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7b4718b-2a67-47ab-a213-6977940a8699-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 02:16:22 crc kubenswrapper[4824]: I1211 02:16:22.157396 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7b4718b-2a67-47ab-a213-6977940a8699-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 02:16:22 crc kubenswrapper[4824]: I1211 02:16:22.607903 4824 generic.go:334] "Generic (PLEG): container finished" podID="b7b4718b-2a67-47ab-a213-6977940a8699" containerID="cee361e8f4e094e83c1bfc3e79903cf93407273ef927ae1d9d444cd574eb2c9c" exitCode=0 Dec 11 02:16:22 crc kubenswrapper[4824]: I1211 02:16:22.607989 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ggtb8" Dec 11 02:16:22 crc kubenswrapper[4824]: I1211 02:16:22.607987 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ggtb8" event={"ID":"b7b4718b-2a67-47ab-a213-6977940a8699","Type":"ContainerDied","Data":"cee361e8f4e094e83c1bfc3e79903cf93407273ef927ae1d9d444cd574eb2c9c"} Dec 11 02:16:22 crc kubenswrapper[4824]: I1211 02:16:22.608377 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ggtb8" event={"ID":"b7b4718b-2a67-47ab-a213-6977940a8699","Type":"ContainerDied","Data":"a7ca67d072930e4222f7b663a7842df3c3da3ab17465fecb636c3c214778a8e6"} Dec 11 02:16:22 crc kubenswrapper[4824]: I1211 02:16:22.608414 4824 scope.go:117] "RemoveContainer" containerID="cee361e8f4e094e83c1bfc3e79903cf93407273ef927ae1d9d444cd574eb2c9c" Dec 11 02:16:22 crc kubenswrapper[4824]: I1211 02:16:22.632440 4824 scope.go:117] "RemoveContainer" containerID="793a7a707c6f5f5da3236b4bb702294ca96907f0f049bcb1436913ab6687ad8d" Dec 11 02:16:22 crc kubenswrapper[4824]: I1211 02:16:22.649434 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ggtb8"] Dec 11 02:16:22 crc kubenswrapper[4824]: I1211 02:16:22.653697 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-ggtb8"] Dec 11 02:16:22 crc kubenswrapper[4824]: I1211 02:16:22.660536 4824 scope.go:117] "RemoveContainer" containerID="313289cb0a834701b0526c6a320f396e28b6f9026c4f1d935827f2da43ff1814" Dec 11 02:16:22 crc kubenswrapper[4824]: I1211 02:16:22.689927 4824 scope.go:117] "RemoveContainer" containerID="cee361e8f4e094e83c1bfc3e79903cf93407273ef927ae1d9d444cd574eb2c9c" Dec 11 02:16:22 crc kubenswrapper[4824]: E1211 02:16:22.690364 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cee361e8f4e094e83c1bfc3e79903cf93407273ef927ae1d9d444cd574eb2c9c\": container with ID starting with cee361e8f4e094e83c1bfc3e79903cf93407273ef927ae1d9d444cd574eb2c9c not found: ID does not exist" containerID="cee361e8f4e094e83c1bfc3e79903cf93407273ef927ae1d9d444cd574eb2c9c" Dec 11 02:16:22 crc kubenswrapper[4824]: I1211 02:16:22.690390 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cee361e8f4e094e83c1bfc3e79903cf93407273ef927ae1d9d444cd574eb2c9c"} err="failed to get container status \"cee361e8f4e094e83c1bfc3e79903cf93407273ef927ae1d9d444cd574eb2c9c\": rpc error: code = NotFound desc = could not find container \"cee361e8f4e094e83c1bfc3e79903cf93407273ef927ae1d9d444cd574eb2c9c\": container with ID starting with cee361e8f4e094e83c1bfc3e79903cf93407273ef927ae1d9d444cd574eb2c9c not found: ID does not exist" Dec 11 02:16:22 crc kubenswrapper[4824]: I1211 02:16:22.690410 4824 scope.go:117] "RemoveContainer" containerID="793a7a707c6f5f5da3236b4bb702294ca96907f0f049bcb1436913ab6687ad8d" Dec 11 02:16:22 crc kubenswrapper[4824]: E1211 02:16:22.690644 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"793a7a707c6f5f5da3236b4bb702294ca96907f0f049bcb1436913ab6687ad8d\": container with ID starting with 793a7a707c6f5f5da3236b4bb702294ca96907f0f049bcb1436913ab6687ad8d not found: ID does not exist" containerID="793a7a707c6f5f5da3236b4bb702294ca96907f0f049bcb1436913ab6687ad8d" Dec 11 02:16:22 crc kubenswrapper[4824]: I1211 02:16:22.690663 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"793a7a707c6f5f5da3236b4bb702294ca96907f0f049bcb1436913ab6687ad8d"} err="failed to get container status \"793a7a707c6f5f5da3236b4bb702294ca96907f0f049bcb1436913ab6687ad8d\": rpc error: code = NotFound desc = could not find container \"793a7a707c6f5f5da3236b4bb702294ca96907f0f049bcb1436913ab6687ad8d\": container with ID starting with 793a7a707c6f5f5da3236b4bb702294ca96907f0f049bcb1436913ab6687ad8d not found: ID does not exist" Dec 11 02:16:22 crc kubenswrapper[4824]: I1211 02:16:22.690676 4824 scope.go:117] "RemoveContainer" containerID="313289cb0a834701b0526c6a320f396e28b6f9026c4f1d935827f2da43ff1814" Dec 11 02:16:22 crc kubenswrapper[4824]: E1211 02:16:22.691214 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"313289cb0a834701b0526c6a320f396e28b6f9026c4f1d935827f2da43ff1814\": container with ID starting with 313289cb0a834701b0526c6a320f396e28b6f9026c4f1d935827f2da43ff1814 not found: ID does not exist" containerID="313289cb0a834701b0526c6a320f396e28b6f9026c4f1d935827f2da43ff1814" Dec 11 02:16:22 crc kubenswrapper[4824]: I1211 02:16:22.691296 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"313289cb0a834701b0526c6a320f396e28b6f9026c4f1d935827f2da43ff1814"} err="failed to get container status \"313289cb0a834701b0526c6a320f396e28b6f9026c4f1d935827f2da43ff1814\": rpc error: code = NotFound desc = could not find container \"313289cb0a834701b0526c6a320f396e28b6f9026c4f1d935827f2da43ff1814\": container with ID starting with 313289cb0a834701b0526c6a320f396e28b6f9026c4f1d935827f2da43ff1814 not found: ID does not exist" Dec 11 02:16:24 crc kubenswrapper[4824]: I1211 02:16:24.648483 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b7b4718b-2a67-47ab-a213-6977940a8699" path="/var/lib/kubelet/pods/b7b4718b-2a67-47ab-a213-6977940a8699/volumes" Dec 11 02:16:25 crc kubenswrapper[4824]: I1211 02:16:25.061733 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-6cb87f7d4c-jt2ns" Dec 11 02:16:37 crc kubenswrapper[4824]: I1211 02:16:37.251575 4824 patch_prober.go:28] interesting pod/machine-config-daemon-gx6xt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 02:16:37 crc kubenswrapper[4824]: I1211 02:16:37.252145 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 02:16:44 crc kubenswrapper[4824]: I1211 02:16:44.832817 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-7d6477f866-c4dqb" Dec 11 02:16:45 crc kubenswrapper[4824]: I1211 02:16:45.800159 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-s7hz7"] Dec 11 02:16:45 crc kubenswrapper[4824]: E1211 02:16:45.800892 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7b4718b-2a67-47ab-a213-6977940a8699" containerName="extract-content" Dec 11 02:16:45 crc kubenswrapper[4824]: I1211 02:16:45.800923 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7b4718b-2a67-47ab-a213-6977940a8699" containerName="extract-content" Dec 11 02:16:45 crc kubenswrapper[4824]: E1211 02:16:45.800951 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7b4718b-2a67-47ab-a213-6977940a8699" containerName="registry-server" Dec 11 02:16:45 crc kubenswrapper[4824]: I1211 02:16:45.800962 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7b4718b-2a67-47ab-a213-6977940a8699" containerName="registry-server" Dec 11 02:16:45 crc kubenswrapper[4824]: E1211 02:16:45.800980 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7b4718b-2a67-47ab-a213-6977940a8699" containerName="extract-utilities" Dec 11 02:16:45 crc kubenswrapper[4824]: I1211 02:16:45.800990 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7b4718b-2a67-47ab-a213-6977940a8699" containerName="extract-utilities" Dec 11 02:16:45 crc kubenswrapper[4824]: I1211 02:16:45.801415 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7b4718b-2a67-47ab-a213-6977940a8699" containerName="registry-server" Dec 11 02:16:45 crc kubenswrapper[4824]: I1211 02:16:45.804652 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-s7hz7" Dec 11 02:16:45 crc kubenswrapper[4824]: I1211 02:16:45.807230 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Dec 11 02:16:45 crc kubenswrapper[4824]: I1211 02:16:45.807916 4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-2tsrf" Dec 11 02:16:45 crc kubenswrapper[4824]: I1211 02:16:45.812008 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-mnh62"] Dec 11 02:16:45 crc kubenswrapper[4824]: I1211 02:16:45.812094 4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Dec 11 02:16:45 crc kubenswrapper[4824]: I1211 02:16:45.813399 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-mnh62" Dec 11 02:16:45 crc kubenswrapper[4824]: I1211 02:16:45.816220 4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Dec 11 02:16:45 crc kubenswrapper[4824]: I1211 02:16:45.818162 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-mnh62"] Dec 11 02:16:45 crc kubenswrapper[4824]: I1211 02:16:45.890999 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/57d0325f-b062-4148-8b2a-c21f42de39a3-frr-startup\") pod \"frr-k8s-s7hz7\" (UID: \"57d0325f-b062-4148-8b2a-c21f42de39a3\") " pod="metallb-system/frr-k8s-s7hz7" Dec 11 02:16:45 crc kubenswrapper[4824]: I1211 02:16:45.891907 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/57d0325f-b062-4148-8b2a-c21f42de39a3-metrics\") pod \"frr-k8s-s7hz7\" (UID: \"57d0325f-b062-4148-8b2a-c21f42de39a3\") " pod="metallb-system/frr-k8s-s7hz7" Dec 11 02:16:45 crc kubenswrapper[4824]: I1211 02:16:45.892034 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/57d0325f-b062-4148-8b2a-c21f42de39a3-frr-conf\") pod \"frr-k8s-s7hz7\" (UID: \"57d0325f-b062-4148-8b2a-c21f42de39a3\") " pod="metallb-system/frr-k8s-s7hz7" Dec 11 02:16:45 crc kubenswrapper[4824]: I1211 02:16:45.892166 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/57d0325f-b062-4148-8b2a-c21f42de39a3-frr-sockets\") pod \"frr-k8s-s7hz7\" (UID: \"57d0325f-b062-4148-8b2a-c21f42de39a3\") " pod="metallb-system/frr-k8s-s7hz7" Dec 11 02:16:45 crc kubenswrapper[4824]: I1211 02:16:45.892246 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t2g2d\" (UniqueName: \"kubernetes.io/projected/053662c4-f168-42b3-a9c9-dc21ca9908ea-kube-api-access-t2g2d\") pod \"frr-k8s-webhook-server-7fcb986d4-mnh62\" (UID: \"053662c4-f168-42b3-a9c9-dc21ca9908ea\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-mnh62" Dec 11 02:16:45 crc kubenswrapper[4824]: I1211 02:16:45.892332 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/053662c4-f168-42b3-a9c9-dc21ca9908ea-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-mnh62\" (UID: \"053662c4-f168-42b3-a9c9-dc21ca9908ea\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-mnh62" Dec 11 02:16:45 crc kubenswrapper[4824]: I1211 02:16:45.892402 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/57d0325f-b062-4148-8b2a-c21f42de39a3-metrics-certs\") pod \"frr-k8s-s7hz7\" (UID: \"57d0325f-b062-4148-8b2a-c21f42de39a3\") " pod="metallb-system/frr-k8s-s7hz7" Dec 11 02:16:45 crc kubenswrapper[4824]: I1211 02:16:45.892484 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/57d0325f-b062-4148-8b2a-c21f42de39a3-reloader\") pod \"frr-k8s-s7hz7\" (UID: \"57d0325f-b062-4148-8b2a-c21f42de39a3\") " pod="metallb-system/frr-k8s-s7hz7" Dec 11 02:16:45 crc kubenswrapper[4824]: I1211 02:16:45.892564 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-875js\" (UniqueName: \"kubernetes.io/projected/57d0325f-b062-4148-8b2a-c21f42de39a3-kube-api-access-875js\") pod \"frr-k8s-s7hz7\" (UID: \"57d0325f-b062-4148-8b2a-c21f42de39a3\") " pod="metallb-system/frr-k8s-s7hz7" Dec 11 02:16:45 crc kubenswrapper[4824]: I1211 02:16:45.898779 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-d99pv"] Dec 11 02:16:45 crc kubenswrapper[4824]: I1211 02:16:45.899716 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-d99pv" Dec 11 02:16:45 crc kubenswrapper[4824]: I1211 02:16:45.903952 4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-rmpvg" Dec 11 02:16:45 crc kubenswrapper[4824]: I1211 02:16:45.904153 4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Dec 11 02:16:45 crc kubenswrapper[4824]: I1211 02:16:45.904510 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Dec 11 02:16:45 crc kubenswrapper[4824]: I1211 02:16:45.904959 4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Dec 11 02:16:45 crc kubenswrapper[4824]: I1211 02:16:45.908564 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-f8648f98b-r49c8"] Dec 11 02:16:45 crc kubenswrapper[4824]: I1211 02:16:45.909411 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-r49c8" Dec 11 02:16:45 crc kubenswrapper[4824]: I1211 02:16:45.911173 4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Dec 11 02:16:45 crc kubenswrapper[4824]: I1211 02:16:45.922796 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-r49c8"] Dec 11 02:16:45 crc kubenswrapper[4824]: I1211 02:16:45.993459 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/57d0325f-b062-4148-8b2a-c21f42de39a3-frr-conf\") pod \"frr-k8s-s7hz7\" (UID: \"57d0325f-b062-4148-8b2a-c21f42de39a3\") " pod="metallb-system/frr-k8s-s7hz7" Dec 11 02:16:45 crc kubenswrapper[4824]: I1211 02:16:45.993509 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7fbb3013-54ad-421f-8022-26b0ffad44f6-metrics-certs\") pod \"controller-f8648f98b-r49c8\" (UID: \"7fbb3013-54ad-421f-8022-26b0ffad44f6\") " pod="metallb-system/controller-f8648f98b-r49c8" Dec 11 02:16:45 crc kubenswrapper[4824]: I1211 02:16:45.993541 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/6e59f040-b574-48e8-9a0d-05072fc1a3f6-memberlist\") pod \"speaker-d99pv\" (UID: \"6e59f040-b574-48e8-9a0d-05072fc1a3f6\") " pod="metallb-system/speaker-d99pv" Dec 11 02:16:45 crc kubenswrapper[4824]: I1211 02:16:45.993558 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7fbb3013-54ad-421f-8022-26b0ffad44f6-cert\") pod \"controller-f8648f98b-r49c8\" (UID: \"7fbb3013-54ad-421f-8022-26b0ffad44f6\") " pod="metallb-system/controller-f8648f98b-r49c8" Dec 11 02:16:45 crc kubenswrapper[4824]: I1211 02:16:45.993586 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vpk6w\" (UniqueName: \"kubernetes.io/projected/6e59f040-b574-48e8-9a0d-05072fc1a3f6-kube-api-access-vpk6w\") pod \"speaker-d99pv\" (UID: \"6e59f040-b574-48e8-9a0d-05072fc1a3f6\") " pod="metallb-system/speaker-d99pv" Dec 11 02:16:45 crc kubenswrapper[4824]: I1211 02:16:45.993602 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/57d0325f-b062-4148-8b2a-c21f42de39a3-frr-sockets\") pod \"frr-k8s-s7hz7\" (UID: \"57d0325f-b062-4148-8b2a-c21f42de39a3\") " pod="metallb-system/frr-k8s-s7hz7" Dec 11 02:16:45 crc kubenswrapper[4824]: I1211 02:16:45.993619 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t2g2d\" (UniqueName: \"kubernetes.io/projected/053662c4-f168-42b3-a9c9-dc21ca9908ea-kube-api-access-t2g2d\") pod \"frr-k8s-webhook-server-7fcb986d4-mnh62\" (UID: \"053662c4-f168-42b3-a9c9-dc21ca9908ea\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-mnh62" Dec 11 02:16:45 crc kubenswrapper[4824]: I1211 02:16:45.993706 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6e59f040-b574-48e8-9a0d-05072fc1a3f6-metrics-certs\") pod \"speaker-d99pv\" (UID: \"6e59f040-b574-48e8-9a0d-05072fc1a3f6\") " pod="metallb-system/speaker-d99pv" Dec 11 02:16:45 crc kubenswrapper[4824]: I1211 02:16:45.993756 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/053662c4-f168-42b3-a9c9-dc21ca9908ea-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-mnh62\" (UID: \"053662c4-f168-42b3-a9c9-dc21ca9908ea\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-mnh62" Dec 11 02:16:45 crc kubenswrapper[4824]: I1211 02:16:45.993774 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/57d0325f-b062-4148-8b2a-c21f42de39a3-metrics-certs\") pod \"frr-k8s-s7hz7\" (UID: \"57d0325f-b062-4148-8b2a-c21f42de39a3\") " pod="metallb-system/frr-k8s-s7hz7" Dec 11 02:16:45 crc kubenswrapper[4824]: I1211 02:16:45.993792 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/57d0325f-b062-4148-8b2a-c21f42de39a3-reloader\") pod \"frr-k8s-s7hz7\" (UID: \"57d0325f-b062-4148-8b2a-c21f42de39a3\") " pod="metallb-system/frr-k8s-s7hz7" Dec 11 02:16:45 crc kubenswrapper[4824]: I1211 02:16:45.993812 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/6e59f040-b574-48e8-9a0d-05072fc1a3f6-metallb-excludel2\") pod \"speaker-d99pv\" (UID: \"6e59f040-b574-48e8-9a0d-05072fc1a3f6\") " pod="metallb-system/speaker-d99pv" Dec 11 02:16:45 crc kubenswrapper[4824]: I1211 02:16:45.993829 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-875js\" (UniqueName: \"kubernetes.io/projected/57d0325f-b062-4148-8b2a-c21f42de39a3-kube-api-access-875js\") pod \"frr-k8s-s7hz7\" (UID: \"57d0325f-b062-4148-8b2a-c21f42de39a3\") " pod="metallb-system/frr-k8s-s7hz7" Dec 11 02:16:45 crc kubenswrapper[4824]: I1211 02:16:45.993850 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/57d0325f-b062-4148-8b2a-c21f42de39a3-frr-startup\") pod \"frr-k8s-s7hz7\" (UID: \"57d0325f-b062-4148-8b2a-c21f42de39a3\") " pod="metallb-system/frr-k8s-s7hz7" Dec 11 02:16:45 crc kubenswrapper[4824]: I1211 02:16:45.993863 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/57d0325f-b062-4148-8b2a-c21f42de39a3-metrics\") pod \"frr-k8s-s7hz7\" (UID: \"57d0325f-b062-4148-8b2a-c21f42de39a3\") " pod="metallb-system/frr-k8s-s7hz7" Dec 11 02:16:45 crc kubenswrapper[4824]: I1211 02:16:45.993882 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2vlxh\" (UniqueName: \"kubernetes.io/projected/7fbb3013-54ad-421f-8022-26b0ffad44f6-kube-api-access-2vlxh\") pod \"controller-f8648f98b-r49c8\" (UID: \"7fbb3013-54ad-421f-8022-26b0ffad44f6\") " pod="metallb-system/controller-f8648f98b-r49c8" Dec 11 02:16:45 crc kubenswrapper[4824]: I1211 02:16:45.993992 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/57d0325f-b062-4148-8b2a-c21f42de39a3-frr-conf\") pod \"frr-k8s-s7hz7\" (UID: \"57d0325f-b062-4148-8b2a-c21f42de39a3\") " pod="metallb-system/frr-k8s-s7hz7" Dec 11 02:16:45 crc kubenswrapper[4824]: I1211 02:16:45.994453 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/57d0325f-b062-4148-8b2a-c21f42de39a3-frr-sockets\") pod \"frr-k8s-s7hz7\" (UID: \"57d0325f-b062-4148-8b2a-c21f42de39a3\") " pod="metallb-system/frr-k8s-s7hz7" Dec 11 02:16:45 crc kubenswrapper[4824]: I1211 02:16:45.994691 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/57d0325f-b062-4148-8b2a-c21f42de39a3-reloader\") pod \"frr-k8s-s7hz7\" (UID: \"57d0325f-b062-4148-8b2a-c21f42de39a3\") " pod="metallb-system/frr-k8s-s7hz7" Dec 11 02:16:45 crc kubenswrapper[4824]: E1211 02:16:45.995346 4824 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Dec 11 02:16:45 crc kubenswrapper[4824]: I1211 02:16:45.995470 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/57d0325f-b062-4148-8b2a-c21f42de39a3-metrics\") pod \"frr-k8s-s7hz7\" (UID: \"57d0325f-b062-4148-8b2a-c21f42de39a3\") " pod="metallb-system/frr-k8s-s7hz7" Dec 11 02:16:45 crc kubenswrapper[4824]: E1211 02:16:45.995591 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/57d0325f-b062-4148-8b2a-c21f42de39a3-metrics-certs podName:57d0325f-b062-4148-8b2a-c21f42de39a3 nodeName:}" failed. No retries permitted until 2025-12-11 02:16:46.495462943 +0000 UTC m=+948.184500392 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/57d0325f-b062-4148-8b2a-c21f42de39a3-metrics-certs") pod "frr-k8s-s7hz7" (UID: "57d0325f-b062-4148-8b2a-c21f42de39a3") : secret "frr-k8s-certs-secret" not found Dec 11 02:16:45 crc kubenswrapper[4824]: I1211 02:16:45.995840 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/57d0325f-b062-4148-8b2a-c21f42de39a3-frr-startup\") pod \"frr-k8s-s7hz7\" (UID: \"57d0325f-b062-4148-8b2a-c21f42de39a3\") " pod="metallb-system/frr-k8s-s7hz7" Dec 11 02:16:46 crc kubenswrapper[4824]: I1211 02:16:46.011790 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/053662c4-f168-42b3-a9c9-dc21ca9908ea-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-mnh62\" (UID: \"053662c4-f168-42b3-a9c9-dc21ca9908ea\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-mnh62" Dec 11 02:16:46 crc kubenswrapper[4824]: I1211 02:16:46.017970 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t2g2d\" (UniqueName: \"kubernetes.io/projected/053662c4-f168-42b3-a9c9-dc21ca9908ea-kube-api-access-t2g2d\") pod \"frr-k8s-webhook-server-7fcb986d4-mnh62\" (UID: \"053662c4-f168-42b3-a9c9-dc21ca9908ea\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-mnh62" Dec 11 02:16:46 crc kubenswrapper[4824]: I1211 02:16:46.031600 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-875js\" (UniqueName: \"kubernetes.io/projected/57d0325f-b062-4148-8b2a-c21f42de39a3-kube-api-access-875js\") pod \"frr-k8s-s7hz7\" (UID: \"57d0325f-b062-4148-8b2a-c21f42de39a3\") " pod="metallb-system/frr-k8s-s7hz7" Dec 11 02:16:46 crc kubenswrapper[4824]: I1211 02:16:46.094660 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/6e59f040-b574-48e8-9a0d-05072fc1a3f6-metallb-excludel2\") pod \"speaker-d99pv\" (UID: \"6e59f040-b574-48e8-9a0d-05072fc1a3f6\") " pod="metallb-system/speaker-d99pv" Dec 11 02:16:46 crc kubenswrapper[4824]: I1211 02:16:46.094726 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2vlxh\" (UniqueName: \"kubernetes.io/projected/7fbb3013-54ad-421f-8022-26b0ffad44f6-kube-api-access-2vlxh\") pod \"controller-f8648f98b-r49c8\" (UID: \"7fbb3013-54ad-421f-8022-26b0ffad44f6\") " pod="metallb-system/controller-f8648f98b-r49c8" Dec 11 02:16:46 crc kubenswrapper[4824]: I1211 02:16:46.094774 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7fbb3013-54ad-421f-8022-26b0ffad44f6-metrics-certs\") pod \"controller-f8648f98b-r49c8\" (UID: \"7fbb3013-54ad-421f-8022-26b0ffad44f6\") " pod="metallb-system/controller-f8648f98b-r49c8" Dec 11 02:16:46 crc kubenswrapper[4824]: I1211 02:16:46.094795 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/6e59f040-b574-48e8-9a0d-05072fc1a3f6-memberlist\") pod \"speaker-d99pv\" (UID: \"6e59f040-b574-48e8-9a0d-05072fc1a3f6\") " pod="metallb-system/speaker-d99pv" Dec 11 02:16:46 crc kubenswrapper[4824]: I1211 02:16:46.094809 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7fbb3013-54ad-421f-8022-26b0ffad44f6-cert\") pod \"controller-f8648f98b-r49c8\" (UID: \"7fbb3013-54ad-421f-8022-26b0ffad44f6\") " pod="metallb-system/controller-f8648f98b-r49c8" Dec 11 02:16:46 crc kubenswrapper[4824]: I1211 02:16:46.094836 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vpk6w\" (UniqueName: \"kubernetes.io/projected/6e59f040-b574-48e8-9a0d-05072fc1a3f6-kube-api-access-vpk6w\") pod \"speaker-d99pv\" (UID: \"6e59f040-b574-48e8-9a0d-05072fc1a3f6\") " pod="metallb-system/speaker-d99pv" Dec 11 02:16:46 crc kubenswrapper[4824]: I1211 02:16:46.094856 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6e59f040-b574-48e8-9a0d-05072fc1a3f6-metrics-certs\") pod \"speaker-d99pv\" (UID: \"6e59f040-b574-48e8-9a0d-05072fc1a3f6\") " pod="metallb-system/speaker-d99pv" Dec 11 02:16:46 crc kubenswrapper[4824]: E1211 02:16:46.094915 4824 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 11 02:16:46 crc kubenswrapper[4824]: E1211 02:16:46.094996 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6e59f040-b574-48e8-9a0d-05072fc1a3f6-memberlist podName:6e59f040-b574-48e8-9a0d-05072fc1a3f6 nodeName:}" failed. No retries permitted until 2025-12-11 02:16:46.594973321 +0000 UTC m=+948.284010700 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/6e59f040-b574-48e8-9a0d-05072fc1a3f6-memberlist") pod "speaker-d99pv" (UID: "6e59f040-b574-48e8-9a0d-05072fc1a3f6") : secret "metallb-memberlist" not found Dec 11 02:16:46 crc kubenswrapper[4824]: E1211 02:16:46.095143 4824 secret.go:188] Couldn't get secret metallb-system/controller-certs-secret: secret "controller-certs-secret" not found Dec 11 02:16:46 crc kubenswrapper[4824]: E1211 02:16:46.095211 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7fbb3013-54ad-421f-8022-26b0ffad44f6-metrics-certs podName:7fbb3013-54ad-421f-8022-26b0ffad44f6 nodeName:}" failed. No retries permitted until 2025-12-11 02:16:46.595178937 +0000 UTC m=+948.284216316 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7fbb3013-54ad-421f-8022-26b0ffad44f6-metrics-certs") pod "controller-f8648f98b-r49c8" (UID: "7fbb3013-54ad-421f-8022-26b0ffad44f6") : secret "controller-certs-secret" not found Dec 11 02:16:46 crc kubenswrapper[4824]: I1211 02:16:46.095985 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/6e59f040-b574-48e8-9a0d-05072fc1a3f6-metallb-excludel2\") pod \"speaker-d99pv\" (UID: \"6e59f040-b574-48e8-9a0d-05072fc1a3f6\") " pod="metallb-system/speaker-d99pv" Dec 11 02:16:46 crc kubenswrapper[4824]: I1211 02:16:46.097493 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6e59f040-b574-48e8-9a0d-05072fc1a3f6-metrics-certs\") pod \"speaker-d99pv\" (UID: \"6e59f040-b574-48e8-9a0d-05072fc1a3f6\") " pod="metallb-system/speaker-d99pv" Dec 11 02:16:46 crc kubenswrapper[4824]: I1211 02:16:46.098077 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7fbb3013-54ad-421f-8022-26b0ffad44f6-cert\") pod \"controller-f8648f98b-r49c8\" (UID: \"7fbb3013-54ad-421f-8022-26b0ffad44f6\") " pod="metallb-system/controller-f8648f98b-r49c8" Dec 11 02:16:46 crc kubenswrapper[4824]: I1211 02:16:46.120268 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vpk6w\" (UniqueName: \"kubernetes.io/projected/6e59f040-b574-48e8-9a0d-05072fc1a3f6-kube-api-access-vpk6w\") pod \"speaker-d99pv\" (UID: \"6e59f040-b574-48e8-9a0d-05072fc1a3f6\") " pod="metallb-system/speaker-d99pv" Dec 11 02:16:46 crc kubenswrapper[4824]: I1211 02:16:46.122295 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2vlxh\" (UniqueName: \"kubernetes.io/projected/7fbb3013-54ad-421f-8022-26b0ffad44f6-kube-api-access-2vlxh\") pod \"controller-f8648f98b-r49c8\" (UID: \"7fbb3013-54ad-421f-8022-26b0ffad44f6\") " pod="metallb-system/controller-f8648f98b-r49c8" Dec 11 02:16:46 crc kubenswrapper[4824]: I1211 02:16:46.137673 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-mnh62" Dec 11 02:16:46 crc kubenswrapper[4824]: I1211 02:16:46.348311 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-mnh62"] Dec 11 02:16:46 crc kubenswrapper[4824]: I1211 02:16:46.498804 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/57d0325f-b062-4148-8b2a-c21f42de39a3-metrics-certs\") pod \"frr-k8s-s7hz7\" (UID: \"57d0325f-b062-4148-8b2a-c21f42de39a3\") " pod="metallb-system/frr-k8s-s7hz7" Dec 11 02:16:46 crc kubenswrapper[4824]: I1211 02:16:46.504489 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/57d0325f-b062-4148-8b2a-c21f42de39a3-metrics-certs\") pod \"frr-k8s-s7hz7\" (UID: \"57d0325f-b062-4148-8b2a-c21f42de39a3\") " pod="metallb-system/frr-k8s-s7hz7" Dec 11 02:16:46 crc kubenswrapper[4824]: I1211 02:16:46.599981 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7fbb3013-54ad-421f-8022-26b0ffad44f6-metrics-certs\") pod \"controller-f8648f98b-r49c8\" (UID: \"7fbb3013-54ad-421f-8022-26b0ffad44f6\") " pod="metallb-system/controller-f8648f98b-r49c8" Dec 11 02:16:46 crc kubenswrapper[4824]: I1211 02:16:46.600029 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/6e59f040-b574-48e8-9a0d-05072fc1a3f6-memberlist\") pod \"speaker-d99pv\" (UID: \"6e59f040-b574-48e8-9a0d-05072fc1a3f6\") " pod="metallb-system/speaker-d99pv" Dec 11 02:16:46 crc kubenswrapper[4824]: E1211 02:16:46.600180 4824 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 11 02:16:46 crc kubenswrapper[4824]: E1211 02:16:46.600238 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6e59f040-b574-48e8-9a0d-05072fc1a3f6-memberlist podName:6e59f040-b574-48e8-9a0d-05072fc1a3f6 nodeName:}" failed. No retries permitted until 2025-12-11 02:16:47.600222468 +0000 UTC m=+949.289259847 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/6e59f040-b574-48e8-9a0d-05072fc1a3f6-memberlist") pod "speaker-d99pv" (UID: "6e59f040-b574-48e8-9a0d-05072fc1a3f6") : secret "metallb-memberlist" not found Dec 11 02:16:46 crc kubenswrapper[4824]: I1211 02:16:46.603926 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7fbb3013-54ad-421f-8022-26b0ffad44f6-metrics-certs\") pod \"controller-f8648f98b-r49c8\" (UID: \"7fbb3013-54ad-421f-8022-26b0ffad44f6\") " pod="metallb-system/controller-f8648f98b-r49c8" Dec 11 02:16:46 crc kubenswrapper[4824]: I1211 02:16:46.723478 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-s7hz7" Dec 11 02:16:46 crc kubenswrapper[4824]: I1211 02:16:46.792928 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-mnh62" event={"ID":"053662c4-f168-42b3-a9c9-dc21ca9908ea","Type":"ContainerStarted","Data":"78d03fc838bda228572fdfd37fee4e5a2db0c7b6deba57eb4615345f06d1af48"} Dec 11 02:16:46 crc kubenswrapper[4824]: I1211 02:16:46.822778 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-r49c8" Dec 11 02:16:47 crc kubenswrapper[4824]: I1211 02:16:47.096482 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-r49c8"] Dec 11 02:16:47 crc kubenswrapper[4824]: I1211 02:16:47.613530 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/6e59f040-b574-48e8-9a0d-05072fc1a3f6-memberlist\") pod \"speaker-d99pv\" (UID: \"6e59f040-b574-48e8-9a0d-05072fc1a3f6\") " pod="metallb-system/speaker-d99pv" Dec 11 02:16:47 crc kubenswrapper[4824]: I1211 02:16:47.618446 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/6e59f040-b574-48e8-9a0d-05072fc1a3f6-memberlist\") pod \"speaker-d99pv\" (UID: \"6e59f040-b574-48e8-9a0d-05072fc1a3f6\") " pod="metallb-system/speaker-d99pv" Dec 11 02:16:47 crc kubenswrapper[4824]: I1211 02:16:47.725680 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-d99pv" Dec 11 02:16:47 crc kubenswrapper[4824]: W1211 02:16:47.748707 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6e59f040_b574_48e8_9a0d_05072fc1a3f6.slice/crio-d78e8b21987aeacc8d8d7031461067ad901d11f8f34cca352e578f432d29a401 WatchSource:0}: Error finding container d78e8b21987aeacc8d8d7031461067ad901d11f8f34cca352e578f432d29a401: Status 404 returned error can't find the container with id d78e8b21987aeacc8d8d7031461067ad901d11f8f34cca352e578f432d29a401 Dec 11 02:16:47 crc kubenswrapper[4824]: I1211 02:16:47.816464 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-r49c8" event={"ID":"7fbb3013-54ad-421f-8022-26b0ffad44f6","Type":"ContainerStarted","Data":"2893b4111a00a1e8000ebc97355622c57ff689f1da3a7d584c2ce08dc8681410"} Dec 11 02:16:47 crc kubenswrapper[4824]: I1211 02:16:47.816531 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-r49c8" event={"ID":"7fbb3013-54ad-421f-8022-26b0ffad44f6","Type":"ContainerStarted","Data":"45b9186ec79848ac9bcd6da7d06f1e8f09d5b5ee5306a4b1767b0256cdef7878"} Dec 11 02:16:47 crc kubenswrapper[4824]: I1211 02:16:47.816547 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-r49c8" event={"ID":"7fbb3013-54ad-421f-8022-26b0ffad44f6","Type":"ContainerStarted","Data":"c8758264c7b3dbb95fc961173b47ad07f138b0e20c73efb8d633c38497542035"} Dec 11 02:16:47 crc kubenswrapper[4824]: I1211 02:16:47.817332 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-f8648f98b-r49c8" Dec 11 02:16:47 crc kubenswrapper[4824]: I1211 02:16:47.818350 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-s7hz7" event={"ID":"57d0325f-b062-4148-8b2a-c21f42de39a3","Type":"ContainerStarted","Data":"065b1afbb659c7e1cd15871d933238c81fe464df1ceefea9dba6d5dcd0663e26"} Dec 11 02:16:47 crc kubenswrapper[4824]: I1211 02:16:47.819451 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-d99pv" event={"ID":"6e59f040-b574-48e8-9a0d-05072fc1a3f6","Type":"ContainerStarted","Data":"d78e8b21987aeacc8d8d7031461067ad901d11f8f34cca352e578f432d29a401"} Dec 11 02:16:47 crc kubenswrapper[4824]: I1211 02:16:47.839272 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-f8648f98b-r49c8" podStartSLOduration=2.839254635 podStartE2EDuration="2.839254635s" podCreationTimestamp="2025-12-11 02:16:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:16:47.838403583 +0000 UTC m=+949.527440972" watchObservedRunningTime="2025-12-11 02:16:47.839254635 +0000 UTC m=+949.528292014" Dec 11 02:16:48 crc kubenswrapper[4824]: I1211 02:16:48.837607 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-d99pv" event={"ID":"6e59f040-b574-48e8-9a0d-05072fc1a3f6","Type":"ContainerStarted","Data":"74cc253267f494b2d6ccd30406bcd0dd73f2dfaa224735d549ba1389d2def009"} Dec 11 02:16:48 crc kubenswrapper[4824]: I1211 02:16:48.838369 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-d99pv" event={"ID":"6e59f040-b574-48e8-9a0d-05072fc1a3f6","Type":"ContainerStarted","Data":"fce5b97f88a644e9c084bdfd1df9ff3b699e762f4b2d83dc529303bbfb69d4c5"} Dec 11 02:16:48 crc kubenswrapper[4824]: I1211 02:16:48.838390 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-d99pv" Dec 11 02:16:48 crc kubenswrapper[4824]: I1211 02:16:48.857211 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-d99pv" podStartSLOduration=3.857195213 podStartE2EDuration="3.857195213s" podCreationTimestamp="2025-12-11 02:16:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:16:48.855386476 +0000 UTC m=+950.544423845" watchObservedRunningTime="2025-12-11 02:16:48.857195213 +0000 UTC m=+950.546232592" Dec 11 02:16:54 crc kubenswrapper[4824]: I1211 02:16:54.875070 4824 generic.go:334] "Generic (PLEG): container finished" podID="57d0325f-b062-4148-8b2a-c21f42de39a3" containerID="e16a7f06f81a8a09e138a08a19e8a220b2a18265122714f9e1b904e0a8a61ed8" exitCode=0 Dec 11 02:16:54 crc kubenswrapper[4824]: I1211 02:16:54.875207 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-s7hz7" event={"ID":"57d0325f-b062-4148-8b2a-c21f42de39a3","Type":"ContainerDied","Data":"e16a7f06f81a8a09e138a08a19e8a220b2a18265122714f9e1b904e0a8a61ed8"} Dec 11 02:16:54 crc kubenswrapper[4824]: I1211 02:16:54.878283 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-mnh62" event={"ID":"053662c4-f168-42b3-a9c9-dc21ca9908ea","Type":"ContainerStarted","Data":"56d7a6b80a77d904989ce4cf39937ee78b778475d2675333f8c6eab31f8e2960"} Dec 11 02:16:54 crc kubenswrapper[4824]: I1211 02:16:54.878568 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-mnh62" Dec 11 02:16:54 crc kubenswrapper[4824]: I1211 02:16:54.932644 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-mnh62" podStartSLOduration=2.15295433 podStartE2EDuration="9.932620283s" podCreationTimestamp="2025-12-11 02:16:45 +0000 UTC" firstStartedPulling="2025-12-11 02:16:46.358876643 +0000 UTC m=+948.047914022" lastFinishedPulling="2025-12-11 02:16:54.138542566 +0000 UTC m=+955.827579975" observedRunningTime="2025-12-11 02:16:54.929163933 +0000 UTC m=+956.618201342" watchObservedRunningTime="2025-12-11 02:16:54.932620283 +0000 UTC m=+956.621657692" Dec 11 02:16:55 crc kubenswrapper[4824]: I1211 02:16:55.888902 4824 generic.go:334] "Generic (PLEG): container finished" podID="57d0325f-b062-4148-8b2a-c21f42de39a3" containerID="45e56840865960ba6eef0576af82cab230592dacc7f76038e6834a66b0001f16" exitCode=0 Dec 11 02:16:55 crc kubenswrapper[4824]: I1211 02:16:55.888973 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-s7hz7" event={"ID":"57d0325f-b062-4148-8b2a-c21f42de39a3","Type":"ContainerDied","Data":"45e56840865960ba6eef0576af82cab230592dacc7f76038e6834a66b0001f16"} Dec 11 02:16:56 crc kubenswrapper[4824]: I1211 02:16:56.901804 4824 generic.go:334] "Generic (PLEG): container finished" podID="57d0325f-b062-4148-8b2a-c21f42de39a3" containerID="462c5f896c7039fb7509d68574aa555b7a9673392c0b465f186ed4efbae25269" exitCode=0 Dec 11 02:16:56 crc kubenswrapper[4824]: I1211 02:16:56.901914 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-s7hz7" event={"ID":"57d0325f-b062-4148-8b2a-c21f42de39a3","Type":"ContainerDied","Data":"462c5f896c7039fb7509d68574aa555b7a9673392c0b465f186ed4efbae25269"} Dec 11 02:16:57 crc kubenswrapper[4824]: I1211 02:16:57.915637 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-s7hz7" event={"ID":"57d0325f-b062-4148-8b2a-c21f42de39a3","Type":"ContainerStarted","Data":"e78f93d8f84f82b5fb26f2873db57c31eec5e533507e9901e8f1573ab34ae725"} Dec 11 02:16:57 crc kubenswrapper[4824]: I1211 02:16:57.915759 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-s7hz7" event={"ID":"57d0325f-b062-4148-8b2a-c21f42de39a3","Type":"ContainerStarted","Data":"22e73980628bd3a7a4768d93c9801289ac5fd53daa7bc78fa9d06be61ed8ffe2"} Dec 11 02:16:57 crc kubenswrapper[4824]: I1211 02:16:57.915792 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-s7hz7" event={"ID":"57d0325f-b062-4148-8b2a-c21f42de39a3","Type":"ContainerStarted","Data":"daf13a62611c41e9ce957f995e1037726edd5ff143dfb2781a201276df3a5b34"} Dec 11 02:16:57 crc kubenswrapper[4824]: I1211 02:16:57.915818 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-s7hz7" event={"ID":"57d0325f-b062-4148-8b2a-c21f42de39a3","Type":"ContainerStarted","Data":"8337d2e668df78e7068c66ef05d0e5a297aaf554f9ed4e0784adafa84d2ba0e1"} Dec 11 02:16:58 crc kubenswrapper[4824]: I1211 02:16:58.930655 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-s7hz7" event={"ID":"57d0325f-b062-4148-8b2a-c21f42de39a3","Type":"ContainerStarted","Data":"b807c3c3e9c0c782790884ad87619ad70e5b4e1bce3f591114fbeda7ce320f81"} Dec 11 02:16:59 crc kubenswrapper[4824]: I1211 02:16:59.948532 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-s7hz7" event={"ID":"57d0325f-b062-4148-8b2a-c21f42de39a3","Type":"ContainerStarted","Data":"515a100f0f268659dbd16dc7444287f998ac3f3b053b7b51bb9ea34a4048669e"} Dec 11 02:16:59 crc kubenswrapper[4824]: I1211 02:16:59.949033 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-s7hz7" Dec 11 02:16:59 crc kubenswrapper[4824]: I1211 02:16:59.988743 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-s7hz7" podStartSLOduration=7.775361441 podStartE2EDuration="14.988723859s" podCreationTimestamp="2025-12-11 02:16:45 +0000 UTC" firstStartedPulling="2025-12-11 02:16:46.919362397 +0000 UTC m=+948.608399816" lastFinishedPulling="2025-12-11 02:16:54.132724845 +0000 UTC m=+955.821762234" observedRunningTime="2025-12-11 02:16:59.984259554 +0000 UTC m=+961.673296963" watchObservedRunningTime="2025-12-11 02:16:59.988723859 +0000 UTC m=+961.677761238" Dec 11 02:17:01 crc kubenswrapper[4824]: I1211 02:17:01.723898 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-s7hz7" Dec 11 02:17:01 crc kubenswrapper[4824]: I1211 02:17:01.784424 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-s7hz7" Dec 11 02:17:06 crc kubenswrapper[4824]: I1211 02:17:06.147736 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-mnh62" Dec 11 02:17:06 crc kubenswrapper[4824]: I1211 02:17:06.830211 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-f8648f98b-r49c8" Dec 11 02:17:07 crc kubenswrapper[4824]: I1211 02:17:07.251365 4824 patch_prober.go:28] interesting pod/machine-config-daemon-gx6xt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 02:17:07 crc kubenswrapper[4824]: I1211 02:17:07.251471 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 02:17:07 crc kubenswrapper[4824]: I1211 02:17:07.251536 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" Dec 11 02:17:07 crc kubenswrapper[4824]: I1211 02:17:07.252527 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"65dbaed5184dd7e582c2d25d2c2754654e16b28021f7e2281401ebdcef00a13e"} pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 11 02:17:07 crc kubenswrapper[4824]: I1211 02:17:07.252625 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" containerName="machine-config-daemon" containerID="cri-o://65dbaed5184dd7e582c2d25d2c2754654e16b28021f7e2281401ebdcef00a13e" gracePeriod=600 Dec 11 02:17:07 crc kubenswrapper[4824]: I1211 02:17:07.729291 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-d99pv" Dec 11 02:17:08 crc kubenswrapper[4824]: I1211 02:17:08.019925 4824 generic.go:334] "Generic (PLEG): container finished" podID="44b156e4-64a4-4d45-aa5e-9b10a862faed" containerID="65dbaed5184dd7e582c2d25d2c2754654e16b28021f7e2281401ebdcef00a13e" exitCode=0 Dec 11 02:17:08 crc kubenswrapper[4824]: I1211 02:17:08.020024 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" event={"ID":"44b156e4-64a4-4d45-aa5e-9b10a862faed","Type":"ContainerDied","Data":"65dbaed5184dd7e582c2d25d2c2754654e16b28021f7e2281401ebdcef00a13e"} Dec 11 02:17:08 crc kubenswrapper[4824]: I1211 02:17:08.020398 4824 scope.go:117] "RemoveContainer" containerID="db8b560cfc3483491a894f8e917ea1a5928fa44440c12ac1a044fd79b218d54a" Dec 11 02:17:08 crc kubenswrapper[4824]: I1211 02:17:08.048498 4824 scope.go:117] "RemoveContainer" containerID="820ab8fffa074e96d20e064678952b74a92387b5649face8929acb3791ff6e88" Dec 11 02:17:09 crc kubenswrapper[4824]: I1211 02:17:09.032462 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" event={"ID":"44b156e4-64a4-4d45-aa5e-9b10a862faed","Type":"ContainerStarted","Data":"d1afa0a444a8a2d00d6d09bdd1c078096ee2bfd1a3af1878373a7a899d9f6039"} Dec 11 02:17:11 crc kubenswrapper[4824]: I1211 02:17:11.501647 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-n44ms"] Dec 11 02:17:11 crc kubenswrapper[4824]: I1211 02:17:11.505062 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-n44ms" Dec 11 02:17:11 crc kubenswrapper[4824]: I1211 02:17:11.507716 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Dec 11 02:17:11 crc kubenswrapper[4824]: I1211 02:17:11.508922 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Dec 11 02:17:11 crc kubenswrapper[4824]: I1211 02:17:11.512877 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-k4l4x" Dec 11 02:17:11 crc kubenswrapper[4824]: I1211 02:17:11.522105 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-n44ms"] Dec 11 02:17:11 crc kubenswrapper[4824]: I1211 02:17:11.549761 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8cjsq\" (UniqueName: \"kubernetes.io/projected/a98a64f0-92df-444c-91ff-961f6fed7d5f-kube-api-access-8cjsq\") pod \"openstack-operator-index-n44ms\" (UID: \"a98a64f0-92df-444c-91ff-961f6fed7d5f\") " pod="openstack-operators/openstack-operator-index-n44ms" Dec 11 02:17:11 crc kubenswrapper[4824]: I1211 02:17:11.650557 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8cjsq\" (UniqueName: \"kubernetes.io/projected/a98a64f0-92df-444c-91ff-961f6fed7d5f-kube-api-access-8cjsq\") pod \"openstack-operator-index-n44ms\" (UID: \"a98a64f0-92df-444c-91ff-961f6fed7d5f\") " pod="openstack-operators/openstack-operator-index-n44ms" Dec 11 02:17:11 crc kubenswrapper[4824]: I1211 02:17:11.668868 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8cjsq\" (UniqueName: \"kubernetes.io/projected/a98a64f0-92df-444c-91ff-961f6fed7d5f-kube-api-access-8cjsq\") pod \"openstack-operator-index-n44ms\" (UID: \"a98a64f0-92df-444c-91ff-961f6fed7d5f\") " pod="openstack-operators/openstack-operator-index-n44ms" Dec 11 02:17:11 crc kubenswrapper[4824]: I1211 02:17:11.833360 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-n44ms" Dec 11 02:17:12 crc kubenswrapper[4824]: I1211 02:17:12.119920 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-n44ms"] Dec 11 02:17:12 crc kubenswrapper[4824]: W1211 02:17:12.129752 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda98a64f0_92df_444c_91ff_961f6fed7d5f.slice/crio-4dfe8a57de2532f6d8f6dbef2a6993773a4ab320ebd245aba2b21b850596578e WatchSource:0}: Error finding container 4dfe8a57de2532f6d8f6dbef2a6993773a4ab320ebd245aba2b21b850596578e: Status 404 returned error can't find the container with id 4dfe8a57de2532f6d8f6dbef2a6993773a4ab320ebd245aba2b21b850596578e Dec 11 02:17:13 crc kubenswrapper[4824]: I1211 02:17:13.066362 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-n44ms" event={"ID":"a98a64f0-92df-444c-91ff-961f6fed7d5f","Type":"ContainerStarted","Data":"4dfe8a57de2532f6d8f6dbef2a6993773a4ab320ebd245aba2b21b850596578e"} Dec 11 02:17:14 crc kubenswrapper[4824]: I1211 02:17:14.849501 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-n44ms"] Dec 11 02:17:15 crc kubenswrapper[4824]: I1211 02:17:15.089834 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-n44ms" event={"ID":"a98a64f0-92df-444c-91ff-961f6fed7d5f","Type":"ContainerStarted","Data":"00c6ef5673dee71054f00d79f64596e54b002cdd7de065d4163eca096e04b217"} Dec 11 02:17:15 crc kubenswrapper[4824]: I1211 02:17:15.114823 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-n44ms" podStartSLOduration=2.247337235 podStartE2EDuration="4.114798862s" podCreationTimestamp="2025-12-11 02:17:11 +0000 UTC" firstStartedPulling="2025-12-11 02:17:12.135062644 +0000 UTC m=+973.824100033" lastFinishedPulling="2025-12-11 02:17:14.002524281 +0000 UTC m=+975.691561660" observedRunningTime="2025-12-11 02:17:15.108891808 +0000 UTC m=+976.797929237" watchObservedRunningTime="2025-12-11 02:17:15.114798862 +0000 UTC m=+976.803836271" Dec 11 02:17:15 crc kubenswrapper[4824]: I1211 02:17:15.462342 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-nb526"] Dec 11 02:17:15 crc kubenswrapper[4824]: I1211 02:17:15.463516 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-nb526" Dec 11 02:17:15 crc kubenswrapper[4824]: I1211 02:17:15.495083 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-nb526"] Dec 11 02:17:15 crc kubenswrapper[4824]: I1211 02:17:15.606509 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q99q9\" (UniqueName: \"kubernetes.io/projected/ab26e0f0-f195-481f-84bd-0a7bca8d0432-kube-api-access-q99q9\") pod \"openstack-operator-index-nb526\" (UID: \"ab26e0f0-f195-481f-84bd-0a7bca8d0432\") " pod="openstack-operators/openstack-operator-index-nb526" Dec 11 02:17:15 crc kubenswrapper[4824]: I1211 02:17:15.708715 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q99q9\" (UniqueName: \"kubernetes.io/projected/ab26e0f0-f195-481f-84bd-0a7bca8d0432-kube-api-access-q99q9\") pod \"openstack-operator-index-nb526\" (UID: \"ab26e0f0-f195-481f-84bd-0a7bca8d0432\") " pod="openstack-operators/openstack-operator-index-nb526" Dec 11 02:17:15 crc kubenswrapper[4824]: I1211 02:17:15.742064 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q99q9\" (UniqueName: \"kubernetes.io/projected/ab26e0f0-f195-481f-84bd-0a7bca8d0432-kube-api-access-q99q9\") pod \"openstack-operator-index-nb526\" (UID: \"ab26e0f0-f195-481f-84bd-0a7bca8d0432\") " pod="openstack-operators/openstack-operator-index-nb526" Dec 11 02:17:15 crc kubenswrapper[4824]: I1211 02:17:15.801833 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-nb526" Dec 11 02:17:16 crc kubenswrapper[4824]: I1211 02:17:16.095454 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-n44ms" podUID="a98a64f0-92df-444c-91ff-961f6fed7d5f" containerName="registry-server" containerID="cri-o://00c6ef5673dee71054f00d79f64596e54b002cdd7de065d4163eca096e04b217" gracePeriod=2 Dec 11 02:17:16 crc kubenswrapper[4824]: I1211 02:17:16.268732 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-nb526"] Dec 11 02:17:16 crc kubenswrapper[4824]: W1211 02:17:16.276646 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podab26e0f0_f195_481f_84bd_0a7bca8d0432.slice/crio-c18e6665caded0b2383159934e0d45be5f9368eae9fa31e29c7536c2d8085625 WatchSource:0}: Error finding container c18e6665caded0b2383159934e0d45be5f9368eae9fa31e29c7536c2d8085625: Status 404 returned error can't find the container with id c18e6665caded0b2383159934e0d45be5f9368eae9fa31e29c7536c2d8085625 Dec 11 02:17:16 crc kubenswrapper[4824]: I1211 02:17:16.422054 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-n44ms" Dec 11 02:17:16 crc kubenswrapper[4824]: I1211 02:17:16.523022 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8cjsq\" (UniqueName: \"kubernetes.io/projected/a98a64f0-92df-444c-91ff-961f6fed7d5f-kube-api-access-8cjsq\") pod \"a98a64f0-92df-444c-91ff-961f6fed7d5f\" (UID: \"a98a64f0-92df-444c-91ff-961f6fed7d5f\") " Dec 11 02:17:16 crc kubenswrapper[4824]: I1211 02:17:16.528816 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a98a64f0-92df-444c-91ff-961f6fed7d5f-kube-api-access-8cjsq" (OuterVolumeSpecName: "kube-api-access-8cjsq") pod "a98a64f0-92df-444c-91ff-961f6fed7d5f" (UID: "a98a64f0-92df-444c-91ff-961f6fed7d5f"). InnerVolumeSpecName "kube-api-access-8cjsq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:17:16 crc kubenswrapper[4824]: I1211 02:17:16.625242 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8cjsq\" (UniqueName: \"kubernetes.io/projected/a98a64f0-92df-444c-91ff-961f6fed7d5f-kube-api-access-8cjsq\") on node \"crc\" DevicePath \"\"" Dec 11 02:17:16 crc kubenswrapper[4824]: I1211 02:17:16.728382 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-s7hz7" Dec 11 02:17:17 crc kubenswrapper[4824]: I1211 02:17:17.103758 4824 generic.go:334] "Generic (PLEG): container finished" podID="a98a64f0-92df-444c-91ff-961f6fed7d5f" containerID="00c6ef5673dee71054f00d79f64596e54b002cdd7de065d4163eca096e04b217" exitCode=0 Dec 11 02:17:17 crc kubenswrapper[4824]: I1211 02:17:17.103872 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-n44ms" Dec 11 02:17:17 crc kubenswrapper[4824]: I1211 02:17:17.103874 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-n44ms" event={"ID":"a98a64f0-92df-444c-91ff-961f6fed7d5f","Type":"ContainerDied","Data":"00c6ef5673dee71054f00d79f64596e54b002cdd7de065d4163eca096e04b217"} Dec 11 02:17:17 crc kubenswrapper[4824]: I1211 02:17:17.104040 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-n44ms" event={"ID":"a98a64f0-92df-444c-91ff-961f6fed7d5f","Type":"ContainerDied","Data":"4dfe8a57de2532f6d8f6dbef2a6993773a4ab320ebd245aba2b21b850596578e"} Dec 11 02:17:17 crc kubenswrapper[4824]: I1211 02:17:17.104077 4824 scope.go:117] "RemoveContainer" containerID="00c6ef5673dee71054f00d79f64596e54b002cdd7de065d4163eca096e04b217" Dec 11 02:17:17 crc kubenswrapper[4824]: I1211 02:17:17.107190 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-nb526" event={"ID":"ab26e0f0-f195-481f-84bd-0a7bca8d0432","Type":"ContainerStarted","Data":"0bdb7e3b51b8d71c95ba22b5b022f3e69b7347b66f17114000c0ecbdb6645994"} Dec 11 02:17:17 crc kubenswrapper[4824]: I1211 02:17:17.107223 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-nb526" event={"ID":"ab26e0f0-f195-481f-84bd-0a7bca8d0432","Type":"ContainerStarted","Data":"c18e6665caded0b2383159934e0d45be5f9368eae9fa31e29c7536c2d8085625"} Dec 11 02:17:17 crc kubenswrapper[4824]: I1211 02:17:17.134163 4824 scope.go:117] "RemoveContainer" containerID="00c6ef5673dee71054f00d79f64596e54b002cdd7de065d4163eca096e04b217" Dec 11 02:17:17 crc kubenswrapper[4824]: E1211 02:17:17.135331 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"00c6ef5673dee71054f00d79f64596e54b002cdd7de065d4163eca096e04b217\": container with ID starting with 00c6ef5673dee71054f00d79f64596e54b002cdd7de065d4163eca096e04b217 not found: ID does not exist" containerID="00c6ef5673dee71054f00d79f64596e54b002cdd7de065d4163eca096e04b217" Dec 11 02:17:17 crc kubenswrapper[4824]: I1211 02:17:17.135409 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00c6ef5673dee71054f00d79f64596e54b002cdd7de065d4163eca096e04b217"} err="failed to get container status \"00c6ef5673dee71054f00d79f64596e54b002cdd7de065d4163eca096e04b217\": rpc error: code = NotFound desc = could not find container \"00c6ef5673dee71054f00d79f64596e54b002cdd7de065d4163eca096e04b217\": container with ID starting with 00c6ef5673dee71054f00d79f64596e54b002cdd7de065d4163eca096e04b217 not found: ID does not exist" Dec 11 02:17:17 crc kubenswrapper[4824]: I1211 02:17:17.137385 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-nb526" podStartSLOduration=2.082368582 podStartE2EDuration="2.137359782s" podCreationTimestamp="2025-12-11 02:17:15 +0000 UTC" firstStartedPulling="2025-12-11 02:17:16.281181159 +0000 UTC m=+977.970218568" lastFinishedPulling="2025-12-11 02:17:16.336172349 +0000 UTC m=+978.025209768" observedRunningTime="2025-12-11 02:17:17.133811689 +0000 UTC m=+978.822849098" watchObservedRunningTime="2025-12-11 02:17:17.137359782 +0000 UTC m=+978.826397191" Dec 11 02:17:17 crc kubenswrapper[4824]: I1211 02:17:17.151106 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-n44ms"] Dec 11 02:17:17 crc kubenswrapper[4824]: I1211 02:17:17.156398 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-n44ms"] Dec 11 02:17:18 crc kubenswrapper[4824]: I1211 02:17:18.647454 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a98a64f0-92df-444c-91ff-961f6fed7d5f" path="/var/lib/kubelet/pods/a98a64f0-92df-444c-91ff-961f6fed7d5f/volumes" Dec 11 02:17:25 crc kubenswrapper[4824]: I1211 02:17:25.802626 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-nb526" Dec 11 02:17:25 crc kubenswrapper[4824]: I1211 02:17:25.804263 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-nb526" Dec 11 02:17:25 crc kubenswrapper[4824]: I1211 02:17:25.838569 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-nb526" Dec 11 02:17:26 crc kubenswrapper[4824]: I1211 02:17:26.205925 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-nb526" Dec 11 02:17:39 crc kubenswrapper[4824]: I1211 02:17:39.317623 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05vd4cz"] Dec 11 02:17:39 crc kubenswrapper[4824]: E1211 02:17:39.318501 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a98a64f0-92df-444c-91ff-961f6fed7d5f" containerName="registry-server" Dec 11 02:17:39 crc kubenswrapper[4824]: I1211 02:17:39.318518 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="a98a64f0-92df-444c-91ff-961f6fed7d5f" containerName="registry-server" Dec 11 02:17:39 crc kubenswrapper[4824]: I1211 02:17:39.318660 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="a98a64f0-92df-444c-91ff-961f6fed7d5f" containerName="registry-server" Dec 11 02:17:39 crc kubenswrapper[4824]: I1211 02:17:39.319604 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05vd4cz" Dec 11 02:17:39 crc kubenswrapper[4824]: I1211 02:17:39.324643 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-n4nnw" Dec 11 02:17:39 crc kubenswrapper[4824]: I1211 02:17:39.334368 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05vd4cz"] Dec 11 02:17:39 crc kubenswrapper[4824]: I1211 02:17:39.481228 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/487cf0fc-4368-4f0e-878a-d3cd3d1e54f4-util\") pod \"50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05vd4cz\" (UID: \"487cf0fc-4368-4f0e-878a-d3cd3d1e54f4\") " pod="openstack-operators/50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05vd4cz" Dec 11 02:17:39 crc kubenswrapper[4824]: I1211 02:17:39.481730 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/487cf0fc-4368-4f0e-878a-d3cd3d1e54f4-bundle\") pod \"50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05vd4cz\" (UID: \"487cf0fc-4368-4f0e-878a-d3cd3d1e54f4\") " pod="openstack-operators/50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05vd4cz" Dec 11 02:17:39 crc kubenswrapper[4824]: I1211 02:17:39.482009 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-26b6m\" (UniqueName: \"kubernetes.io/projected/487cf0fc-4368-4f0e-878a-d3cd3d1e54f4-kube-api-access-26b6m\") pod \"50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05vd4cz\" (UID: \"487cf0fc-4368-4f0e-878a-d3cd3d1e54f4\") " pod="openstack-operators/50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05vd4cz" Dec 11 02:17:39 crc kubenswrapper[4824]: I1211 02:17:39.583459 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-26b6m\" (UniqueName: \"kubernetes.io/projected/487cf0fc-4368-4f0e-878a-d3cd3d1e54f4-kube-api-access-26b6m\") pod \"50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05vd4cz\" (UID: \"487cf0fc-4368-4f0e-878a-d3cd3d1e54f4\") " pod="openstack-operators/50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05vd4cz" Dec 11 02:17:39 crc kubenswrapper[4824]: I1211 02:17:39.583908 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/487cf0fc-4368-4f0e-878a-d3cd3d1e54f4-util\") pod \"50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05vd4cz\" (UID: \"487cf0fc-4368-4f0e-878a-d3cd3d1e54f4\") " pod="openstack-operators/50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05vd4cz" Dec 11 02:17:39 crc kubenswrapper[4824]: I1211 02:17:39.584246 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/487cf0fc-4368-4f0e-878a-d3cd3d1e54f4-bundle\") pod \"50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05vd4cz\" (UID: \"487cf0fc-4368-4f0e-878a-d3cd3d1e54f4\") " pod="openstack-operators/50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05vd4cz" Dec 11 02:17:39 crc kubenswrapper[4824]: I1211 02:17:39.585092 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/487cf0fc-4368-4f0e-878a-d3cd3d1e54f4-bundle\") pod \"50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05vd4cz\" (UID: \"487cf0fc-4368-4f0e-878a-d3cd3d1e54f4\") " pod="openstack-operators/50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05vd4cz" Dec 11 02:17:39 crc kubenswrapper[4824]: I1211 02:17:39.585094 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/487cf0fc-4368-4f0e-878a-d3cd3d1e54f4-util\") pod \"50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05vd4cz\" (UID: \"487cf0fc-4368-4f0e-878a-d3cd3d1e54f4\") " pod="openstack-operators/50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05vd4cz" Dec 11 02:17:39 crc kubenswrapper[4824]: I1211 02:17:39.620087 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-26b6m\" (UniqueName: \"kubernetes.io/projected/487cf0fc-4368-4f0e-878a-d3cd3d1e54f4-kube-api-access-26b6m\") pod \"50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05vd4cz\" (UID: \"487cf0fc-4368-4f0e-878a-d3cd3d1e54f4\") " pod="openstack-operators/50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05vd4cz" Dec 11 02:17:39 crc kubenswrapper[4824]: I1211 02:17:39.655400 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05vd4cz" Dec 11 02:17:40 crc kubenswrapper[4824]: I1211 02:17:40.113856 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05vd4cz"] Dec 11 02:17:40 crc kubenswrapper[4824]: I1211 02:17:40.327081 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05vd4cz" event={"ID":"487cf0fc-4368-4f0e-878a-d3cd3d1e54f4","Type":"ContainerStarted","Data":"1342840a98157710e8b6a3be21f7b8605cd28ed27fe142b7d4076aaaad40f117"} Dec 11 02:17:41 crc kubenswrapper[4824]: I1211 02:17:41.339554 4824 generic.go:334] "Generic (PLEG): container finished" podID="487cf0fc-4368-4f0e-878a-d3cd3d1e54f4" containerID="5a527dbc62258e2c6b0c4f15b7ac5e81c964c22bffc55e381485e1b806cb0fbe" exitCode=0 Dec 11 02:17:41 crc kubenswrapper[4824]: I1211 02:17:41.340981 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05vd4cz" event={"ID":"487cf0fc-4368-4f0e-878a-d3cd3d1e54f4","Type":"ContainerDied","Data":"5a527dbc62258e2c6b0c4f15b7ac5e81c964c22bffc55e381485e1b806cb0fbe"} Dec 11 02:17:42 crc kubenswrapper[4824]: I1211 02:17:42.352647 4824 generic.go:334] "Generic (PLEG): container finished" podID="487cf0fc-4368-4f0e-878a-d3cd3d1e54f4" containerID="52b0a9a0c15b98cc091e05ee1c17f1e1fa97796e590e63a4a64052ec449ada26" exitCode=0 Dec 11 02:17:42 crc kubenswrapper[4824]: I1211 02:17:42.352759 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05vd4cz" event={"ID":"487cf0fc-4368-4f0e-878a-d3cd3d1e54f4","Type":"ContainerDied","Data":"52b0a9a0c15b98cc091e05ee1c17f1e1fa97796e590e63a4a64052ec449ada26"} Dec 11 02:17:43 crc kubenswrapper[4824]: I1211 02:17:43.364938 4824 generic.go:334] "Generic (PLEG): container finished" podID="487cf0fc-4368-4f0e-878a-d3cd3d1e54f4" containerID="a683bd77803c7352dd3878cbada4676a9396613548d767cc18bb646482c69fbb" exitCode=0 Dec 11 02:17:43 crc kubenswrapper[4824]: I1211 02:17:43.365088 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05vd4cz" event={"ID":"487cf0fc-4368-4f0e-878a-d3cd3d1e54f4","Type":"ContainerDied","Data":"a683bd77803c7352dd3878cbada4676a9396613548d767cc18bb646482c69fbb"} Dec 11 02:17:44 crc kubenswrapper[4824]: I1211 02:17:44.714753 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05vd4cz" Dec 11 02:17:44 crc kubenswrapper[4824]: I1211 02:17:44.880956 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/487cf0fc-4368-4f0e-878a-d3cd3d1e54f4-bundle\") pod \"487cf0fc-4368-4f0e-878a-d3cd3d1e54f4\" (UID: \"487cf0fc-4368-4f0e-878a-d3cd3d1e54f4\") " Dec 11 02:17:44 crc kubenswrapper[4824]: I1211 02:17:44.881040 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-26b6m\" (UniqueName: \"kubernetes.io/projected/487cf0fc-4368-4f0e-878a-d3cd3d1e54f4-kube-api-access-26b6m\") pod \"487cf0fc-4368-4f0e-878a-d3cd3d1e54f4\" (UID: \"487cf0fc-4368-4f0e-878a-d3cd3d1e54f4\") " Dec 11 02:17:44 crc kubenswrapper[4824]: I1211 02:17:44.881069 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/487cf0fc-4368-4f0e-878a-d3cd3d1e54f4-util\") pod \"487cf0fc-4368-4f0e-878a-d3cd3d1e54f4\" (UID: \"487cf0fc-4368-4f0e-878a-d3cd3d1e54f4\") " Dec 11 02:17:44 crc kubenswrapper[4824]: I1211 02:17:44.883694 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/487cf0fc-4368-4f0e-878a-d3cd3d1e54f4-bundle" (OuterVolumeSpecName: "bundle") pod "487cf0fc-4368-4f0e-878a-d3cd3d1e54f4" (UID: "487cf0fc-4368-4f0e-878a-d3cd3d1e54f4"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:17:44 crc kubenswrapper[4824]: I1211 02:17:44.894378 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/487cf0fc-4368-4f0e-878a-d3cd3d1e54f4-kube-api-access-26b6m" (OuterVolumeSpecName: "kube-api-access-26b6m") pod "487cf0fc-4368-4f0e-878a-d3cd3d1e54f4" (UID: "487cf0fc-4368-4f0e-878a-d3cd3d1e54f4"). InnerVolumeSpecName "kube-api-access-26b6m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:17:44 crc kubenswrapper[4824]: I1211 02:17:44.896550 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/487cf0fc-4368-4f0e-878a-d3cd3d1e54f4-util" (OuterVolumeSpecName: "util") pod "487cf0fc-4368-4f0e-878a-d3cd3d1e54f4" (UID: "487cf0fc-4368-4f0e-878a-d3cd3d1e54f4"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:17:44 crc kubenswrapper[4824]: I1211 02:17:44.982800 4824 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/487cf0fc-4368-4f0e-878a-d3cd3d1e54f4-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 02:17:44 crc kubenswrapper[4824]: I1211 02:17:44.982866 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-26b6m\" (UniqueName: \"kubernetes.io/projected/487cf0fc-4368-4f0e-878a-d3cd3d1e54f4-kube-api-access-26b6m\") on node \"crc\" DevicePath \"\"" Dec 11 02:17:44 crc kubenswrapper[4824]: I1211 02:17:44.982919 4824 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/487cf0fc-4368-4f0e-878a-d3cd3d1e54f4-util\") on node \"crc\" DevicePath \"\"" Dec 11 02:17:45 crc kubenswrapper[4824]: I1211 02:17:45.387102 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05vd4cz" event={"ID":"487cf0fc-4368-4f0e-878a-d3cd3d1e54f4","Type":"ContainerDied","Data":"1342840a98157710e8b6a3be21f7b8605cd28ed27fe142b7d4076aaaad40f117"} Dec 11 02:17:45 crc kubenswrapper[4824]: I1211 02:17:45.387207 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1342840a98157710e8b6a3be21f7b8605cd28ed27fe142b7d4076aaaad40f117" Dec 11 02:17:45 crc kubenswrapper[4824]: I1211 02:17:45.387221 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05vd4cz" Dec 11 02:17:47 crc kubenswrapper[4824]: I1211 02:17:47.395376 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-7795b79658-tzqvv"] Dec 11 02:17:47 crc kubenswrapper[4824]: E1211 02:17:47.396159 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="487cf0fc-4368-4f0e-878a-d3cd3d1e54f4" containerName="extract" Dec 11 02:17:47 crc kubenswrapper[4824]: I1211 02:17:47.396181 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="487cf0fc-4368-4f0e-878a-d3cd3d1e54f4" containerName="extract" Dec 11 02:17:47 crc kubenswrapper[4824]: E1211 02:17:47.396208 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="487cf0fc-4368-4f0e-878a-d3cd3d1e54f4" containerName="util" Dec 11 02:17:47 crc kubenswrapper[4824]: I1211 02:17:47.396222 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="487cf0fc-4368-4f0e-878a-d3cd3d1e54f4" containerName="util" Dec 11 02:17:47 crc kubenswrapper[4824]: E1211 02:17:47.396247 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="487cf0fc-4368-4f0e-878a-d3cd3d1e54f4" containerName="pull" Dec 11 02:17:47 crc kubenswrapper[4824]: I1211 02:17:47.396264 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="487cf0fc-4368-4f0e-878a-d3cd3d1e54f4" containerName="pull" Dec 11 02:17:47 crc kubenswrapper[4824]: I1211 02:17:47.396501 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="487cf0fc-4368-4f0e-878a-d3cd3d1e54f4" containerName="extract" Dec 11 02:17:47 crc kubenswrapper[4824]: I1211 02:17:47.397374 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-7795b79658-tzqvv" Dec 11 02:17:47 crc kubenswrapper[4824]: I1211 02:17:47.400089 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-tdl8p" Dec 11 02:17:47 crc kubenswrapper[4824]: I1211 02:17:47.420460 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-7795b79658-tzqvv"] Dec 11 02:17:47 crc kubenswrapper[4824]: I1211 02:17:47.557354 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nwwtk\" (UniqueName: \"kubernetes.io/projected/52218f50-e21e-4566-b323-9b5ece7a5e35-kube-api-access-nwwtk\") pod \"openstack-operator-controller-operator-7795b79658-tzqvv\" (UID: \"52218f50-e21e-4566-b323-9b5ece7a5e35\") " pod="openstack-operators/openstack-operator-controller-operator-7795b79658-tzqvv" Dec 11 02:17:47 crc kubenswrapper[4824]: I1211 02:17:47.659137 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nwwtk\" (UniqueName: \"kubernetes.io/projected/52218f50-e21e-4566-b323-9b5ece7a5e35-kube-api-access-nwwtk\") pod \"openstack-operator-controller-operator-7795b79658-tzqvv\" (UID: \"52218f50-e21e-4566-b323-9b5ece7a5e35\") " pod="openstack-operators/openstack-operator-controller-operator-7795b79658-tzqvv" Dec 11 02:17:47 crc kubenswrapper[4824]: I1211 02:17:47.685055 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nwwtk\" (UniqueName: \"kubernetes.io/projected/52218f50-e21e-4566-b323-9b5ece7a5e35-kube-api-access-nwwtk\") pod \"openstack-operator-controller-operator-7795b79658-tzqvv\" (UID: \"52218f50-e21e-4566-b323-9b5ece7a5e35\") " pod="openstack-operators/openstack-operator-controller-operator-7795b79658-tzqvv" Dec 11 02:17:47 crc kubenswrapper[4824]: I1211 02:17:47.714729 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-7795b79658-tzqvv" Dec 11 02:17:47 crc kubenswrapper[4824]: I1211 02:17:47.960609 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-7795b79658-tzqvv"] Dec 11 02:17:47 crc kubenswrapper[4824]: W1211 02:17:47.966225 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod52218f50_e21e_4566_b323_9b5ece7a5e35.slice/crio-80a4c7e1252a139500dccccf818f39b437217a3d993834620a9adcf74d376905 WatchSource:0}: Error finding container 80a4c7e1252a139500dccccf818f39b437217a3d993834620a9adcf74d376905: Status 404 returned error can't find the container with id 80a4c7e1252a139500dccccf818f39b437217a3d993834620a9adcf74d376905 Dec 11 02:17:48 crc kubenswrapper[4824]: I1211 02:17:48.409015 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-7795b79658-tzqvv" event={"ID":"52218f50-e21e-4566-b323-9b5ece7a5e35","Type":"ContainerStarted","Data":"80a4c7e1252a139500dccccf818f39b437217a3d993834620a9adcf74d376905"} Dec 11 02:17:52 crc kubenswrapper[4824]: I1211 02:17:52.436469 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-7795b79658-tzqvv" event={"ID":"52218f50-e21e-4566-b323-9b5ece7a5e35","Type":"ContainerStarted","Data":"ea97d39476ab0d05f1992b4e0aafd6dc9b171a3d7fc8fd3187275d791b67180a"} Dec 11 02:17:52 crc kubenswrapper[4824]: I1211 02:17:52.437075 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-7795b79658-tzqvv" Dec 11 02:17:52 crc kubenswrapper[4824]: I1211 02:17:52.476008 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-7795b79658-tzqvv" podStartSLOduration=1.3484112879999999 podStartE2EDuration="5.475993501s" podCreationTimestamp="2025-12-11 02:17:47 +0000 UTC" firstStartedPulling="2025-12-11 02:17:47.967652218 +0000 UTC m=+1009.656689597" lastFinishedPulling="2025-12-11 02:17:52.095234421 +0000 UTC m=+1013.784271810" observedRunningTime="2025-12-11 02:17:52.469479692 +0000 UTC m=+1014.158517081" watchObservedRunningTime="2025-12-11 02:17:52.475993501 +0000 UTC m=+1014.165030880" Dec 11 02:17:57 crc kubenswrapper[4824]: I1211 02:17:57.718941 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-7795b79658-tzqvv" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.425365 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-r8tnj"] Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.427840 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-r8tnj" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.430182 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-2z5qw" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.433079 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6c677c69b-8hvxh"] Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.435814 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-8hvxh" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.438642 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-697fb699cf-4b2cb"] Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.440141 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-vrnk5" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.440342 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-4b2cb" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.441770 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-cqnq6" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.455307 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6c677c69b-8hvxh"] Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.463770 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-r8tnj"] Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.467986 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-5697bb5779-4cjld"] Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.469343 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-4cjld" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.472423 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-chgdl" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.479868 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-697fb699cf-4b2cb"] Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.485732 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-5697bb5779-4cjld"] Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.489597 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-l9jrd"] Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.490575 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-l9jrd" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.492545 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-xt86w" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.493012 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-5qp6v"] Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.493886 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-5qp6v" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.496299 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-6kgvf" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.510341 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-l9jrd"] Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.517149 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-78d48bff9d-j68mf"] Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.518125 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-j68mf" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.521883 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.522138 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-dxbpz" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.523042 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f219b9bf-34b1-48c4-96c8-20580b6ec9a4-cert\") pod \"infra-operator-controller-manager-78d48bff9d-j68mf\" (UID: \"f219b9bf-34b1-48c4-96c8-20580b6ec9a4\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-j68mf" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.523075 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s48sm\" (UniqueName: \"kubernetes.io/projected/df6864cf-9559-4398-bb3c-0f58f4c45563-kube-api-access-s48sm\") pod \"barbican-operator-controller-manager-7d9dfd778-r8tnj\" (UID: \"df6864cf-9559-4398-bb3c-0f58f4c45563\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-r8tnj" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.523102 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrj7p\" (UniqueName: \"kubernetes.io/projected/0b7b142b-04e2-4bb3-98ff-ed3b4da63746-kube-api-access-lrj7p\") pod \"cinder-operator-controller-manager-6c677c69b-8hvxh\" (UID: \"0b7b142b-04e2-4bb3-98ff-ed3b4da63746\") " pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-8hvxh" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.523133 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5s2w2\" (UniqueName: \"kubernetes.io/projected/0eeea3b6-c514-4547-a2e0-52870ef572c6-kube-api-access-5s2w2\") pod \"glance-operator-controller-manager-5697bb5779-4cjld\" (UID: \"0eeea3b6-c514-4547-a2e0-52870ef572c6\") " pod="openstack-operators/glance-operator-controller-manager-5697bb5779-4cjld" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.523216 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lkkgn\" (UniqueName: \"kubernetes.io/projected/1fc0b6e3-15c9-4f14-90f8-3f39a04ebc2c-kube-api-access-lkkgn\") pod \"horizon-operator-controller-manager-68c6d99b8f-5qp6v\" (UID: \"1fc0b6e3-15c9-4f14-90f8-3f39a04ebc2c\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-5qp6v" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.523255 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2vqd\" (UniqueName: \"kubernetes.io/projected/6743e907-9570-4dde-bb99-83e67202b224-kube-api-access-x2vqd\") pod \"designate-operator-controller-manager-697fb699cf-4b2cb\" (UID: \"6743e907-9570-4dde-bb99-83e67202b224\") " pod="openstack-operators/designate-operator-controller-manager-697fb699cf-4b2cb" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.523279 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sjfvc\" (UniqueName: \"kubernetes.io/projected/db4453b9-ead1-45fc-91c5-d90541fa9c78-kube-api-access-sjfvc\") pod \"heat-operator-controller-manager-5f64f6f8bb-l9jrd\" (UID: \"db4453b9-ead1-45fc-91c5-d90541fa9c78\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-l9jrd" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.523310 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6krgg\" (UniqueName: \"kubernetes.io/projected/f219b9bf-34b1-48c4-96c8-20580b6ec9a4-kube-api-access-6krgg\") pod \"infra-operator-controller-manager-78d48bff9d-j68mf\" (UID: \"f219b9bf-34b1-48c4-96c8-20580b6ec9a4\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-j68mf" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.528360 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-5qp6v"] Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.548449 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-78d48bff9d-j68mf"] Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.560553 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-967d97867-5tpj5"] Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.561665 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-967d97867-5tpj5" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.567248 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-d5nb8" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.586764 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-4np4x"] Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.588393 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-4np4x" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.592077 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-l7txh" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.605397 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-967d97867-5tpj5"] Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.624424 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5fvm2\" (UniqueName: \"kubernetes.io/projected/cec1f29f-a1db-495f-b59c-e9308a9b53a2-kube-api-access-5fvm2\") pod \"keystone-operator-controller-manager-7765d96ddf-4np4x\" (UID: \"cec1f29f-a1db-495f-b59c-e9308a9b53a2\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-4np4x" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.624500 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sjfvc\" (UniqueName: \"kubernetes.io/projected/db4453b9-ead1-45fc-91c5-d90541fa9c78-kube-api-access-sjfvc\") pod \"heat-operator-controller-manager-5f64f6f8bb-l9jrd\" (UID: \"db4453b9-ead1-45fc-91c5-d90541fa9c78\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-l9jrd" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.624546 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x2vqd\" (UniqueName: \"kubernetes.io/projected/6743e907-9570-4dde-bb99-83e67202b224-kube-api-access-x2vqd\") pod \"designate-operator-controller-manager-697fb699cf-4b2cb\" (UID: \"6743e907-9570-4dde-bb99-83e67202b224\") " pod="openstack-operators/designate-operator-controller-manager-697fb699cf-4b2cb" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.624605 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6krgg\" (UniqueName: \"kubernetes.io/projected/f219b9bf-34b1-48c4-96c8-20580b6ec9a4-kube-api-access-6krgg\") pod \"infra-operator-controller-manager-78d48bff9d-j68mf\" (UID: \"f219b9bf-34b1-48c4-96c8-20580b6ec9a4\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-j68mf" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.624633 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f219b9bf-34b1-48c4-96c8-20580b6ec9a4-cert\") pod \"infra-operator-controller-manager-78d48bff9d-j68mf\" (UID: \"f219b9bf-34b1-48c4-96c8-20580b6ec9a4\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-j68mf" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.624661 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7tnsf\" (UniqueName: \"kubernetes.io/projected/aea47aaf-ee88-45a0-87b7-375b9f656828-kube-api-access-7tnsf\") pod \"ironic-operator-controller-manager-967d97867-5tpj5\" (UID: \"aea47aaf-ee88-45a0-87b7-375b9f656828\") " pod="openstack-operators/ironic-operator-controller-manager-967d97867-5tpj5" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.624712 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s48sm\" (UniqueName: \"kubernetes.io/projected/df6864cf-9559-4398-bb3c-0f58f4c45563-kube-api-access-s48sm\") pod \"barbican-operator-controller-manager-7d9dfd778-r8tnj\" (UID: \"df6864cf-9559-4398-bb3c-0f58f4c45563\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-r8tnj" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.624766 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrj7p\" (UniqueName: \"kubernetes.io/projected/0b7b142b-04e2-4bb3-98ff-ed3b4da63746-kube-api-access-lrj7p\") pod \"cinder-operator-controller-manager-6c677c69b-8hvxh\" (UID: \"0b7b142b-04e2-4bb3-98ff-ed3b4da63746\") " pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-8hvxh" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.624791 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5s2w2\" (UniqueName: \"kubernetes.io/projected/0eeea3b6-c514-4547-a2e0-52870ef572c6-kube-api-access-5s2w2\") pod \"glance-operator-controller-manager-5697bb5779-4cjld\" (UID: \"0eeea3b6-c514-4547-a2e0-52870ef572c6\") " pod="openstack-operators/glance-operator-controller-manager-5697bb5779-4cjld" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.624863 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lkkgn\" (UniqueName: \"kubernetes.io/projected/1fc0b6e3-15c9-4f14-90f8-3f39a04ebc2c-kube-api-access-lkkgn\") pod \"horizon-operator-controller-manager-68c6d99b8f-5qp6v\" (UID: \"1fc0b6e3-15c9-4f14-90f8-3f39a04ebc2c\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-5qp6v" Dec 11 02:18:34 crc kubenswrapper[4824]: E1211 02:18:34.625811 4824 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 11 02:18:34 crc kubenswrapper[4824]: E1211 02:18:34.625884 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f219b9bf-34b1-48c4-96c8-20580b6ec9a4-cert podName:f219b9bf-34b1-48c4-96c8-20580b6ec9a4 nodeName:}" failed. No retries permitted until 2025-12-11 02:18:35.125866326 +0000 UTC m=+1056.814903705 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f219b9bf-34b1-48c4-96c8-20580b6ec9a4-cert") pod "infra-operator-controller-manager-78d48bff9d-j68mf" (UID: "f219b9bf-34b1-48c4-96c8-20580b6ec9a4") : secret "infra-operator-webhook-server-cert" not found Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.628272 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-5b5fd79c9c-w24mx"] Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.629284 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-w24mx" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.632229 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-4np4x"] Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.635449 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-b79kh" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.655912 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-5b5fd79c9c-w24mx"] Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.670537 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lkkgn\" (UniqueName: \"kubernetes.io/projected/1fc0b6e3-15c9-4f14-90f8-3f39a04ebc2c-kube-api-access-lkkgn\") pod \"horizon-operator-controller-manager-68c6d99b8f-5qp6v\" (UID: \"1fc0b6e3-15c9-4f14-90f8-3f39a04ebc2c\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-5qp6v" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.676619 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6krgg\" (UniqueName: \"kubernetes.io/projected/f219b9bf-34b1-48c4-96c8-20580b6ec9a4-kube-api-access-6krgg\") pod \"infra-operator-controller-manager-78d48bff9d-j68mf\" (UID: \"f219b9bf-34b1-48c4-96c8-20580b6ec9a4\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-j68mf" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.676708 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-79c8c4686c-hpjng"] Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.677424 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s48sm\" (UniqueName: \"kubernetes.io/projected/df6864cf-9559-4398-bb3c-0f58f4c45563-kube-api-access-s48sm\") pod \"barbican-operator-controller-manager-7d9dfd778-r8tnj\" (UID: \"df6864cf-9559-4398-bb3c-0f58f4c45563\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-r8tnj" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.680478 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-hpjng" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.682895 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-rj5xw" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.688043 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5s2w2\" (UniqueName: \"kubernetes.io/projected/0eeea3b6-c514-4547-a2e0-52870ef572c6-kube-api-access-5s2w2\") pod \"glance-operator-controller-manager-5697bb5779-4cjld\" (UID: \"0eeea3b6-c514-4547-a2e0-52870ef572c6\") " pod="openstack-operators/glance-operator-controller-manager-5697bb5779-4cjld" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.699863 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x2vqd\" (UniqueName: \"kubernetes.io/projected/6743e907-9570-4dde-bb99-83e67202b224-kube-api-access-x2vqd\") pod \"designate-operator-controller-manager-697fb699cf-4b2cb\" (UID: \"6743e907-9570-4dde-bb99-83e67202b224\") " pod="openstack-operators/designate-operator-controller-manager-697fb699cf-4b2cb" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.714262 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrj7p\" (UniqueName: \"kubernetes.io/projected/0b7b142b-04e2-4bb3-98ff-ed3b4da63746-kube-api-access-lrj7p\") pod \"cinder-operator-controller-manager-6c677c69b-8hvxh\" (UID: \"0b7b142b-04e2-4bb3-98ff-ed3b4da63746\") " pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-8hvxh" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.717830 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sjfvc\" (UniqueName: \"kubernetes.io/projected/db4453b9-ead1-45fc-91c5-d90541fa9c78-kube-api-access-sjfvc\") pod \"heat-operator-controller-manager-5f64f6f8bb-l9jrd\" (UID: \"db4453b9-ead1-45fc-91c5-d90541fa9c78\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-l9jrd" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.726230 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5fvm2\" (UniqueName: \"kubernetes.io/projected/cec1f29f-a1db-495f-b59c-e9308a9b53a2-kube-api-access-5fvm2\") pod \"keystone-operator-controller-manager-7765d96ddf-4np4x\" (UID: \"cec1f29f-a1db-495f-b59c-e9308a9b53a2\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-4np4x" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.726334 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7tnsf\" (UniqueName: \"kubernetes.io/projected/aea47aaf-ee88-45a0-87b7-375b9f656828-kube-api-access-7tnsf\") pod \"ironic-operator-controller-manager-967d97867-5tpj5\" (UID: \"aea47aaf-ee88-45a0-87b7-375b9f656828\") " pod="openstack-operators/ironic-operator-controller-manager-967d97867-5tpj5" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.732913 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-79c8c4686c-hpjng"] Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.742476 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5fvm2\" (UniqueName: \"kubernetes.io/projected/cec1f29f-a1db-495f-b59c-e9308a9b53a2-kube-api-access-5fvm2\") pod \"keystone-operator-controller-manager-7765d96ddf-4np4x\" (UID: \"cec1f29f-a1db-495f-b59c-e9308a9b53a2\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-4np4x" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.744828 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7tnsf\" (UniqueName: \"kubernetes.io/projected/aea47aaf-ee88-45a0-87b7-375b9f656828-kube-api-access-7tnsf\") pod \"ironic-operator-controller-manager-967d97867-5tpj5\" (UID: \"aea47aaf-ee88-45a0-87b7-375b9f656828\") " pod="openstack-operators/ironic-operator-controller-manager-967d97867-5tpj5" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.748296 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-lmszx"] Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.751916 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-lmszx" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.754544 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-qrkzg" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.763241 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-r8tnj" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.768473 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-zkz72"] Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.769547 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-zkz72" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.771721 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-52bkb" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.774963 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-lmszx"] Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.781219 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-8hvxh" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.789911 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-zkz72"] Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.792291 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-4b2cb" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.804496 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-4cjld" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.820983 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-l9jrd" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.824460 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-qqpf2"] Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.825924 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-qqpf2" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.827081 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6mpbl\" (UniqueName: \"kubernetes.io/projected/46bf5c63-42b4-45bd-ade3-446a4b2e4f6e-kube-api-access-6mpbl\") pod \"manila-operator-controller-manager-5b5fd79c9c-w24mx\" (UID: \"46bf5c63-42b4-45bd-ade3-446a4b2e4f6e\") " pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-w24mx" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.827153 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdppb\" (UniqueName: \"kubernetes.io/projected/fd3368a6-77e2-4c8f-b535-1d602b794d7d-kube-api-access-hdppb\") pod \"mariadb-operator-controller-manager-79c8c4686c-hpjng\" (UID: \"fd3368a6-77e2-4c8f-b535-1d602b794d7d\") " pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-hpjng" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.831412 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-4ktjw" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.834396 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-5qp6v" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.842011 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-qqpf2"] Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.846375 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f4k8rk"] Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.847664 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f4k8rk" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.849463 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-vdk5t" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.849639 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.852235 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-r2bsg"] Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.853440 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-r2bsg" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.855710 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-k64td" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.858239 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-xx9gc"] Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.859505 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-xx9gc" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.861251 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-v9c2c" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.865545 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-r2bsg"] Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.880695 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-967d97867-5tpj5" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.893912 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-xx9gc"] Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.905461 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f4k8rk"] Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.908966 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-4np4x" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.913078 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-9d58d64bc-fmvm7"] Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.914256 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-fmvm7" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.922227 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-29v77" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.927199 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-9d58d64bc-fmvm7"] Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.927909 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z65ct\" (UniqueName: \"kubernetes.io/projected/dca85063-ba7a-45ac-ab88-5498a44834ba-kube-api-access-z65ct\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-lmszx\" (UID: \"dca85063-ba7a-45ac-ab88-5498a44834ba\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-lmszx" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.927956 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sgb9f\" (UniqueName: \"kubernetes.io/projected/f1183bb7-356d-43b7-8ee7-55ab60dcbe54-kube-api-access-sgb9f\") pod \"octavia-operator-controller-manager-998648c74-qqpf2\" (UID: \"f1183bb7-356d-43b7-8ee7-55ab60dcbe54\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-qqpf2" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.928094 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6mpbl\" (UniqueName: \"kubernetes.io/projected/46bf5c63-42b4-45bd-ade3-446a4b2e4f6e-kube-api-access-6mpbl\") pod \"manila-operator-controller-manager-5b5fd79c9c-w24mx\" (UID: \"46bf5c63-42b4-45bd-ade3-446a4b2e4f6e\") " pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-w24mx" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.928147 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdppb\" (UniqueName: \"kubernetes.io/projected/fd3368a6-77e2-4c8f-b535-1d602b794d7d-kube-api-access-hdppb\") pod \"mariadb-operator-controller-manager-79c8c4686c-hpjng\" (UID: \"fd3368a6-77e2-4c8f-b535-1d602b794d7d\") " pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-hpjng" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.928168 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4kcrt\" (UniqueName: \"kubernetes.io/projected/b206f71d-32cd-4b7d-ae38-a5125ab481dd-kube-api-access-4kcrt\") pod \"nova-operator-controller-manager-697bc559fc-zkz72\" (UID: \"b206f71d-32cd-4b7d-ae38-a5125ab481dd\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-zkz72" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.949193 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-58d5ff84df-qhhzn"] Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.952227 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdppb\" (UniqueName: \"kubernetes.io/projected/fd3368a6-77e2-4c8f-b535-1d602b794d7d-kube-api-access-hdppb\") pod \"mariadb-operator-controller-manager-79c8c4686c-hpjng\" (UID: \"fd3368a6-77e2-4c8f-b535-1d602b794d7d\") " pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-hpjng" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.956875 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6mpbl\" (UniqueName: \"kubernetes.io/projected/46bf5c63-42b4-45bd-ade3-446a4b2e4f6e-kube-api-access-6mpbl\") pod \"manila-operator-controller-manager-5b5fd79c9c-w24mx\" (UID: \"46bf5c63-42b4-45bd-ade3-446a4b2e4f6e\") " pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-w24mx" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.959308 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-58d5ff84df-qhhzn"] Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.959412 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-qhhzn" Dec 11 02:18:34 crc kubenswrapper[4824]: I1211 02:18:34.961593 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-9ckfm" Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.014345 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-4zfbk"] Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.015696 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-4zfbk" Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.017494 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-pdsph" Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.029286 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zrm4p\" (UniqueName: \"kubernetes.io/projected/8047af9c-3a6d-4b2e-a088-df967e42671c-kube-api-access-zrm4p\") pod \"swift-operator-controller-manager-9d58d64bc-fmvm7\" (UID: \"8047af9c-3a6d-4b2e-a088-df967e42671c\") " pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-fmvm7" Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.029341 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zvhsp\" (UniqueName: \"kubernetes.io/projected/10325d55-7efd-49e8-95e9-46d68c7cd7ae-kube-api-access-zvhsp\") pod \"placement-operator-controller-manager-78f8948974-xx9gc\" (UID: \"10325d55-7efd-49e8-95e9-46d68c7cd7ae\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-xx9gc" Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.029369 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z65ct\" (UniqueName: \"kubernetes.io/projected/dca85063-ba7a-45ac-ab88-5498a44834ba-kube-api-access-z65ct\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-lmszx\" (UID: \"dca85063-ba7a-45ac-ab88-5498a44834ba\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-lmszx" Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.029393 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sgb9f\" (UniqueName: \"kubernetes.io/projected/f1183bb7-356d-43b7-8ee7-55ab60dcbe54-kube-api-access-sgb9f\") pod \"octavia-operator-controller-manager-998648c74-qqpf2\" (UID: \"f1183bb7-356d-43b7-8ee7-55ab60dcbe54\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-qqpf2" Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.029428 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l6xnd\" (UniqueName: \"kubernetes.io/projected/4d26d90a-05a8-4960-b2ca-12fa220383b7-kube-api-access-l6xnd\") pod \"ovn-operator-controller-manager-b6456fdb6-r2bsg\" (UID: \"4d26d90a-05a8-4960-b2ca-12fa220383b7\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-r2bsg" Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.029446 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/cb370fee-74a4-4f43-98a9-3df8c6e61335-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879f4k8rk\" (UID: \"cb370fee-74a4-4f43-98a9-3df8c6e61335\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f4k8rk" Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.029761 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-4zfbk"] Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.029516 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4kcrt\" (UniqueName: \"kubernetes.io/projected/b206f71d-32cd-4b7d-ae38-a5125ab481dd-kube-api-access-4kcrt\") pod \"nova-operator-controller-manager-697bc559fc-zkz72\" (UID: \"b206f71d-32cd-4b7d-ae38-a5125ab481dd\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-zkz72" Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.029924 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vdg5c\" (UniqueName: \"kubernetes.io/projected/cb370fee-74a4-4f43-98a9-3df8c6e61335-kube-api-access-vdg5c\") pod \"openstack-baremetal-operator-controller-manager-84b575879f4k8rk\" (UID: \"cb370fee-74a4-4f43-98a9-3df8c6e61335\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f4k8rk" Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.050993 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z65ct\" (UniqueName: \"kubernetes.io/projected/dca85063-ba7a-45ac-ab88-5498a44834ba-kube-api-access-z65ct\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-lmszx\" (UID: \"dca85063-ba7a-45ac-ab88-5498a44834ba\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-lmszx" Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.050996 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4kcrt\" (UniqueName: \"kubernetes.io/projected/b206f71d-32cd-4b7d-ae38-a5125ab481dd-kube-api-access-4kcrt\") pod \"nova-operator-controller-manager-697bc559fc-zkz72\" (UID: \"b206f71d-32cd-4b7d-ae38-a5125ab481dd\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-zkz72" Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.051515 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sgb9f\" (UniqueName: \"kubernetes.io/projected/f1183bb7-356d-43b7-8ee7-55ab60dcbe54-kube-api-access-sgb9f\") pod \"octavia-operator-controller-manager-998648c74-qqpf2\" (UID: \"f1183bb7-356d-43b7-8ee7-55ab60dcbe54\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-qqpf2" Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.064833 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-75944c9b7-h6bgf"] Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.067024 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-h6bgf" Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.068363 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-w24mx" Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.068598 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-rrw2d" Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.083442 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-75944c9b7-h6bgf"] Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.091153 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-hpjng" Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.119562 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-lmszx" Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.131649 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ljct\" (UniqueName: \"kubernetes.io/projected/aff71eda-47bc-4b47-89df-fe9bf8357fba-kube-api-access-8ljct\") pod \"telemetry-operator-controller-manager-58d5ff84df-qhhzn\" (UID: \"aff71eda-47bc-4b47-89df-fe9bf8357fba\") " pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-qhhzn" Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.131717 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zrm4p\" (UniqueName: \"kubernetes.io/projected/8047af9c-3a6d-4b2e-a088-df967e42671c-kube-api-access-zrm4p\") pod \"swift-operator-controller-manager-9d58d64bc-fmvm7\" (UID: \"8047af9c-3a6d-4b2e-a088-df967e42671c\") " pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-fmvm7" Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.131750 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zvhsp\" (UniqueName: \"kubernetes.io/projected/10325d55-7efd-49e8-95e9-46d68c7cd7ae-kube-api-access-zvhsp\") pod \"placement-operator-controller-manager-78f8948974-xx9gc\" (UID: \"10325d55-7efd-49e8-95e9-46d68c7cd7ae\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-xx9gc" Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.131802 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l6xnd\" (UniqueName: \"kubernetes.io/projected/4d26d90a-05a8-4960-b2ca-12fa220383b7-kube-api-access-l6xnd\") pod \"ovn-operator-controller-manager-b6456fdb6-r2bsg\" (UID: \"4d26d90a-05a8-4960-b2ca-12fa220383b7\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-r2bsg" Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.131825 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9shh8\" (UniqueName: \"kubernetes.io/projected/cf19e8a0-8f29-48a4-b715-f499644fc896-kube-api-access-9shh8\") pod \"test-operator-controller-manager-5854674fcc-4zfbk\" (UID: \"cf19e8a0-8f29-48a4-b715-f499644fc896\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-4zfbk" Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.131848 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/cb370fee-74a4-4f43-98a9-3df8c6e61335-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879f4k8rk\" (UID: \"cb370fee-74a4-4f43-98a9-3df8c6e61335\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f4k8rk" Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.131879 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f219b9bf-34b1-48c4-96c8-20580b6ec9a4-cert\") pod \"infra-operator-controller-manager-78d48bff9d-j68mf\" (UID: \"f219b9bf-34b1-48c4-96c8-20580b6ec9a4\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-j68mf" Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.131918 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vdg5c\" (UniqueName: \"kubernetes.io/projected/cb370fee-74a4-4f43-98a9-3df8c6e61335-kube-api-access-vdg5c\") pod \"openstack-baremetal-operator-controller-manager-84b575879f4k8rk\" (UID: \"cb370fee-74a4-4f43-98a9-3df8c6e61335\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f4k8rk" Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.131660 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-zkz72" Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.133411 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-56cf947496-nhjjr"] Dec 11 02:18:35 crc kubenswrapper[4824]: E1211 02:18:35.143069 4824 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 11 02:18:35 crc kubenswrapper[4824]: E1211 02:18:35.143171 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cb370fee-74a4-4f43-98a9-3df8c6e61335-cert podName:cb370fee-74a4-4f43-98a9-3df8c6e61335 nodeName:}" failed. No retries permitted until 2025-12-11 02:18:35.643147713 +0000 UTC m=+1057.332185092 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/cb370fee-74a4-4f43-98a9-3df8c6e61335-cert") pod "openstack-baremetal-operator-controller-manager-84b575879f4k8rk" (UID: "cb370fee-74a4-4f43-98a9-3df8c6e61335") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.143445 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-qqpf2" Dec 11 02:18:35 crc kubenswrapper[4824]: E1211 02:18:35.143619 4824 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 11 02:18:35 crc kubenswrapper[4824]: E1211 02:18:35.143718 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f219b9bf-34b1-48c4-96c8-20580b6ec9a4-cert podName:f219b9bf-34b1-48c4-96c8-20580b6ec9a4 nodeName:}" failed. No retries permitted until 2025-12-11 02:18:36.143696907 +0000 UTC m=+1057.832734286 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f219b9bf-34b1-48c4-96c8-20580b6ec9a4-cert") pod "infra-operator-controller-manager-78d48bff9d-j68mf" (UID: "f219b9bf-34b1-48c4-96c8-20580b6ec9a4") : secret "infra-operator-webhook-server-cert" not found Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.156984 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-56cf947496-nhjjr" Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.182264 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-hqsm4" Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.182454 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.183186 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.186689 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-56cf947496-nhjjr"] Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.227010 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vdg5c\" (UniqueName: \"kubernetes.io/projected/cb370fee-74a4-4f43-98a9-3df8c6e61335-kube-api-access-vdg5c\") pod \"openstack-baremetal-operator-controller-manager-84b575879f4k8rk\" (UID: \"cb370fee-74a4-4f43-98a9-3df8c6e61335\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f4k8rk" Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.230401 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zvhsp\" (UniqueName: \"kubernetes.io/projected/10325d55-7efd-49e8-95e9-46d68c7cd7ae-kube-api-access-zvhsp\") pod \"placement-operator-controller-manager-78f8948974-xx9gc\" (UID: \"10325d55-7efd-49e8-95e9-46d68c7cd7ae\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-xx9gc" Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.233000 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zrm4p\" (UniqueName: \"kubernetes.io/projected/8047af9c-3a6d-4b2e-a088-df967e42671c-kube-api-access-zrm4p\") pod \"swift-operator-controller-manager-9d58d64bc-fmvm7\" (UID: \"8047af9c-3a6d-4b2e-a088-df967e42671c\") " pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-fmvm7" Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.233252 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dnwhx\" (UniqueName: \"kubernetes.io/projected/9793f7fe-2e1b-46c0-985e-1f77733f925f-kube-api-access-dnwhx\") pod \"watcher-operator-controller-manager-75944c9b7-h6bgf\" (UID: \"9793f7fe-2e1b-46c0-985e-1f77733f925f\") " pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-h6bgf" Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.233380 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9shh8\" (UniqueName: \"kubernetes.io/projected/cf19e8a0-8f29-48a4-b715-f499644fc896-kube-api-access-9shh8\") pod \"test-operator-controller-manager-5854674fcc-4zfbk\" (UID: \"cf19e8a0-8f29-48a4-b715-f499644fc896\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-4zfbk" Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.233478 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ljct\" (UniqueName: \"kubernetes.io/projected/aff71eda-47bc-4b47-89df-fe9bf8357fba-kube-api-access-8ljct\") pod \"telemetry-operator-controller-manager-58d5ff84df-qhhzn\" (UID: \"aff71eda-47bc-4b47-89df-fe9bf8357fba\") " pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-qhhzn" Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.237738 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-wsjbh"] Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.238582 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-wsjbh" Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.242083 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l6xnd\" (UniqueName: \"kubernetes.io/projected/4d26d90a-05a8-4960-b2ca-12fa220383b7-kube-api-access-l6xnd\") pod \"ovn-operator-controller-manager-b6456fdb6-r2bsg\" (UID: \"4d26d90a-05a8-4960-b2ca-12fa220383b7\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-r2bsg" Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.248423 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-8hfnv" Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.248479 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-fmvm7" Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.270261 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-wsjbh"] Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.306648 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ljct\" (UniqueName: \"kubernetes.io/projected/aff71eda-47bc-4b47-89df-fe9bf8357fba-kube-api-access-8ljct\") pod \"telemetry-operator-controller-manager-58d5ff84df-qhhzn\" (UID: \"aff71eda-47bc-4b47-89df-fe9bf8357fba\") " pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-qhhzn" Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.306950 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9shh8\" (UniqueName: \"kubernetes.io/projected/cf19e8a0-8f29-48a4-b715-f499644fc896-kube-api-access-9shh8\") pod \"test-operator-controller-manager-5854674fcc-4zfbk\" (UID: \"cf19e8a0-8f29-48a4-b715-f499644fc896\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-4zfbk" Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.335304 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7180c14a-0c3a-4267-b122-5c30c5685d60-metrics-certs\") pod \"openstack-operator-controller-manager-56cf947496-nhjjr\" (UID: \"7180c14a-0c3a-4267-b122-5c30c5685d60\") " pod="openstack-operators/openstack-operator-controller-manager-56cf947496-nhjjr" Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.335353 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hwv5j\" (UniqueName: \"kubernetes.io/projected/34bc4184-9cfb-4278-a9b9-80bf2f1ae612-kube-api-access-hwv5j\") pod \"rabbitmq-cluster-operator-manager-668c99d594-wsjbh\" (UID: \"34bc4184-9cfb-4278-a9b9-80bf2f1ae612\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-wsjbh" Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.335379 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dnwhx\" (UniqueName: \"kubernetes.io/projected/9793f7fe-2e1b-46c0-985e-1f77733f925f-kube-api-access-dnwhx\") pod \"watcher-operator-controller-manager-75944c9b7-h6bgf\" (UID: \"9793f7fe-2e1b-46c0-985e-1f77733f925f\") " pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-h6bgf" Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.335408 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bvdl9\" (UniqueName: \"kubernetes.io/projected/7180c14a-0c3a-4267-b122-5c30c5685d60-kube-api-access-bvdl9\") pod \"openstack-operator-controller-manager-56cf947496-nhjjr\" (UID: \"7180c14a-0c3a-4267-b122-5c30c5685d60\") " pod="openstack-operators/openstack-operator-controller-manager-56cf947496-nhjjr" Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.335449 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/7180c14a-0c3a-4267-b122-5c30c5685d60-webhook-certs\") pod \"openstack-operator-controller-manager-56cf947496-nhjjr\" (UID: \"7180c14a-0c3a-4267-b122-5c30c5685d60\") " pod="openstack-operators/openstack-operator-controller-manager-56cf947496-nhjjr" Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.356567 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dnwhx\" (UniqueName: \"kubernetes.io/projected/9793f7fe-2e1b-46c0-985e-1f77733f925f-kube-api-access-dnwhx\") pod \"watcher-operator-controller-manager-75944c9b7-h6bgf\" (UID: \"9793f7fe-2e1b-46c0-985e-1f77733f925f\") " pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-h6bgf" Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.367484 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-4zfbk" Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.436993 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7180c14a-0c3a-4267-b122-5c30c5685d60-metrics-certs\") pod \"openstack-operator-controller-manager-56cf947496-nhjjr\" (UID: \"7180c14a-0c3a-4267-b122-5c30c5685d60\") " pod="openstack-operators/openstack-operator-controller-manager-56cf947496-nhjjr" Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.437264 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hwv5j\" (UniqueName: \"kubernetes.io/projected/34bc4184-9cfb-4278-a9b9-80bf2f1ae612-kube-api-access-hwv5j\") pod \"rabbitmq-cluster-operator-manager-668c99d594-wsjbh\" (UID: \"34bc4184-9cfb-4278-a9b9-80bf2f1ae612\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-wsjbh" Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.437304 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bvdl9\" (UniqueName: \"kubernetes.io/projected/7180c14a-0c3a-4267-b122-5c30c5685d60-kube-api-access-bvdl9\") pod \"openstack-operator-controller-manager-56cf947496-nhjjr\" (UID: \"7180c14a-0c3a-4267-b122-5c30c5685d60\") " pod="openstack-operators/openstack-operator-controller-manager-56cf947496-nhjjr" Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.437334 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/7180c14a-0c3a-4267-b122-5c30c5685d60-webhook-certs\") pod \"openstack-operator-controller-manager-56cf947496-nhjjr\" (UID: \"7180c14a-0c3a-4267-b122-5c30c5685d60\") " pod="openstack-operators/openstack-operator-controller-manager-56cf947496-nhjjr" Dec 11 02:18:35 crc kubenswrapper[4824]: E1211 02:18:35.437465 4824 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 11 02:18:35 crc kubenswrapper[4824]: E1211 02:18:35.437510 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7180c14a-0c3a-4267-b122-5c30c5685d60-webhook-certs podName:7180c14a-0c3a-4267-b122-5c30c5685d60 nodeName:}" failed. No retries permitted until 2025-12-11 02:18:35.937495562 +0000 UTC m=+1057.626532941 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/7180c14a-0c3a-4267-b122-5c30c5685d60-webhook-certs") pod "openstack-operator-controller-manager-56cf947496-nhjjr" (UID: "7180c14a-0c3a-4267-b122-5c30c5685d60") : secret "webhook-server-cert" not found Dec 11 02:18:35 crc kubenswrapper[4824]: E1211 02:18:35.437744 4824 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 11 02:18:35 crc kubenswrapper[4824]: E1211 02:18:35.437765 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7180c14a-0c3a-4267-b122-5c30c5685d60-metrics-certs podName:7180c14a-0c3a-4267-b122-5c30c5685d60 nodeName:}" failed. No retries permitted until 2025-12-11 02:18:35.937758568 +0000 UTC m=+1057.626795937 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7180c14a-0c3a-4267-b122-5c30c5685d60-metrics-certs") pod "openstack-operator-controller-manager-56cf947496-nhjjr" (UID: "7180c14a-0c3a-4267-b122-5c30c5685d60") : secret "metrics-server-cert" not found Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.459496 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bvdl9\" (UniqueName: \"kubernetes.io/projected/7180c14a-0c3a-4267-b122-5c30c5685d60-kube-api-access-bvdl9\") pod \"openstack-operator-controller-manager-56cf947496-nhjjr\" (UID: \"7180c14a-0c3a-4267-b122-5c30c5685d60\") " pod="openstack-operators/openstack-operator-controller-manager-56cf947496-nhjjr" Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.461305 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hwv5j\" (UniqueName: \"kubernetes.io/projected/34bc4184-9cfb-4278-a9b9-80bf2f1ae612-kube-api-access-hwv5j\") pod \"rabbitmq-cluster-operator-manager-668c99d594-wsjbh\" (UID: \"34bc4184-9cfb-4278-a9b9-80bf2f1ae612\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-wsjbh" Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.475953 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-r2bsg" Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.508275 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-xx9gc" Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.584421 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-qhhzn" Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.609380 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-h6bgf" Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.688550 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-wsjbh" Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.699890 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-r8tnj"] Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.704904 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6c677c69b-8hvxh"] Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.724384 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-697fb699cf-4b2cb"] Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.735564 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-4np4x"] Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.741015 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/cb370fee-74a4-4f43-98a9-3df8c6e61335-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879f4k8rk\" (UID: \"cb370fee-74a4-4f43-98a9-3df8c6e61335\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f4k8rk" Dec 11 02:18:35 crc kubenswrapper[4824]: E1211 02:18:35.742091 4824 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 11 02:18:35 crc kubenswrapper[4824]: E1211 02:18:35.742155 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cb370fee-74a4-4f43-98a9-3df8c6e61335-cert podName:cb370fee-74a4-4f43-98a9-3df8c6e61335 nodeName:}" failed. No retries permitted until 2025-12-11 02:18:36.742136515 +0000 UTC m=+1058.431173894 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/cb370fee-74a4-4f43-98a9-3df8c6e61335-cert") pod "openstack-baremetal-operator-controller-manager-84b575879f4k8rk" (UID: "cb370fee-74a4-4f43-98a9-3df8c6e61335") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.742185 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-967d97867-5tpj5"] Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.747230 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-5697bb5779-4cjld"] Dec 11 02:18:35 crc kubenswrapper[4824]: W1211 02:18:35.779253 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0eeea3b6_c514_4547_a2e0_52870ef572c6.slice/crio-c6f225eb96531d981fdd8e4cf55bc0bc2b4cb75262f0208df9623d062ac31b05 WatchSource:0}: Error finding container c6f225eb96531d981fdd8e4cf55bc0bc2b4cb75262f0208df9623d062ac31b05: Status 404 returned error can't find the container with id c6f225eb96531d981fdd8e4cf55bc0bc2b4cb75262f0208df9623d062ac31b05 Dec 11 02:18:35 crc kubenswrapper[4824]: W1211 02:18:35.786462 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaea47aaf_ee88_45a0_87b7_375b9f656828.slice/crio-9c2952afe2ae475c50710adb8e97613379664a37c8ad6fbfad60e6b9ec1d1a8f WatchSource:0}: Error finding container 9c2952afe2ae475c50710adb8e97613379664a37c8ad6fbfad60e6b9ec1d1a8f: Status 404 returned error can't find the container with id 9c2952afe2ae475c50710adb8e97613379664a37c8ad6fbfad60e6b9ec1d1a8f Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.787002 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-r8tnj" event={"ID":"df6864cf-9559-4398-bb3c-0f58f4c45563","Type":"ContainerStarted","Data":"26f39166462b89e7641044f25ff05ad3c34539d8540df561c41be12ecf2a4254"} Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.788669 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-4np4x" event={"ID":"cec1f29f-a1db-495f-b59c-e9308a9b53a2","Type":"ContainerStarted","Data":"cc48c3f9528eb1e008acc424d65168afd9815262769f537c15d0199a80684ab1"} Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.789482 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-8hvxh" event={"ID":"0b7b142b-04e2-4bb3-98ff-ed3b4da63746","Type":"ContainerStarted","Data":"ea711354df5e094163c91e8c641be1c110ee9e531fe73f3556857494e14d88a1"} Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.790128 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-4b2cb" event={"ID":"6743e907-9570-4dde-bb99-83e67202b224","Type":"ContainerStarted","Data":"9a918f9aa822c24548ed4a531e1eef49edc793b26ad78317999f472a564b97a7"} Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.887934 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-5qp6v"] Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.894014 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-l9jrd"] Dec 11 02:18:35 crc kubenswrapper[4824]: W1211 02:18:35.897320 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddb4453b9_ead1_45fc_91c5_d90541fa9c78.slice/crio-6f698c69a3fdaa187e173ee1ff9aa37a40beedf16688eb97ba2d5befbd2a8239 WatchSource:0}: Error finding container 6f698c69a3fdaa187e173ee1ff9aa37a40beedf16688eb97ba2d5befbd2a8239: Status 404 returned error can't find the container with id 6f698c69a3fdaa187e173ee1ff9aa37a40beedf16688eb97ba2d5befbd2a8239 Dec 11 02:18:35 crc kubenswrapper[4824]: W1211 02:18:35.897624 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1fc0b6e3_15c9_4f14_90f8_3f39a04ebc2c.slice/crio-b70c6fd8f8725251adc6c8952fbea7fe7ddae2d72a1834861bd6e4672f088a63 WatchSource:0}: Error finding container b70c6fd8f8725251adc6c8952fbea7fe7ddae2d72a1834861bd6e4672f088a63: Status 404 returned error can't find the container with id b70c6fd8f8725251adc6c8952fbea7fe7ddae2d72a1834861bd6e4672f088a63 Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.902696 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-lmszx"] Dec 11 02:18:35 crc kubenswrapper[4824]: W1211 02:18:35.913975 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddca85063_ba7a_45ac_ab88_5498a44834ba.slice/crio-baa1304f5391116552ed1351a125fa1be7a4fd7d825b15be88a6d3ed2269b6a0 WatchSource:0}: Error finding container baa1304f5391116552ed1351a125fa1be7a4fd7d825b15be88a6d3ed2269b6a0: Status 404 returned error can't find the container with id baa1304f5391116552ed1351a125fa1be7a4fd7d825b15be88a6d3ed2269b6a0 Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.916160 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-5b5fd79c9c-w24mx"] Dec 11 02:18:35 crc kubenswrapper[4824]: E1211 02:18:35.929023 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/manila-operator@sha256:44126f9c6b1d2bf752ddf989e20a4fc4cc1c07723d4fcb78465ccb2f55da6b3a,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6mpbl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-5b5fd79c9c-w24mx_openstack-operators(46bf5c63-42b4-45bd-ade3-446a4b2e4f6e): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 11 02:18:35 crc kubenswrapper[4824]: E1211 02:18:35.930774 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6mpbl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-5b5fd79c9c-w24mx_openstack-operators(46bf5c63-42b4-45bd-ade3-446a4b2e4f6e): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.930840 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-79c8c4686c-hpjng"] Dec 11 02:18:35 crc kubenswrapper[4824]: E1211 02:18:35.932046 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-w24mx" podUID="46bf5c63-42b4-45bd-ade3-446a4b2e4f6e" Dec 11 02:18:35 crc kubenswrapper[4824]: E1211 02:18:35.932235 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/mariadb-operator@sha256:424da951f13f1fbe9083215dc9f5088f90676dd813f01fdf3c1a8639b61cbaad,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hdppb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-79c8c4686c-hpjng_openstack-operators(fd3368a6-77e2-4c8f-b535-1d602b794d7d): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 11 02:18:35 crc kubenswrapper[4824]: E1211 02:18:35.936308 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hdppb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-79c8c4686c-hpjng_openstack-operators(fd3368a6-77e2-4c8f-b535-1d602b794d7d): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 11 02:18:35 crc kubenswrapper[4824]: E1211 02:18:35.937677 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-hpjng" podUID="fd3368a6-77e2-4c8f-b535-1d602b794d7d" Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.937786 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-zkz72"] Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.943098 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7180c14a-0c3a-4267-b122-5c30c5685d60-metrics-certs\") pod \"openstack-operator-controller-manager-56cf947496-nhjjr\" (UID: \"7180c14a-0c3a-4267-b122-5c30c5685d60\") " pod="openstack-operators/openstack-operator-controller-manager-56cf947496-nhjjr" Dec 11 02:18:35 crc kubenswrapper[4824]: I1211 02:18:35.943184 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/7180c14a-0c3a-4267-b122-5c30c5685d60-webhook-certs\") pod \"openstack-operator-controller-manager-56cf947496-nhjjr\" (UID: \"7180c14a-0c3a-4267-b122-5c30c5685d60\") " pod="openstack-operators/openstack-operator-controller-manager-56cf947496-nhjjr" Dec 11 02:18:35 crc kubenswrapper[4824]: E1211 02:18:35.943374 4824 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 11 02:18:35 crc kubenswrapper[4824]: E1211 02:18:35.943428 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7180c14a-0c3a-4267-b122-5c30c5685d60-webhook-certs podName:7180c14a-0c3a-4267-b122-5c30c5685d60 nodeName:}" failed. No retries permitted until 2025-12-11 02:18:36.943411415 +0000 UTC m=+1058.632448784 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/7180c14a-0c3a-4267-b122-5c30c5685d60-webhook-certs") pod "openstack-operator-controller-manager-56cf947496-nhjjr" (UID: "7180c14a-0c3a-4267-b122-5c30c5685d60") : secret "webhook-server-cert" not found Dec 11 02:18:35 crc kubenswrapper[4824]: E1211 02:18:35.943456 4824 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 11 02:18:35 crc kubenswrapper[4824]: E1211 02:18:35.943478 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7180c14a-0c3a-4267-b122-5c30c5685d60-metrics-certs podName:7180c14a-0c3a-4267-b122-5c30c5685d60 nodeName:}" failed. No retries permitted until 2025-12-11 02:18:36.943470407 +0000 UTC m=+1058.632507786 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7180c14a-0c3a-4267-b122-5c30c5685d60-metrics-certs") pod "openstack-operator-controller-manager-56cf947496-nhjjr" (UID: "7180c14a-0c3a-4267-b122-5c30c5685d60") : secret "metrics-server-cert" not found Dec 11 02:18:36 crc kubenswrapper[4824]: I1211 02:18:36.102568 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-4zfbk"] Dec 11 02:18:36 crc kubenswrapper[4824]: I1211 02:18:36.111849 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-9d58d64bc-fmvm7"] Dec 11 02:18:36 crc kubenswrapper[4824]: W1211 02:18:36.122992 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf1183bb7_356d_43b7_8ee7_55ab60dcbe54.slice/crio-9a12eed523dc195a1715507199ea468a66f20e84fdcb73ac0db4c6019f949429 WatchSource:0}: Error finding container 9a12eed523dc195a1715507199ea468a66f20e84fdcb73ac0db4c6019f949429: Status 404 returned error can't find the container with id 9a12eed523dc195a1715507199ea468a66f20e84fdcb73ac0db4c6019f949429 Dec 11 02:18:36 crc kubenswrapper[4824]: W1211 02:18:36.126053 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9793f7fe_2e1b_46c0_985e_1f77733f925f.slice/crio-45b1e19ef9e04408bc5df78b3a6aa641773c288ca59959172faa9c95bdc34890 WatchSource:0}: Error finding container 45b1e19ef9e04408bc5df78b3a6aa641773c288ca59959172faa9c95bdc34890: Status 404 returned error can't find the container with id 45b1e19ef9e04408bc5df78b3a6aa641773c288ca59959172faa9c95bdc34890 Dec 11 02:18:36 crc kubenswrapper[4824]: E1211 02:18:36.126262 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-sgb9f,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-998648c74-qqpf2_openstack-operators(f1183bb7-356d-43b7-8ee7-55ab60dcbe54): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 11 02:18:36 crc kubenswrapper[4824]: E1211 02:18:36.129517 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-sgb9f,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-998648c74-qqpf2_openstack-operators(f1183bb7-356d-43b7-8ee7-55ab60dcbe54): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 11 02:18:36 crc kubenswrapper[4824]: I1211 02:18:36.131568 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-75944c9b7-h6bgf"] Dec 11 02:18:36 crc kubenswrapper[4824]: E1211 02:18:36.131651 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/octavia-operator-controller-manager-998648c74-qqpf2" podUID="f1183bb7-356d-43b7-8ee7-55ab60dcbe54" Dec 11 02:18:36 crc kubenswrapper[4824]: I1211 02:18:36.139041 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-qqpf2"] Dec 11 02:18:36 crc kubenswrapper[4824]: E1211 02:18:36.140680 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:961417d59f527d925ac48ff6a11de747d0493315e496e34dc83d76a1a1fff58a,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-dnwhx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-75944c9b7-h6bgf_openstack-operators(9793f7fe-2e1b-46c0-985e-1f77733f925f): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 11 02:18:36 crc kubenswrapper[4824]: I1211 02:18:36.145793 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f219b9bf-34b1-48c4-96c8-20580b6ec9a4-cert\") pod \"infra-operator-controller-manager-78d48bff9d-j68mf\" (UID: \"f219b9bf-34b1-48c4-96c8-20580b6ec9a4\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-j68mf" Dec 11 02:18:36 crc kubenswrapper[4824]: E1211 02:18:36.145957 4824 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 11 02:18:36 crc kubenswrapper[4824]: E1211 02:18:36.145996 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f219b9bf-34b1-48c4-96c8-20580b6ec9a4-cert podName:f219b9bf-34b1-48c4-96c8-20580b6ec9a4 nodeName:}" failed. No retries permitted until 2025-12-11 02:18:38.1459835 +0000 UTC m=+1059.835020879 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f219b9bf-34b1-48c4-96c8-20580b6ec9a4-cert") pod "infra-operator-controller-manager-78d48bff9d-j68mf" (UID: "f219b9bf-34b1-48c4-96c8-20580b6ec9a4") : secret "infra-operator-webhook-server-cert" not found Dec 11 02:18:36 crc kubenswrapper[4824]: I1211 02:18:36.158069 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-wsjbh"] Dec 11 02:18:36 crc kubenswrapper[4824]: W1211 02:18:36.159049 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8047af9c_3a6d_4b2e_a088_df967e42671c.slice/crio-68bb518fe02ff67350159297c06e256387310ad1841a2a5a8c28d712651794f9 WatchSource:0}: Error finding container 68bb518fe02ff67350159297c06e256387310ad1841a2a5a8c28d712651794f9: Status 404 returned error can't find the container with id 68bb518fe02ff67350159297c06e256387310ad1841a2a5a8c28d712651794f9 Dec 11 02:18:36 crc kubenswrapper[4824]: W1211 02:18:36.159991 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod34bc4184_9cfb_4278_a9b9_80bf2f1ae612.slice/crio-bdf6ce6e6d9da10a6d9c46d76d9d0fd0dbf1bbb9a5a4944de992ba3a3e216be4 WatchSource:0}: Error finding container bdf6ce6e6d9da10a6d9c46d76d9d0fd0dbf1bbb9a5a4944de992ba3a3e216be4: Status 404 returned error can't find the container with id bdf6ce6e6d9da10a6d9c46d76d9d0fd0dbf1bbb9a5a4944de992ba3a3e216be4 Dec 11 02:18:36 crc kubenswrapper[4824]: E1211 02:18:36.163760 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hwv5j,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-wsjbh_openstack-operators(34bc4184-9cfb-4278-a9b9-80bf2f1ae612): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 11 02:18:36 crc kubenswrapper[4824]: E1211 02:18:36.163845 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:3aa109bb973253ae9dcf339b9b65abbd1176cdb4be672c93e538a5f113816991,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-zrm4p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-9d58d64bc-fmvm7_openstack-operators(8047af9c-3a6d-4b2e-a088-df967e42671c): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 11 02:18:36 crc kubenswrapper[4824]: E1211 02:18:36.165081 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-wsjbh" podUID="34bc4184-9cfb-4278-a9b9-80bf2f1ae612" Dec 11 02:18:36 crc kubenswrapper[4824]: E1211 02:18:36.166006 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-zrm4p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-9d58d64bc-fmvm7_openstack-operators(8047af9c-3a6d-4b2e-a088-df967e42671c): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 11 02:18:36 crc kubenswrapper[4824]: E1211 02:18:36.167262 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-fmvm7" podUID="8047af9c-3a6d-4b2e-a088-df967e42671c" Dec 11 02:18:36 crc kubenswrapper[4824]: W1211 02:18:36.191763 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod10325d55_7efd_49e8_95e9_46d68c7cd7ae.slice/crio-8674f1f065ebd816438f92357beb09a3ae3a2095cf7d60a7d9769530111d6ac4 WatchSource:0}: Error finding container 8674f1f065ebd816438f92357beb09a3ae3a2095cf7d60a7d9769530111d6ac4: Status 404 returned error can't find the container with id 8674f1f065ebd816438f92357beb09a3ae3a2095cf7d60a7d9769530111d6ac4 Dec 11 02:18:36 crc kubenswrapper[4824]: I1211 02:18:36.192017 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-r2bsg"] Dec 11 02:18:36 crc kubenswrapper[4824]: W1211 02:18:36.194380 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4d26d90a_05a8_4960_b2ca_12fa220383b7.slice/crio-9c1741de995294d11dabf44a39e51d54e43312f8add7acde2526e38b504ba41d WatchSource:0}: Error finding container 9c1741de995294d11dabf44a39e51d54e43312f8add7acde2526e38b504ba41d: Status 404 returned error can't find the container with id 9c1741de995294d11dabf44a39e51d54e43312f8add7acde2526e38b504ba41d Dec 11 02:18:36 crc kubenswrapper[4824]: E1211 02:18:36.194372 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-zvhsp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-78f8948974-xx9gc_openstack-operators(10325d55-7efd-49e8-95e9-46d68c7cd7ae): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 11 02:18:36 crc kubenswrapper[4824]: E1211 02:18:36.196642 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-zvhsp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-78f8948974-xx9gc_openstack-operators(10325d55-7efd-49e8-95e9-46d68c7cd7ae): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 11 02:18:36 crc kubenswrapper[4824]: E1211 02:18:36.196918 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-l6xnd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-b6456fdb6-r2bsg_openstack-operators(4d26d90a-05a8-4960-b2ca-12fa220383b7): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 11 02:18:36 crc kubenswrapper[4824]: I1211 02:18:36.197690 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-xx9gc"] Dec 11 02:18:36 crc kubenswrapper[4824]: E1211 02:18:36.197773 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/placement-operator-controller-manager-78f8948974-xx9gc" podUID="10325d55-7efd-49e8-95e9-46d68c7cd7ae" Dec 11 02:18:36 crc kubenswrapper[4824]: E1211 02:18:36.198659 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-l6xnd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-b6456fdb6-r2bsg_openstack-operators(4d26d90a-05a8-4960-b2ca-12fa220383b7): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 11 02:18:36 crc kubenswrapper[4824]: E1211 02:18:36.200497 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-r2bsg" podUID="4d26d90a-05a8-4960-b2ca-12fa220383b7" Dec 11 02:18:36 crc kubenswrapper[4824]: I1211 02:18:36.242996 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-58d5ff84df-qhhzn"] Dec 11 02:18:36 crc kubenswrapper[4824]: W1211 02:18:36.245250 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaff71eda_47bc_4b47_89df_fe9bf8357fba.slice/crio-c406b809d98b74bfaa168b32c72e427f2c3b91210103228ef9df37b33dbb1641 WatchSource:0}: Error finding container c406b809d98b74bfaa168b32c72e427f2c3b91210103228ef9df37b33dbb1641: Status 404 returned error can't find the container with id c406b809d98b74bfaa168b32c72e427f2c3b91210103228ef9df37b33dbb1641 Dec 11 02:18:36 crc kubenswrapper[4824]: E1211 02:18:36.248840 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:f27e732ec1faee765461bf137d9be81278b2fa39675019a73622755e1e610b6f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-8ljct,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-58d5ff84df-qhhzn_openstack-operators(aff71eda-47bc-4b47-89df-fe9bf8357fba): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 11 02:18:36 crc kubenswrapper[4824]: E1211 02:18:36.250730 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-8ljct,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-58d5ff84df-qhhzn_openstack-operators(aff71eda-47bc-4b47-89df-fe9bf8357fba): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 11 02:18:36 crc kubenswrapper[4824]: E1211 02:18:36.251872 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-qhhzn" podUID="aff71eda-47bc-4b47-89df-fe9bf8357fba" Dec 11 02:18:36 crc kubenswrapper[4824]: I1211 02:18:36.759718 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/cb370fee-74a4-4f43-98a9-3df8c6e61335-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879f4k8rk\" (UID: \"cb370fee-74a4-4f43-98a9-3df8c6e61335\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f4k8rk" Dec 11 02:18:36 crc kubenswrapper[4824]: E1211 02:18:36.759974 4824 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 11 02:18:36 crc kubenswrapper[4824]: E1211 02:18:36.760090 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cb370fee-74a4-4f43-98a9-3df8c6e61335-cert podName:cb370fee-74a4-4f43-98a9-3df8c6e61335 nodeName:}" failed. No retries permitted until 2025-12-11 02:18:38.760066383 +0000 UTC m=+1060.449103762 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/cb370fee-74a4-4f43-98a9-3df8c6e61335-cert") pod "openstack-baremetal-operator-controller-manager-84b575879f4k8rk" (UID: "cb370fee-74a4-4f43-98a9-3df8c6e61335") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 11 02:18:36 crc kubenswrapper[4824]: I1211 02:18:36.798027 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-967d97867-5tpj5" event={"ID":"aea47aaf-ee88-45a0-87b7-375b9f656828","Type":"ContainerStarted","Data":"9c2952afe2ae475c50710adb8e97613379664a37c8ad6fbfad60e6b9ec1d1a8f"} Dec 11 02:18:36 crc kubenswrapper[4824]: I1211 02:18:36.799259 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-qqpf2" event={"ID":"f1183bb7-356d-43b7-8ee7-55ab60dcbe54","Type":"ContainerStarted","Data":"9a12eed523dc195a1715507199ea468a66f20e84fdcb73ac0db4c6019f949429"} Dec 11 02:18:36 crc kubenswrapper[4824]: I1211 02:18:36.800639 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-wsjbh" event={"ID":"34bc4184-9cfb-4278-a9b9-80bf2f1ae612","Type":"ContainerStarted","Data":"bdf6ce6e6d9da10a6d9c46d76d9d0fd0dbf1bbb9a5a4944de992ba3a3e216be4"} Dec 11 02:18:36 crc kubenswrapper[4824]: E1211 02:18:36.801730 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/octavia-operator-controller-manager-998648c74-qqpf2" podUID="f1183bb7-356d-43b7-8ee7-55ab60dcbe54" Dec 11 02:18:36 crc kubenswrapper[4824]: E1211 02:18:36.802356 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-wsjbh" podUID="34bc4184-9cfb-4278-a9b9-80bf2f1ae612" Dec 11 02:18:36 crc kubenswrapper[4824]: I1211 02:18:36.803103 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-5qp6v" event={"ID":"1fc0b6e3-15c9-4f14-90f8-3f39a04ebc2c","Type":"ContainerStarted","Data":"b70c6fd8f8725251adc6c8952fbea7fe7ddae2d72a1834861bd6e4672f088a63"} Dec 11 02:18:36 crc kubenswrapper[4824]: I1211 02:18:36.804434 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-fmvm7" event={"ID":"8047af9c-3a6d-4b2e-a088-df967e42671c","Type":"ContainerStarted","Data":"68bb518fe02ff67350159297c06e256387310ad1841a2a5a8c28d712651794f9"} Dec 11 02:18:36 crc kubenswrapper[4824]: I1211 02:18:36.805724 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-h6bgf" event={"ID":"9793f7fe-2e1b-46c0-985e-1f77733f925f","Type":"ContainerStarted","Data":"45b1e19ef9e04408bc5df78b3a6aa641773c288ca59959172faa9c95bdc34890"} Dec 11 02:18:36 crc kubenswrapper[4824]: E1211 02:18:36.806180 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:3aa109bb973253ae9dcf339b9b65abbd1176cdb4be672c93e538a5f113816991\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-fmvm7" podUID="8047af9c-3a6d-4b2e-a088-df967e42671c" Dec 11 02:18:36 crc kubenswrapper[4824]: I1211 02:18:36.807724 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-qhhzn" event={"ID":"aff71eda-47bc-4b47-89df-fe9bf8357fba","Type":"ContainerStarted","Data":"c406b809d98b74bfaa168b32c72e427f2c3b91210103228ef9df37b33dbb1641"} Dec 11 02:18:36 crc kubenswrapper[4824]: I1211 02:18:36.809463 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-zkz72" event={"ID":"b206f71d-32cd-4b7d-ae38-a5125ab481dd","Type":"ContainerStarted","Data":"dbb2e8393255f32192fad4c4e37633403f49b40926e59cb6e3b0c2e9b564d6ca"} Dec 11 02:18:36 crc kubenswrapper[4824]: E1211 02:18:36.812465 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:f27e732ec1faee765461bf137d9be81278b2fa39675019a73622755e1e610b6f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-qhhzn" podUID="aff71eda-47bc-4b47-89df-fe9bf8357fba" Dec 11 02:18:36 crc kubenswrapper[4824]: I1211 02:18:36.818752 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-hpjng" event={"ID":"fd3368a6-77e2-4c8f-b535-1d602b794d7d","Type":"ContainerStarted","Data":"abfe18f6b806d49555825010ec20d34ccc737de987d03e1630c2df3b79b376f8"} Dec 11 02:18:36 crc kubenswrapper[4824]: I1211 02:18:36.828794 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-l9jrd" event={"ID":"db4453b9-ead1-45fc-91c5-d90541fa9c78","Type":"ContainerStarted","Data":"6f698c69a3fdaa187e173ee1ff9aa37a40beedf16688eb97ba2d5befbd2a8239"} Dec 11 02:18:36 crc kubenswrapper[4824]: E1211 02:18:36.829176 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:424da951f13f1fbe9083215dc9f5088f90676dd813f01fdf3c1a8639b61cbaad\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-hpjng" podUID="fd3368a6-77e2-4c8f-b535-1d602b794d7d" Dec 11 02:18:36 crc kubenswrapper[4824]: I1211 02:18:36.830768 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-r2bsg" event={"ID":"4d26d90a-05a8-4960-b2ca-12fa220383b7","Type":"ContainerStarted","Data":"9c1741de995294d11dabf44a39e51d54e43312f8add7acde2526e38b504ba41d"} Dec 11 02:18:36 crc kubenswrapper[4824]: I1211 02:18:36.831727 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-4cjld" event={"ID":"0eeea3b6-c514-4547-a2e0-52870ef572c6","Type":"ContainerStarted","Data":"c6f225eb96531d981fdd8e4cf55bc0bc2b4cb75262f0208df9623d062ac31b05"} Dec 11 02:18:36 crc kubenswrapper[4824]: I1211 02:18:36.833892 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-xx9gc" event={"ID":"10325d55-7efd-49e8-95e9-46d68c7cd7ae","Type":"ContainerStarted","Data":"8674f1f065ebd816438f92357beb09a3ae3a2095cf7d60a7d9769530111d6ac4"} Dec 11 02:18:36 crc kubenswrapper[4824]: E1211 02:18:36.834342 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-r2bsg" podUID="4d26d90a-05a8-4960-b2ca-12fa220383b7" Dec 11 02:18:36 crc kubenswrapper[4824]: I1211 02:18:36.835627 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-lmszx" event={"ID":"dca85063-ba7a-45ac-ab88-5498a44834ba","Type":"ContainerStarted","Data":"baa1304f5391116552ed1351a125fa1be7a4fd7d825b15be88a6d3ed2269b6a0"} Dec 11 02:18:36 crc kubenswrapper[4824]: I1211 02:18:36.837131 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-4zfbk" event={"ID":"cf19e8a0-8f29-48a4-b715-f499644fc896","Type":"ContainerStarted","Data":"e4276aec761625f6b553e069683d31829ed631aeaab8cb31097e49766e60f972"} Dec 11 02:18:36 crc kubenswrapper[4824]: E1211 02:18:36.837713 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/placement-operator-controller-manager-78f8948974-xx9gc" podUID="10325d55-7efd-49e8-95e9-46d68c7cd7ae" Dec 11 02:18:36 crc kubenswrapper[4824]: I1211 02:18:36.838095 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-w24mx" event={"ID":"46bf5c63-42b4-45bd-ade3-446a4b2e4f6e","Type":"ContainerStarted","Data":"87d19efa5c704b355904f0f1fe57ad59844152fe7e42a932b8629a497413cc4f"} Dec 11 02:18:36 crc kubenswrapper[4824]: E1211 02:18:36.841479 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/manila-operator@sha256:44126f9c6b1d2bf752ddf989e20a4fc4cc1c07723d4fcb78465ccb2f55da6b3a\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-w24mx" podUID="46bf5c63-42b4-45bd-ade3-446a4b2e4f6e" Dec 11 02:18:36 crc kubenswrapper[4824]: I1211 02:18:36.962524 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/7180c14a-0c3a-4267-b122-5c30c5685d60-webhook-certs\") pod \"openstack-operator-controller-manager-56cf947496-nhjjr\" (UID: \"7180c14a-0c3a-4267-b122-5c30c5685d60\") " pod="openstack-operators/openstack-operator-controller-manager-56cf947496-nhjjr" Dec 11 02:18:36 crc kubenswrapper[4824]: I1211 02:18:36.962754 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7180c14a-0c3a-4267-b122-5c30c5685d60-metrics-certs\") pod \"openstack-operator-controller-manager-56cf947496-nhjjr\" (UID: \"7180c14a-0c3a-4267-b122-5c30c5685d60\") " pod="openstack-operators/openstack-operator-controller-manager-56cf947496-nhjjr" Dec 11 02:18:36 crc kubenswrapper[4824]: E1211 02:18:36.963294 4824 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 11 02:18:36 crc kubenswrapper[4824]: E1211 02:18:36.963348 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7180c14a-0c3a-4267-b122-5c30c5685d60-metrics-certs podName:7180c14a-0c3a-4267-b122-5c30c5685d60 nodeName:}" failed. No retries permitted until 2025-12-11 02:18:38.963332007 +0000 UTC m=+1060.652369386 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7180c14a-0c3a-4267-b122-5c30c5685d60-metrics-certs") pod "openstack-operator-controller-manager-56cf947496-nhjjr" (UID: "7180c14a-0c3a-4267-b122-5c30c5685d60") : secret "metrics-server-cert" not found Dec 11 02:18:36 crc kubenswrapper[4824]: E1211 02:18:36.963663 4824 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 11 02:18:36 crc kubenswrapper[4824]: E1211 02:18:36.963739 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7180c14a-0c3a-4267-b122-5c30c5685d60-webhook-certs podName:7180c14a-0c3a-4267-b122-5c30c5685d60 nodeName:}" failed. No retries permitted until 2025-12-11 02:18:38.963723338 +0000 UTC m=+1060.652760717 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/7180c14a-0c3a-4267-b122-5c30c5685d60-webhook-certs") pod "openstack-operator-controller-manager-56cf947496-nhjjr" (UID: "7180c14a-0c3a-4267-b122-5c30c5685d60") : secret "webhook-server-cert" not found Dec 11 02:18:37 crc kubenswrapper[4824]: E1211 02:18:37.846775 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:3aa109bb973253ae9dcf339b9b65abbd1176cdb4be672c93e538a5f113816991\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-fmvm7" podUID="8047af9c-3a6d-4b2e-a088-df967e42671c" Dec 11 02:18:37 crc kubenswrapper[4824]: E1211 02:18:37.846790 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-r2bsg" podUID="4d26d90a-05a8-4960-b2ca-12fa220383b7" Dec 11 02:18:37 crc kubenswrapper[4824]: E1211 02:18:37.847473 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-wsjbh" podUID="34bc4184-9cfb-4278-a9b9-80bf2f1ae612" Dec 11 02:18:37 crc kubenswrapper[4824]: E1211 02:18:37.848658 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:f27e732ec1faee765461bf137d9be81278b2fa39675019a73622755e1e610b6f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-qhhzn" podUID="aff71eda-47bc-4b47-89df-fe9bf8357fba" Dec 11 02:18:37 crc kubenswrapper[4824]: E1211 02:18:37.848718 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/manila-operator@sha256:44126f9c6b1d2bf752ddf989e20a4fc4cc1c07723d4fcb78465ccb2f55da6b3a\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-w24mx" podUID="46bf5c63-42b4-45bd-ade3-446a4b2e4f6e" Dec 11 02:18:37 crc kubenswrapper[4824]: E1211 02:18:37.850057 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/placement-operator-controller-manager-78f8948974-xx9gc" podUID="10325d55-7efd-49e8-95e9-46d68c7cd7ae" Dec 11 02:18:37 crc kubenswrapper[4824]: E1211 02:18:37.850163 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:424da951f13f1fbe9083215dc9f5088f90676dd813f01fdf3c1a8639b61cbaad\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-hpjng" podUID="fd3368a6-77e2-4c8f-b535-1d602b794d7d" Dec 11 02:18:37 crc kubenswrapper[4824]: E1211 02:18:37.850405 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/octavia-operator-controller-manager-998648c74-qqpf2" podUID="f1183bb7-356d-43b7-8ee7-55ab60dcbe54" Dec 11 02:18:38 crc kubenswrapper[4824]: I1211 02:18:38.178040 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f219b9bf-34b1-48c4-96c8-20580b6ec9a4-cert\") pod \"infra-operator-controller-manager-78d48bff9d-j68mf\" (UID: \"f219b9bf-34b1-48c4-96c8-20580b6ec9a4\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-j68mf" Dec 11 02:18:38 crc kubenswrapper[4824]: E1211 02:18:38.178200 4824 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 11 02:18:38 crc kubenswrapper[4824]: E1211 02:18:38.178247 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f219b9bf-34b1-48c4-96c8-20580b6ec9a4-cert podName:f219b9bf-34b1-48c4-96c8-20580b6ec9a4 nodeName:}" failed. No retries permitted until 2025-12-11 02:18:42.178234272 +0000 UTC m=+1063.867271641 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f219b9bf-34b1-48c4-96c8-20580b6ec9a4-cert") pod "infra-operator-controller-manager-78d48bff9d-j68mf" (UID: "f219b9bf-34b1-48c4-96c8-20580b6ec9a4") : secret "infra-operator-webhook-server-cert" not found Dec 11 02:18:38 crc kubenswrapper[4824]: I1211 02:18:38.790328 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/cb370fee-74a4-4f43-98a9-3df8c6e61335-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879f4k8rk\" (UID: \"cb370fee-74a4-4f43-98a9-3df8c6e61335\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f4k8rk" Dec 11 02:18:38 crc kubenswrapper[4824]: E1211 02:18:38.791095 4824 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 11 02:18:38 crc kubenswrapper[4824]: E1211 02:18:38.791182 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cb370fee-74a4-4f43-98a9-3df8c6e61335-cert podName:cb370fee-74a4-4f43-98a9-3df8c6e61335 nodeName:}" failed. No retries permitted until 2025-12-11 02:18:42.791162325 +0000 UTC m=+1064.480199904 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/cb370fee-74a4-4f43-98a9-3df8c6e61335-cert") pod "openstack-baremetal-operator-controller-manager-84b575879f4k8rk" (UID: "cb370fee-74a4-4f43-98a9-3df8c6e61335") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 11 02:18:38 crc kubenswrapper[4824]: I1211 02:18:38.992748 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7180c14a-0c3a-4267-b122-5c30c5685d60-metrics-certs\") pod \"openstack-operator-controller-manager-56cf947496-nhjjr\" (UID: \"7180c14a-0c3a-4267-b122-5c30c5685d60\") " pod="openstack-operators/openstack-operator-controller-manager-56cf947496-nhjjr" Dec 11 02:18:38 crc kubenswrapper[4824]: E1211 02:18:38.992994 4824 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 11 02:18:38 crc kubenswrapper[4824]: E1211 02:18:38.994287 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7180c14a-0c3a-4267-b122-5c30c5685d60-metrics-certs podName:7180c14a-0c3a-4267-b122-5c30c5685d60 nodeName:}" failed. No retries permitted until 2025-12-11 02:18:42.994264654 +0000 UTC m=+1064.683302033 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7180c14a-0c3a-4267-b122-5c30c5685d60-metrics-certs") pod "openstack-operator-controller-manager-56cf947496-nhjjr" (UID: "7180c14a-0c3a-4267-b122-5c30c5685d60") : secret "metrics-server-cert" not found Dec 11 02:18:38 crc kubenswrapper[4824]: I1211 02:18:38.994010 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/7180c14a-0c3a-4267-b122-5c30c5685d60-webhook-certs\") pod \"openstack-operator-controller-manager-56cf947496-nhjjr\" (UID: \"7180c14a-0c3a-4267-b122-5c30c5685d60\") " pod="openstack-operators/openstack-operator-controller-manager-56cf947496-nhjjr" Dec 11 02:18:38 crc kubenswrapper[4824]: E1211 02:18:38.994538 4824 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 11 02:18:38 crc kubenswrapper[4824]: E1211 02:18:38.994622 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7180c14a-0c3a-4267-b122-5c30c5685d60-webhook-certs podName:7180c14a-0c3a-4267-b122-5c30c5685d60 nodeName:}" failed. No retries permitted until 2025-12-11 02:18:42.994603803 +0000 UTC m=+1064.683641182 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/7180c14a-0c3a-4267-b122-5c30c5685d60-webhook-certs") pod "openstack-operator-controller-manager-56cf947496-nhjjr" (UID: "7180c14a-0c3a-4267-b122-5c30c5685d60") : secret "webhook-server-cert" not found Dec 11 02:18:42 crc kubenswrapper[4824]: I1211 02:18:42.251203 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f219b9bf-34b1-48c4-96c8-20580b6ec9a4-cert\") pod \"infra-operator-controller-manager-78d48bff9d-j68mf\" (UID: \"f219b9bf-34b1-48c4-96c8-20580b6ec9a4\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-j68mf" Dec 11 02:18:42 crc kubenswrapper[4824]: E1211 02:18:42.251423 4824 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 11 02:18:42 crc kubenswrapper[4824]: E1211 02:18:42.251528 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f219b9bf-34b1-48c4-96c8-20580b6ec9a4-cert podName:f219b9bf-34b1-48c4-96c8-20580b6ec9a4 nodeName:}" failed. No retries permitted until 2025-12-11 02:18:50.2515053 +0000 UTC m=+1071.940542679 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f219b9bf-34b1-48c4-96c8-20580b6ec9a4-cert") pod "infra-operator-controller-manager-78d48bff9d-j68mf" (UID: "f219b9bf-34b1-48c4-96c8-20580b6ec9a4") : secret "infra-operator-webhook-server-cert" not found Dec 11 02:18:42 crc kubenswrapper[4824]: I1211 02:18:42.860718 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/cb370fee-74a4-4f43-98a9-3df8c6e61335-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879f4k8rk\" (UID: \"cb370fee-74a4-4f43-98a9-3df8c6e61335\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f4k8rk" Dec 11 02:18:42 crc kubenswrapper[4824]: E1211 02:18:42.861003 4824 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 11 02:18:42 crc kubenswrapper[4824]: E1211 02:18:42.861092 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cb370fee-74a4-4f43-98a9-3df8c6e61335-cert podName:cb370fee-74a4-4f43-98a9-3df8c6e61335 nodeName:}" failed. No retries permitted until 2025-12-11 02:18:50.861071133 +0000 UTC m=+1072.550108512 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/cb370fee-74a4-4f43-98a9-3df8c6e61335-cert") pod "openstack-baremetal-operator-controller-manager-84b575879f4k8rk" (UID: "cb370fee-74a4-4f43-98a9-3df8c6e61335") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 11 02:18:43 crc kubenswrapper[4824]: I1211 02:18:43.065078 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7180c14a-0c3a-4267-b122-5c30c5685d60-metrics-certs\") pod \"openstack-operator-controller-manager-56cf947496-nhjjr\" (UID: \"7180c14a-0c3a-4267-b122-5c30c5685d60\") " pod="openstack-operators/openstack-operator-controller-manager-56cf947496-nhjjr" Dec 11 02:18:43 crc kubenswrapper[4824]: I1211 02:18:43.065306 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/7180c14a-0c3a-4267-b122-5c30c5685d60-webhook-certs\") pod \"openstack-operator-controller-manager-56cf947496-nhjjr\" (UID: \"7180c14a-0c3a-4267-b122-5c30c5685d60\") " pod="openstack-operators/openstack-operator-controller-manager-56cf947496-nhjjr" Dec 11 02:18:43 crc kubenswrapper[4824]: E1211 02:18:43.065391 4824 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 11 02:18:43 crc kubenswrapper[4824]: E1211 02:18:43.065434 4824 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 11 02:18:43 crc kubenswrapper[4824]: E1211 02:18:43.065472 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7180c14a-0c3a-4267-b122-5c30c5685d60-metrics-certs podName:7180c14a-0c3a-4267-b122-5c30c5685d60 nodeName:}" failed. No retries permitted until 2025-12-11 02:18:51.065454506 +0000 UTC m=+1072.754491885 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7180c14a-0c3a-4267-b122-5c30c5685d60-metrics-certs") pod "openstack-operator-controller-manager-56cf947496-nhjjr" (UID: "7180c14a-0c3a-4267-b122-5c30c5685d60") : secret "metrics-server-cert" not found Dec 11 02:18:43 crc kubenswrapper[4824]: E1211 02:18:43.065491 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7180c14a-0c3a-4267-b122-5c30c5685d60-webhook-certs podName:7180c14a-0c3a-4267-b122-5c30c5685d60 nodeName:}" failed. No retries permitted until 2025-12-11 02:18:51.065479787 +0000 UTC m=+1072.754517166 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/7180c14a-0c3a-4267-b122-5c30c5685d60-webhook-certs") pod "openstack-operator-controller-manager-56cf947496-nhjjr" (UID: "7180c14a-0c3a-4267-b122-5c30c5685d60") : secret "webhook-server-cert" not found Dec 11 02:18:46 crc kubenswrapper[4824]: E1211 02:18:46.715777 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-h6bgf" podUID="9793f7fe-2e1b-46c0-985e-1f77733f925f" Dec 11 02:18:46 crc kubenswrapper[4824]: I1211 02:18:46.951300 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-r8tnj" event={"ID":"df6864cf-9559-4398-bb3c-0f58f4c45563","Type":"ContainerStarted","Data":"9bb49f28bfad1301583ccea853b6adc00f1a8617537a3becc8f13c7e12888db5"} Dec 11 02:18:46 crc kubenswrapper[4824]: I1211 02:18:46.982266 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-h6bgf" event={"ID":"9793f7fe-2e1b-46c0-985e-1f77733f925f","Type":"ContainerStarted","Data":"3dec1dd9e97e5a3d6390d294e4b41af393a463f18bedf59b1fbc0aa6707c77d8"} Dec 11 02:18:47 crc kubenswrapper[4824]: I1211 02:18:46.999306 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-zkz72" event={"ID":"b206f71d-32cd-4b7d-ae38-a5125ab481dd","Type":"ContainerStarted","Data":"f9d507a4e6a34fc3cd84c0fea67aeb4893e5d5a0c21a0e7beecbfc5e12cc429b"} Dec 11 02:18:47 crc kubenswrapper[4824]: I1211 02:18:47.040651 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-967d97867-5tpj5" event={"ID":"aea47aaf-ee88-45a0-87b7-375b9f656828","Type":"ContainerStarted","Data":"145c32cc31e5cfdc97fd0ce0f82820e1b80f3757a75db94c2d1aa00a705a11e5"} Dec 11 02:18:47 crc kubenswrapper[4824]: I1211 02:18:47.060266 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-4b2cb" event={"ID":"6743e907-9570-4dde-bb99-83e67202b224","Type":"ContainerStarted","Data":"c19bc2ad40d11ff335f7428fefa79b404d9dc448d53df394ef80e64697182ac4"} Dec 11 02:18:47 crc kubenswrapper[4824]: I1211 02:18:47.060310 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-4b2cb" event={"ID":"6743e907-9570-4dde-bb99-83e67202b224","Type":"ContainerStarted","Data":"2e70b46413c42487dee83e4d636ab5e4efe60bbe97684b0d63286ade48226100"} Dec 11 02:18:47 crc kubenswrapper[4824]: I1211 02:18:47.061060 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-4b2cb" Dec 11 02:18:47 crc kubenswrapper[4824]: I1211 02:18:47.083633 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-4cjld" event={"ID":"0eeea3b6-c514-4547-a2e0-52870ef572c6","Type":"ContainerStarted","Data":"7aed13f394b2cdcc75acb82d65a0b528407d2d721bf56828465b29c8ed08faf9"} Dec 11 02:18:47 crc kubenswrapper[4824]: I1211 02:18:47.083681 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-4cjld" event={"ID":"0eeea3b6-c514-4547-a2e0-52870ef572c6","Type":"ContainerStarted","Data":"f9fce47ae2a3d990bddc2248e0160e5614473d5938fdbdc98632b20b611d0782"} Dec 11 02:18:47 crc kubenswrapper[4824]: I1211 02:18:47.083876 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-4cjld" Dec 11 02:18:47 crc kubenswrapper[4824]: I1211 02:18:47.102619 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-4zfbk" event={"ID":"cf19e8a0-8f29-48a4-b715-f499644fc896","Type":"ContainerStarted","Data":"ec977eb476cd64d03a6bfdaad96f58ccad694fb4d7ed5b87095601049f805bbe"} Dec 11 02:18:47 crc kubenswrapper[4824]: I1211 02:18:47.120363 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-4np4x" event={"ID":"cec1f29f-a1db-495f-b59c-e9308a9b53a2","Type":"ContainerStarted","Data":"c7e38feba2f9060d72887b5f5eee3df0469fb896915c93f959a9c215144ec6ef"} Dec 11 02:18:47 crc kubenswrapper[4824]: I1211 02:18:47.137819 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-5qp6v" event={"ID":"1fc0b6e3-15c9-4f14-90f8-3f39a04ebc2c","Type":"ContainerStarted","Data":"8f9a7f73450dfb8b6e5ad9eb23d15568d193b9215bee9586d571dc7cb21f731d"} Dec 11 02:18:47 crc kubenswrapper[4824]: I1211 02:18:47.137858 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-5qp6v" event={"ID":"1fc0b6e3-15c9-4f14-90f8-3f39a04ebc2c","Type":"ContainerStarted","Data":"ac03b80a27d11ba2d77da59e1e16e7a421a5235974a79da0ce9d7f34d69ce98f"} Dec 11 02:18:47 crc kubenswrapper[4824]: I1211 02:18:47.138644 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-5qp6v" Dec 11 02:18:47 crc kubenswrapper[4824]: I1211 02:18:47.144157 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-8hvxh" event={"ID":"0b7b142b-04e2-4bb3-98ff-ed3b4da63746","Type":"ContainerStarted","Data":"b11c1f8ef0ebe5c1ac5b81f1f63db9aa36ee2e9064080709f96e9eab0c18491a"} Dec 11 02:18:47 crc kubenswrapper[4824]: I1211 02:18:47.144203 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-8hvxh" event={"ID":"0b7b142b-04e2-4bb3-98ff-ed3b4da63746","Type":"ContainerStarted","Data":"c320ee8145033835abc7432383270022918b52ecf304717240058c05f212d1f3"} Dec 11 02:18:47 crc kubenswrapper[4824]: I1211 02:18:47.144836 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-8hvxh" Dec 11 02:18:47 crc kubenswrapper[4824]: I1211 02:18:47.150872 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-lmszx" event={"ID":"dca85063-ba7a-45ac-ab88-5498a44834ba","Type":"ContainerStarted","Data":"69e2bacdaa57d4c57a9bea945744193d917be50dd7f55ce0b5aaba7b0a817f8d"} Dec 11 02:18:47 crc kubenswrapper[4824]: I1211 02:18:47.151399 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-lmszx" Dec 11 02:18:47 crc kubenswrapper[4824]: I1211 02:18:47.161902 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-l9jrd" event={"ID":"db4453b9-ead1-45fc-91c5-d90541fa9c78","Type":"ContainerStarted","Data":"e15a54ab0b0f6aee1d12cf828c300bb7d63eb4675caa2665ec352f8f8b3bbc8b"} Dec 11 02:18:47 crc kubenswrapper[4824]: I1211 02:18:47.162478 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-l9jrd" Dec 11 02:18:47 crc kubenswrapper[4824]: I1211 02:18:47.205032 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-5qp6v" podStartSLOduration=4.451448208 podStartE2EDuration="13.205012869s" podCreationTimestamp="2025-12-11 02:18:34 +0000 UTC" firstStartedPulling="2025-12-11 02:18:35.904640753 +0000 UTC m=+1057.593678132" lastFinishedPulling="2025-12-11 02:18:44.658205404 +0000 UTC m=+1066.347242793" observedRunningTime="2025-12-11 02:18:47.195250329 +0000 UTC m=+1068.884287708" watchObservedRunningTime="2025-12-11 02:18:47.205012869 +0000 UTC m=+1068.894050248" Dec 11 02:18:47 crc kubenswrapper[4824]: I1211 02:18:47.206833 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-4b2cb" podStartSLOduration=4.31825476 podStartE2EDuration="13.206826487s" podCreationTimestamp="2025-12-11 02:18:34 +0000 UTC" firstStartedPulling="2025-12-11 02:18:35.767957482 +0000 UTC m=+1057.456994861" lastFinishedPulling="2025-12-11 02:18:44.656529209 +0000 UTC m=+1066.345566588" observedRunningTime="2025-12-11 02:18:47.139763052 +0000 UTC m=+1068.828800431" watchObservedRunningTime="2025-12-11 02:18:47.206826487 +0000 UTC m=+1068.895863866" Dec 11 02:18:47 crc kubenswrapper[4824]: I1211 02:18:47.227589 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-4cjld" podStartSLOduration=4.367393639 podStartE2EDuration="13.22757666s" podCreationTimestamp="2025-12-11 02:18:34 +0000 UTC" firstStartedPulling="2025-12-11 02:18:35.796466671 +0000 UTC m=+1057.485504050" lastFinishedPulling="2025-12-11 02:18:44.656649662 +0000 UTC m=+1066.345687071" observedRunningTime="2025-12-11 02:18:47.221688883 +0000 UTC m=+1068.910726262" watchObservedRunningTime="2025-12-11 02:18:47.22757666 +0000 UTC m=+1068.916614039" Dec 11 02:18:47 crc kubenswrapper[4824]: I1211 02:18:47.254938 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-8hvxh" podStartSLOduration=4.357379434 podStartE2EDuration="13.254922469s" podCreationTimestamp="2025-12-11 02:18:34 +0000 UTC" firstStartedPulling="2025-12-11 02:18:35.758987194 +0000 UTC m=+1057.448024583" lastFinishedPulling="2025-12-11 02:18:44.656530239 +0000 UTC m=+1066.345567618" observedRunningTime="2025-12-11 02:18:47.250591573 +0000 UTC m=+1068.939628952" watchObservedRunningTime="2025-12-11 02:18:47.254922469 +0000 UTC m=+1068.943959838" Dec 11 02:18:47 crc kubenswrapper[4824]: I1211 02:18:47.327976 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-lmszx" podStartSLOduration=3.917348155 podStartE2EDuration="13.327961964s" podCreationTimestamp="2025-12-11 02:18:34 +0000 UTC" firstStartedPulling="2025-12-11 02:18:35.916071297 +0000 UTC m=+1057.605108676" lastFinishedPulling="2025-12-11 02:18:45.326685106 +0000 UTC m=+1067.015722485" observedRunningTime="2025-12-11 02:18:47.318330737 +0000 UTC m=+1069.007368116" watchObservedRunningTime="2025-12-11 02:18:47.327961964 +0000 UTC m=+1069.016999343" Dec 11 02:18:47 crc kubenswrapper[4824]: I1211 02:18:47.357756 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-l9jrd" podStartSLOduration=4.599976644 podStartE2EDuration="13.357742186s" podCreationTimestamp="2025-12-11 02:18:34 +0000 UTC" firstStartedPulling="2025-12-11 02:18:35.89889234 +0000 UTC m=+1057.587929719" lastFinishedPulling="2025-12-11 02:18:44.656657872 +0000 UTC m=+1066.345695261" observedRunningTime="2025-12-11 02:18:47.353333209 +0000 UTC m=+1069.042370588" watchObservedRunningTime="2025-12-11 02:18:47.357742186 +0000 UTC m=+1069.046779565" Dec 11 02:18:48 crc kubenswrapper[4824]: I1211 02:18:48.168042 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-967d97867-5tpj5" event={"ID":"aea47aaf-ee88-45a0-87b7-375b9f656828","Type":"ContainerStarted","Data":"d1cd688153125913163c2c387ae5ea994918eb09c0b539ce6e9bdcbdaae682c3"} Dec 11 02:18:48 crc kubenswrapper[4824]: I1211 02:18:48.168938 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-967d97867-5tpj5" Dec 11 02:18:48 crc kubenswrapper[4824]: I1211 02:18:48.170194 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-4zfbk" event={"ID":"cf19e8a0-8f29-48a4-b715-f499644fc896","Type":"ContainerStarted","Data":"ff9e38c53215c90b2bdfe0af1ba59d1fc88473ad0f0f9a3710c9b66660094ac2"} Dec 11 02:18:48 crc kubenswrapper[4824]: I1211 02:18:48.170529 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5854674fcc-4zfbk" Dec 11 02:18:48 crc kubenswrapper[4824]: I1211 02:18:48.172170 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-4np4x" event={"ID":"cec1f29f-a1db-495f-b59c-e9308a9b53a2","Type":"ContainerStarted","Data":"9279dc75ab3d2f99d424dde605eae76eb13a5e3ce95c2672c02a593664364910"} Dec 11 02:18:48 crc kubenswrapper[4824]: I1211 02:18:48.172251 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-4np4x" Dec 11 02:18:48 crc kubenswrapper[4824]: I1211 02:18:48.173590 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-r8tnj" event={"ID":"df6864cf-9559-4398-bb3c-0f58f4c45563","Type":"ContainerStarted","Data":"17046af5f4fcba12692d80c0cac1b4fecbb6a883d1e4eeda986ad0bd62dc5e68"} Dec 11 02:18:48 crc kubenswrapper[4824]: I1211 02:18:48.173921 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-r8tnj" Dec 11 02:18:48 crc kubenswrapper[4824]: I1211 02:18:48.175315 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-zkz72" event={"ID":"b206f71d-32cd-4b7d-ae38-a5125ab481dd","Type":"ContainerStarted","Data":"d131dfed7309ab21a8ec6147551fc9b82da16bbfc03bf9c8b7fde1daf722ae84"} Dec 11 02:18:48 crc kubenswrapper[4824]: I1211 02:18:48.175648 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-zkz72" Dec 11 02:18:48 crc kubenswrapper[4824]: I1211 02:18:48.176958 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-lmszx" event={"ID":"dca85063-ba7a-45ac-ab88-5498a44834ba","Type":"ContainerStarted","Data":"f4fd2181a4711423c54444c3db8b8029fbca3b158916134466df3c71197271ef"} Dec 11 02:18:48 crc kubenswrapper[4824]: I1211 02:18:48.178395 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-l9jrd" event={"ID":"db4453b9-ead1-45fc-91c5-d90541fa9c78","Type":"ContainerStarted","Data":"648f4dc9e4473a8328cc8c4c7744075b8199d8116f157455ee057441a79f40c4"} Dec 11 02:18:48 crc kubenswrapper[4824]: I1211 02:18:48.190143 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-967d97867-5tpj5" podStartSLOduration=4.6032745219999995 podStartE2EDuration="14.190127334s" podCreationTimestamp="2025-12-11 02:18:34 +0000 UTC" firstStartedPulling="2025-12-11 02:18:35.809760556 +0000 UTC m=+1057.498797935" lastFinishedPulling="2025-12-11 02:18:45.396613378 +0000 UTC m=+1067.085650747" observedRunningTime="2025-12-11 02:18:48.184486693 +0000 UTC m=+1069.873524072" watchObservedRunningTime="2025-12-11 02:18:48.190127334 +0000 UTC m=+1069.879164713" Dec 11 02:18:48 crc kubenswrapper[4824]: I1211 02:18:48.230892 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-zkz72" podStartSLOduration=4.615279591 podStartE2EDuration="14.230872489s" podCreationTimestamp="2025-12-11 02:18:34 +0000 UTC" firstStartedPulling="2025-12-11 02:18:35.923422253 +0000 UTC m=+1057.612459632" lastFinishedPulling="2025-12-11 02:18:45.539015151 +0000 UTC m=+1067.228052530" observedRunningTime="2025-12-11 02:18:48.229204454 +0000 UTC m=+1069.918241853" watchObservedRunningTime="2025-12-11 02:18:48.230872489 +0000 UTC m=+1069.919909868" Dec 11 02:18:48 crc kubenswrapper[4824]: I1211 02:18:48.233346 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5854674fcc-4zfbk" podStartSLOduration=4.945797063 podStartE2EDuration="14.233337355s" podCreationTimestamp="2025-12-11 02:18:34 +0000 UTC" firstStartedPulling="2025-12-11 02:18:36.109312173 +0000 UTC m=+1057.798349552" lastFinishedPulling="2025-12-11 02:18:45.396852425 +0000 UTC m=+1067.085889844" observedRunningTime="2025-12-11 02:18:48.203927271 +0000 UTC m=+1069.892964660" watchObservedRunningTime="2025-12-11 02:18:48.233337355 +0000 UTC m=+1069.922374734" Dec 11 02:18:48 crc kubenswrapper[4824]: I1211 02:18:48.246490 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-4np4x" podStartSLOduration=4.469048837 podStartE2EDuration="14.246471445s" podCreationTimestamp="2025-12-11 02:18:34 +0000 UTC" firstStartedPulling="2025-12-11 02:18:35.760869664 +0000 UTC m=+1057.449907043" lastFinishedPulling="2025-12-11 02:18:45.538292272 +0000 UTC m=+1067.227329651" observedRunningTime="2025-12-11 02:18:48.24216289 +0000 UTC m=+1069.931200279" watchObservedRunningTime="2025-12-11 02:18:48.246471445 +0000 UTC m=+1069.935508824" Dec 11 02:18:48 crc kubenswrapper[4824]: I1211 02:18:48.271843 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-r8tnj" podStartSLOduration=5.356791449 podStartE2EDuration="14.27182467s" podCreationTimestamp="2025-12-11 02:18:34 +0000 UTC" firstStartedPulling="2025-12-11 02:18:35.741724964 +0000 UTC m=+1057.430762343" lastFinishedPulling="2025-12-11 02:18:44.656758145 +0000 UTC m=+1066.345795564" observedRunningTime="2025-12-11 02:18:48.269083327 +0000 UTC m=+1069.958120706" watchObservedRunningTime="2025-12-11 02:18:48.27182467 +0000 UTC m=+1069.960862049" Dec 11 02:18:50 crc kubenswrapper[4824]: I1211 02:18:50.193837 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-h6bgf" event={"ID":"9793f7fe-2e1b-46c0-985e-1f77733f925f","Type":"ContainerStarted","Data":"5e307afcdad23cdf8577d0cfc3c783e5698b5995a45cd7f8b0db20d5e0649148"} Dec 11 02:18:50 crc kubenswrapper[4824]: I1211 02:18:50.212185 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-h6bgf" podStartSLOduration=3.122219678 podStartE2EDuration="16.212167904s" podCreationTimestamp="2025-12-11 02:18:34 +0000 UTC" firstStartedPulling="2025-12-11 02:18:36.140594986 +0000 UTC m=+1057.829632355" lastFinishedPulling="2025-12-11 02:18:49.230543192 +0000 UTC m=+1070.919580581" observedRunningTime="2025-12-11 02:18:50.207089188 +0000 UTC m=+1071.896126587" watchObservedRunningTime="2025-12-11 02:18:50.212167904 +0000 UTC m=+1071.901205283" Dec 11 02:18:50 crc kubenswrapper[4824]: I1211 02:18:50.281060 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f219b9bf-34b1-48c4-96c8-20580b6ec9a4-cert\") pod \"infra-operator-controller-manager-78d48bff9d-j68mf\" (UID: \"f219b9bf-34b1-48c4-96c8-20580b6ec9a4\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-j68mf" Dec 11 02:18:50 crc kubenswrapper[4824]: E1211 02:18:50.281372 4824 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 11 02:18:50 crc kubenswrapper[4824]: E1211 02:18:50.281531 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f219b9bf-34b1-48c4-96c8-20580b6ec9a4-cert podName:f219b9bf-34b1-48c4-96c8-20580b6ec9a4 nodeName:}" failed. No retries permitted until 2025-12-11 02:19:06.281448259 +0000 UTC m=+1087.970485668 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f219b9bf-34b1-48c4-96c8-20580b6ec9a4-cert") pod "infra-operator-controller-manager-78d48bff9d-j68mf" (UID: "f219b9bf-34b1-48c4-96c8-20580b6ec9a4") : secret "infra-operator-webhook-server-cert" not found Dec 11 02:18:50 crc kubenswrapper[4824]: I1211 02:18:50.889225 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/cb370fee-74a4-4f43-98a9-3df8c6e61335-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879f4k8rk\" (UID: \"cb370fee-74a4-4f43-98a9-3df8c6e61335\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f4k8rk" Dec 11 02:18:50 crc kubenswrapper[4824]: I1211 02:18:50.911632 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/cb370fee-74a4-4f43-98a9-3df8c6e61335-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879f4k8rk\" (UID: \"cb370fee-74a4-4f43-98a9-3df8c6e61335\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f4k8rk" Dec 11 02:18:51 crc kubenswrapper[4824]: I1211 02:18:51.064686 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f4k8rk" Dec 11 02:18:51 crc kubenswrapper[4824]: I1211 02:18:51.092123 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/7180c14a-0c3a-4267-b122-5c30c5685d60-webhook-certs\") pod \"openstack-operator-controller-manager-56cf947496-nhjjr\" (UID: \"7180c14a-0c3a-4267-b122-5c30c5685d60\") " pod="openstack-operators/openstack-operator-controller-manager-56cf947496-nhjjr" Dec 11 02:18:51 crc kubenswrapper[4824]: I1211 02:18:51.092262 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7180c14a-0c3a-4267-b122-5c30c5685d60-metrics-certs\") pod \"openstack-operator-controller-manager-56cf947496-nhjjr\" (UID: \"7180c14a-0c3a-4267-b122-5c30c5685d60\") " pod="openstack-operators/openstack-operator-controller-manager-56cf947496-nhjjr" Dec 11 02:18:51 crc kubenswrapper[4824]: E1211 02:18:51.092285 4824 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 11 02:18:51 crc kubenswrapper[4824]: E1211 02:18:51.092382 4824 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 11 02:18:51 crc kubenswrapper[4824]: E1211 02:18:51.092387 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7180c14a-0c3a-4267-b122-5c30c5685d60-webhook-certs podName:7180c14a-0c3a-4267-b122-5c30c5685d60 nodeName:}" failed. No retries permitted until 2025-12-11 02:19:07.092363725 +0000 UTC m=+1088.781401124 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/7180c14a-0c3a-4267-b122-5c30c5685d60-webhook-certs") pod "openstack-operator-controller-manager-56cf947496-nhjjr" (UID: "7180c14a-0c3a-4267-b122-5c30c5685d60") : secret "webhook-server-cert" not found Dec 11 02:18:51 crc kubenswrapper[4824]: E1211 02:18:51.092446 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7180c14a-0c3a-4267-b122-5c30c5685d60-metrics-certs podName:7180c14a-0c3a-4267-b122-5c30c5685d60 nodeName:}" failed. No retries permitted until 2025-12-11 02:19:07.092430407 +0000 UTC m=+1088.781467786 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7180c14a-0c3a-4267-b122-5c30c5685d60-metrics-certs") pod "openstack-operator-controller-manager-56cf947496-nhjjr" (UID: "7180c14a-0c3a-4267-b122-5c30c5685d60") : secret "metrics-server-cert" not found Dec 11 02:18:51 crc kubenswrapper[4824]: I1211 02:18:51.218828 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-qhhzn" event={"ID":"aff71eda-47bc-4b47-89df-fe9bf8357fba","Type":"ContainerStarted","Data":"6b6939a57b96f350b358209d1c76e02957d0a036a2fb7ad4c984ee225e116f8a"} Dec 11 02:18:51 crc kubenswrapper[4824]: I1211 02:18:51.219325 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-qhhzn" event={"ID":"aff71eda-47bc-4b47-89df-fe9bf8357fba","Type":"ContainerStarted","Data":"7302c18546fbfa9c86aed0aad510ba1fed2aa16e70ae877cfee4afcc39469461"} Dec 11 02:18:51 crc kubenswrapper[4824]: I1211 02:18:51.219447 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-h6bgf" Dec 11 02:18:51 crc kubenswrapper[4824]: I1211 02:18:51.219511 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-qhhzn" Dec 11 02:18:51 crc kubenswrapper[4824]: I1211 02:18:51.237101 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-qhhzn" podStartSLOduration=2.917448696 podStartE2EDuration="17.237084349s" podCreationTimestamp="2025-12-11 02:18:34 +0000 UTC" firstStartedPulling="2025-12-11 02:18:36.248740367 +0000 UTC m=+1057.937777746" lastFinishedPulling="2025-12-11 02:18:50.56837602 +0000 UTC m=+1072.257413399" observedRunningTime="2025-12-11 02:18:51.231250714 +0000 UTC m=+1072.920288093" watchObservedRunningTime="2025-12-11 02:18:51.237084349 +0000 UTC m=+1072.926121728" Dec 11 02:18:51 crc kubenswrapper[4824]: I1211 02:18:51.306771 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f4k8rk"] Dec 11 02:18:51 crc kubenswrapper[4824]: W1211 02:18:51.320861 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcb370fee_74a4_4f43_98a9_3df8c6e61335.slice/crio-3e62591f0bd84af6061cc28e647e77932099d124b1aacc043a562abdd5380f31 WatchSource:0}: Error finding container 3e62591f0bd84af6061cc28e647e77932099d124b1aacc043a562abdd5380f31: Status 404 returned error can't find the container with id 3e62591f0bd84af6061cc28e647e77932099d124b1aacc043a562abdd5380f31 Dec 11 02:18:52 crc kubenswrapper[4824]: I1211 02:18:52.229847 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f4k8rk" event={"ID":"cb370fee-74a4-4f43-98a9-3df8c6e61335","Type":"ContainerStarted","Data":"3e62591f0bd84af6061cc28e647e77932099d124b1aacc043a562abdd5380f31"} Dec 11 02:18:54 crc kubenswrapper[4824]: I1211 02:18:54.767232 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-r8tnj" Dec 11 02:18:54 crc kubenswrapper[4824]: I1211 02:18:54.785270 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-8hvxh" Dec 11 02:18:54 crc kubenswrapper[4824]: I1211 02:18:54.794801 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-4b2cb" Dec 11 02:18:54 crc kubenswrapper[4824]: I1211 02:18:54.808205 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-4cjld" Dec 11 02:18:54 crc kubenswrapper[4824]: I1211 02:18:54.828718 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-l9jrd" Dec 11 02:18:54 crc kubenswrapper[4824]: I1211 02:18:54.837294 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-5qp6v" Dec 11 02:18:54 crc kubenswrapper[4824]: I1211 02:18:54.883577 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-967d97867-5tpj5" Dec 11 02:18:54 crc kubenswrapper[4824]: I1211 02:18:54.919884 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-4np4x" Dec 11 02:18:55 crc kubenswrapper[4824]: I1211 02:18:55.123690 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-lmszx" Dec 11 02:18:55 crc kubenswrapper[4824]: I1211 02:18:55.135283 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-zkz72" Dec 11 02:18:55 crc kubenswrapper[4824]: I1211 02:18:55.370618 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5854674fcc-4zfbk" Dec 11 02:18:55 crc kubenswrapper[4824]: I1211 02:18:55.587392 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-qhhzn" Dec 11 02:18:55 crc kubenswrapper[4824]: I1211 02:18:55.617788 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-h6bgf" Dec 11 02:18:59 crc kubenswrapper[4824]: I1211 02:18:59.296412 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-qqpf2" event={"ID":"f1183bb7-356d-43b7-8ee7-55ab60dcbe54","Type":"ContainerStarted","Data":"aa90354bc0da84e81a876b3fb6c6767d43cd5717ca1bdc29c63bfbb03dc3df80"} Dec 11 02:18:59 crc kubenswrapper[4824]: I1211 02:18:59.300621 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-hpjng" event={"ID":"fd3368a6-77e2-4c8f-b535-1d602b794d7d","Type":"ContainerStarted","Data":"f7d0c53c956260fde9281f4593b11bff5657e4a8b73758954f616955b8626a25"} Dec 11 02:18:59 crc kubenswrapper[4824]: I1211 02:18:59.303481 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-r2bsg" event={"ID":"4d26d90a-05a8-4960-b2ca-12fa220383b7","Type":"ContainerStarted","Data":"397d47a200b9dfa2e56dc84522373898eb8b1484300ccca8478186a212d6280c"} Dec 11 02:18:59 crc kubenswrapper[4824]: I1211 02:18:59.305998 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-xx9gc" event={"ID":"10325d55-7efd-49e8-95e9-46d68c7cd7ae","Type":"ContainerStarted","Data":"82e7e894d53600b77f60b4a66164ba466ef5dfbba322255b22f6421c8cf016a9"} Dec 11 02:19:00 crc kubenswrapper[4824]: I1211 02:19:00.311733 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-wsjbh" event={"ID":"34bc4184-9cfb-4278-a9b9-80bf2f1ae612","Type":"ContainerStarted","Data":"fb7b7e3e0a7153e3c38fc0b60c6bbf885f6db8bebc5fb634a975919e6200206b"} Dec 11 02:19:00 crc kubenswrapper[4824]: I1211 02:19:00.313517 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-hpjng" event={"ID":"fd3368a6-77e2-4c8f-b535-1d602b794d7d","Type":"ContainerStarted","Data":"b5fe7a2f4a308297efaffc3141d2f6c0beee8b10b6eed5f4f5d391cd521bb2a7"} Dec 11 02:19:00 crc kubenswrapper[4824]: I1211 02:19:00.313608 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-hpjng" Dec 11 02:19:00 crc kubenswrapper[4824]: I1211 02:19:00.315337 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-r2bsg" event={"ID":"4d26d90a-05a8-4960-b2ca-12fa220383b7","Type":"ContainerStarted","Data":"d562db2f4a2cfa3eaa946eba1d886eb19d96caa9d91cfcf5eb6b23aee1809dff"} Dec 11 02:19:00 crc kubenswrapper[4824]: I1211 02:19:00.315413 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-r2bsg" Dec 11 02:19:00 crc kubenswrapper[4824]: I1211 02:19:00.316904 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f4k8rk" event={"ID":"cb370fee-74a4-4f43-98a9-3df8c6e61335","Type":"ContainerStarted","Data":"abbb433cc2bb648f9f43063de5cb6bdbebbdb3985d860731250d27d246f36493"} Dec 11 02:19:00 crc kubenswrapper[4824]: I1211 02:19:00.316931 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f4k8rk" event={"ID":"cb370fee-74a4-4f43-98a9-3df8c6e61335","Type":"ContainerStarted","Data":"ac6dd25afab8e90c78fa8fcdb8618c4de40720bd710155e138cfdf37859e2005"} Dec 11 02:19:00 crc kubenswrapper[4824]: I1211 02:19:00.316972 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f4k8rk" Dec 11 02:19:00 crc kubenswrapper[4824]: I1211 02:19:00.318495 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-w24mx" event={"ID":"46bf5c63-42b4-45bd-ade3-446a4b2e4f6e","Type":"ContainerStarted","Data":"c609126aa3f4675a9a77fdcb38941c0fac43074098d396e001e78a60b3f7ae64"} Dec 11 02:19:00 crc kubenswrapper[4824]: I1211 02:19:00.318527 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-w24mx" event={"ID":"46bf5c63-42b4-45bd-ade3-446a4b2e4f6e","Type":"ContainerStarted","Data":"548352a927acf9bf65c2a5d61f6ea486b7eb8aa47cd6141c5bd0c8288bc4d54f"} Dec 11 02:19:00 crc kubenswrapper[4824]: I1211 02:19:00.318645 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-w24mx" Dec 11 02:19:00 crc kubenswrapper[4824]: I1211 02:19:00.320326 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-xx9gc" event={"ID":"10325d55-7efd-49e8-95e9-46d68c7cd7ae","Type":"ContainerStarted","Data":"0205601135fbd885701e4ebc190e362ff5f1843a9c639b5afd3f321b9f77f93e"} Dec 11 02:19:00 crc kubenswrapper[4824]: I1211 02:19:00.320440 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-78f8948974-xx9gc" Dec 11 02:19:00 crc kubenswrapper[4824]: I1211 02:19:00.321923 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-fmvm7" event={"ID":"8047af9c-3a6d-4b2e-a088-df967e42671c","Type":"ContainerStarted","Data":"200241f7358430c212dfe55546a2b525616570bfe98a6cda090a0d860291e529"} Dec 11 02:19:00 crc kubenswrapper[4824]: I1211 02:19:00.321946 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-fmvm7" event={"ID":"8047af9c-3a6d-4b2e-a088-df967e42671c","Type":"ContainerStarted","Data":"5f0bd36f21bbf5e44336a8bc009a2ea76c41e13237212e2beba5d49e643c56e2"} Dec 11 02:19:00 crc kubenswrapper[4824]: I1211 02:19:00.322103 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-fmvm7" Dec 11 02:19:00 crc kubenswrapper[4824]: I1211 02:19:00.353530 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-qqpf2" event={"ID":"f1183bb7-356d-43b7-8ee7-55ab60dcbe54","Type":"ContainerStarted","Data":"6a16ac41160ae3a889aa4a04046518c16829969efaf5493053f4497df7e81996"} Dec 11 02:19:00 crc kubenswrapper[4824]: I1211 02:19:00.353935 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-998648c74-qqpf2" Dec 11 02:19:00 crc kubenswrapper[4824]: I1211 02:19:00.362855 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-wsjbh" podStartSLOduration=2.700777936 podStartE2EDuration="25.362837992s" podCreationTimestamp="2025-12-11 02:18:35 +0000 UTC" firstStartedPulling="2025-12-11 02:18:36.163677901 +0000 UTC m=+1057.852715280" lastFinishedPulling="2025-12-11 02:18:58.825737957 +0000 UTC m=+1080.514775336" observedRunningTime="2025-12-11 02:19:00.338015621 +0000 UTC m=+1082.027053000" watchObservedRunningTime="2025-12-11 02:19:00.362837992 +0000 UTC m=+1082.051875371" Dec 11 02:19:00 crc kubenswrapper[4824]: I1211 02:19:00.371952 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-78f8948974-xx9gc" podStartSLOduration=7.751896503 podStartE2EDuration="26.371936564s" podCreationTimestamp="2025-12-11 02:18:34 +0000 UTC" firstStartedPulling="2025-12-11 02:18:36.194251675 +0000 UTC m=+1057.883289054" lastFinishedPulling="2025-12-11 02:18:54.814291726 +0000 UTC m=+1076.503329115" observedRunningTime="2025-12-11 02:19:00.367507356 +0000 UTC m=+1082.056544735" watchObservedRunningTime="2025-12-11 02:19:00.371936564 +0000 UTC m=+1082.060973943" Dec 11 02:19:00 crc kubenswrapper[4824]: I1211 02:19:00.400340 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-hpjng" podStartSLOduration=7.512237221 podStartE2EDuration="26.40032609s" podCreationTimestamp="2025-12-11 02:18:34 +0000 UTC" firstStartedPulling="2025-12-11 02:18:35.932126825 +0000 UTC m=+1057.621164204" lastFinishedPulling="2025-12-11 02:18:54.820215664 +0000 UTC m=+1076.509253073" observedRunningTime="2025-12-11 02:19:00.394902496 +0000 UTC m=+1082.083939875" watchObservedRunningTime="2025-12-11 02:19:00.40032609 +0000 UTC m=+1082.089363469" Dec 11 02:19:00 crc kubenswrapper[4824]: I1211 02:19:00.412848 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-fmvm7" podStartSLOduration=3.799742732 podStartE2EDuration="26.412836964s" podCreationTimestamp="2025-12-11 02:18:34 +0000 UTC" firstStartedPulling="2025-12-11 02:18:36.163085975 +0000 UTC m=+1057.852123354" lastFinishedPulling="2025-12-11 02:18:58.776180207 +0000 UTC m=+1080.465217586" observedRunningTime="2025-12-11 02:19:00.41121332 +0000 UTC m=+1082.100250699" watchObservedRunningTime="2025-12-11 02:19:00.412836964 +0000 UTC m=+1082.101874343" Dec 11 02:19:00 crc kubenswrapper[4824]: I1211 02:19:00.427598 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-w24mx" podStartSLOduration=3.544603618 podStartE2EDuration="26.427579276s" podCreationTimestamp="2025-12-11 02:18:34 +0000 UTC" firstStartedPulling="2025-12-11 02:18:35.928894909 +0000 UTC m=+1057.617932288" lastFinishedPulling="2025-12-11 02:18:58.811870527 +0000 UTC m=+1080.500907946" observedRunningTime="2025-12-11 02:19:00.426849327 +0000 UTC m=+1082.115886716" watchObservedRunningTime="2025-12-11 02:19:00.427579276 +0000 UTC m=+1082.116616655" Dec 11 02:19:00 crc kubenswrapper[4824]: I1211 02:19:00.460028 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f4k8rk" podStartSLOduration=19.00726136 podStartE2EDuration="26.460002129s" podCreationTimestamp="2025-12-11 02:18:34 +0000 UTC" firstStartedPulling="2025-12-11 02:18:51.323949332 +0000 UTC m=+1073.012986721" lastFinishedPulling="2025-12-11 02:18:58.776690091 +0000 UTC m=+1080.465727490" observedRunningTime="2025-12-11 02:19:00.454932705 +0000 UTC m=+1082.143970084" watchObservedRunningTime="2025-12-11 02:19:00.460002129 +0000 UTC m=+1082.149039508" Dec 11 02:19:00 crc kubenswrapper[4824]: I1211 02:19:00.502514 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-r2bsg" podStartSLOduration=3.910686847 podStartE2EDuration="26.502488642s" podCreationTimestamp="2025-12-11 02:18:34 +0000 UTC" firstStartedPulling="2025-12-11 02:18:36.196746682 +0000 UTC m=+1057.885784061" lastFinishedPulling="2025-12-11 02:18:58.788548447 +0000 UTC m=+1080.477585856" observedRunningTime="2025-12-11 02:19:00.495421493 +0000 UTC m=+1082.184458872" watchObservedRunningTime="2025-12-11 02:19:00.502488642 +0000 UTC m=+1082.191526021" Dec 11 02:19:00 crc kubenswrapper[4824]: I1211 02:19:00.515319 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-998648c74-qqpf2" podStartSLOduration=7.830106997 podStartE2EDuration="26.515305302s" podCreationTimestamp="2025-12-11 02:18:34 +0000 UTC" firstStartedPulling="2025-12-11 02:18:36.126148482 +0000 UTC m=+1057.815185861" lastFinishedPulling="2025-12-11 02:18:54.811346777 +0000 UTC m=+1076.500384166" observedRunningTime="2025-12-11 02:19:00.514523442 +0000 UTC m=+1082.203560821" watchObservedRunningTime="2025-12-11 02:19:00.515305302 +0000 UTC m=+1082.204342671" Dec 11 02:19:05 crc kubenswrapper[4824]: I1211 02:19:05.077416 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-w24mx" Dec 11 02:19:05 crc kubenswrapper[4824]: I1211 02:19:05.094390 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-hpjng" Dec 11 02:19:05 crc kubenswrapper[4824]: I1211 02:19:05.146278 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-998648c74-qqpf2" Dec 11 02:19:05 crc kubenswrapper[4824]: I1211 02:19:05.256311 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-fmvm7" Dec 11 02:19:05 crc kubenswrapper[4824]: I1211 02:19:05.479955 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-r2bsg" Dec 11 02:19:05 crc kubenswrapper[4824]: I1211 02:19:05.515731 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-78f8948974-xx9gc" Dec 11 02:19:06 crc kubenswrapper[4824]: I1211 02:19:06.376879 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f219b9bf-34b1-48c4-96c8-20580b6ec9a4-cert\") pod \"infra-operator-controller-manager-78d48bff9d-j68mf\" (UID: \"f219b9bf-34b1-48c4-96c8-20580b6ec9a4\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-j68mf" Dec 11 02:19:06 crc kubenswrapper[4824]: I1211 02:19:06.386400 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f219b9bf-34b1-48c4-96c8-20580b6ec9a4-cert\") pod \"infra-operator-controller-manager-78d48bff9d-j68mf\" (UID: \"f219b9bf-34b1-48c4-96c8-20580b6ec9a4\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-j68mf" Dec 11 02:19:06 crc kubenswrapper[4824]: I1211 02:19:06.661696 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-dxbpz" Dec 11 02:19:06 crc kubenswrapper[4824]: I1211 02:19:06.669754 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-j68mf" Dec 11 02:19:07 crc kubenswrapper[4824]: I1211 02:19:07.128276 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/7180c14a-0c3a-4267-b122-5c30c5685d60-webhook-certs\") pod \"openstack-operator-controller-manager-56cf947496-nhjjr\" (UID: \"7180c14a-0c3a-4267-b122-5c30c5685d60\") " pod="openstack-operators/openstack-operator-controller-manager-56cf947496-nhjjr" Dec 11 02:19:07 crc kubenswrapper[4824]: I1211 02:19:07.128415 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7180c14a-0c3a-4267-b122-5c30c5685d60-metrics-certs\") pod \"openstack-operator-controller-manager-56cf947496-nhjjr\" (UID: \"7180c14a-0c3a-4267-b122-5c30c5685d60\") " pod="openstack-operators/openstack-operator-controller-manager-56cf947496-nhjjr" Dec 11 02:19:07 crc kubenswrapper[4824]: I1211 02:19:07.136501 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/7180c14a-0c3a-4267-b122-5c30c5685d60-webhook-certs\") pod \"openstack-operator-controller-manager-56cf947496-nhjjr\" (UID: \"7180c14a-0c3a-4267-b122-5c30c5685d60\") " pod="openstack-operators/openstack-operator-controller-manager-56cf947496-nhjjr" Dec 11 02:19:07 crc kubenswrapper[4824]: I1211 02:19:07.139396 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7180c14a-0c3a-4267-b122-5c30c5685d60-metrics-certs\") pod \"openstack-operator-controller-manager-56cf947496-nhjjr\" (UID: \"7180c14a-0c3a-4267-b122-5c30c5685d60\") " pod="openstack-operators/openstack-operator-controller-manager-56cf947496-nhjjr" Dec 11 02:19:07 crc kubenswrapper[4824]: I1211 02:19:07.164658 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-78d48bff9d-j68mf"] Dec 11 02:19:07 crc kubenswrapper[4824]: I1211 02:19:07.165580 4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 11 02:19:07 crc kubenswrapper[4824]: I1211 02:19:07.171408 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-hqsm4" Dec 11 02:19:07 crc kubenswrapper[4824]: I1211 02:19:07.179242 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-56cf947496-nhjjr" Dec 11 02:19:07 crc kubenswrapper[4824]: I1211 02:19:07.421618 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-j68mf" event={"ID":"f219b9bf-34b1-48c4-96c8-20580b6ec9a4","Type":"ContainerStarted","Data":"dc0fa64049e14e55999e5b7164e60d4a361dce3079c648b6c0465c3605109d9c"} Dec 11 02:19:07 crc kubenswrapper[4824]: I1211 02:19:07.672963 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-56cf947496-nhjjr"] Dec 11 02:19:08 crc kubenswrapper[4824]: I1211 02:19:08.438548 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-56cf947496-nhjjr" event={"ID":"7180c14a-0c3a-4267-b122-5c30c5685d60","Type":"ContainerStarted","Data":"3c2998988c3314866c9435b4c614a118a25a7fa0629f5906e4abe4be0bf86073"} Dec 11 02:19:10 crc kubenswrapper[4824]: I1211 02:19:10.456528 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-56cf947496-nhjjr" event={"ID":"7180c14a-0c3a-4267-b122-5c30c5685d60","Type":"ContainerStarted","Data":"0f2793df1c87178e578d3871df0dc12a0047c71c38f640d16211b7dee69789c3"} Dec 11 02:19:10 crc kubenswrapper[4824]: I1211 02:19:10.456991 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-56cf947496-nhjjr" Dec 11 02:19:10 crc kubenswrapper[4824]: I1211 02:19:10.502534 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-56cf947496-nhjjr" podStartSLOduration=35.502512308 podStartE2EDuration="35.502512308s" podCreationTimestamp="2025-12-11 02:18:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:19:10.497237277 +0000 UTC m=+1092.186274666" watchObservedRunningTime="2025-12-11 02:19:10.502512308 +0000 UTC m=+1092.191549687" Dec 11 02:19:11 crc kubenswrapper[4824]: I1211 02:19:11.073364 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f4k8rk" Dec 11 02:19:16 crc kubenswrapper[4824]: I1211 02:19:16.511466 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-j68mf" event={"ID":"f219b9bf-34b1-48c4-96c8-20580b6ec9a4","Type":"ContainerStarted","Data":"4ea2570333d9a80ffae68e06738cb0ae641861086801e09f75552546cae53970"} Dec 11 02:19:16 crc kubenswrapper[4824]: I1211 02:19:16.512163 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-j68mf" Dec 11 02:19:16 crc kubenswrapper[4824]: I1211 02:19:16.512185 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-j68mf" event={"ID":"f219b9bf-34b1-48c4-96c8-20580b6ec9a4","Type":"ContainerStarted","Data":"9502611fa1d806ccc9f588e5693f97074aade2aa8140eb5bd7fd22df2964f7fa"} Dec 11 02:19:16 crc kubenswrapper[4824]: I1211 02:19:16.554861 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-j68mf" podStartSLOduration=34.149581285 podStartE2EDuration="42.554837941s" podCreationTimestamp="2025-12-11 02:18:34 +0000 UTC" firstStartedPulling="2025-12-11 02:19:07.165192509 +0000 UTC m=+1088.854229928" lastFinishedPulling="2025-12-11 02:19:15.570449195 +0000 UTC m=+1097.259486584" observedRunningTime="2025-12-11 02:19:16.538030494 +0000 UTC m=+1098.227067913" watchObservedRunningTime="2025-12-11 02:19:16.554837941 +0000 UTC m=+1098.243875330" Dec 11 02:19:17 crc kubenswrapper[4824]: I1211 02:19:17.184498 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-56cf947496-nhjjr" Dec 11 02:19:26 crc kubenswrapper[4824]: I1211 02:19:26.680074 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-j68mf" Dec 11 02:19:37 crc kubenswrapper[4824]: I1211 02:19:37.251577 4824 patch_prober.go:28] interesting pod/machine-config-daemon-gx6xt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 02:19:37 crc kubenswrapper[4824]: I1211 02:19:37.252460 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 02:19:55 crc kubenswrapper[4824]: I1211 02:19:55.907458 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-qwcd8"] Dec 11 02:19:55 crc kubenswrapper[4824]: I1211 02:19:55.909829 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-qwcd8" Dec 11 02:19:55 crc kubenswrapper[4824]: I1211 02:19:55.913737 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-s9fl4" Dec 11 02:19:55 crc kubenswrapper[4824]: I1211 02:19:55.913932 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Dec 11 02:19:55 crc kubenswrapper[4824]: I1211 02:19:55.914047 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Dec 11 02:19:55 crc kubenswrapper[4824]: I1211 02:19:55.914282 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Dec 11 02:19:55 crc kubenswrapper[4824]: I1211 02:19:55.929021 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-qwcd8"] Dec 11 02:19:55 crc kubenswrapper[4824]: I1211 02:19:55.932594 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tvn4n\" (UniqueName: \"kubernetes.io/projected/1081fac2-0512-4e5a-b05e-3999af5d4159-kube-api-access-tvn4n\") pod \"dnsmasq-dns-675f4bcbfc-qwcd8\" (UID: \"1081fac2-0512-4e5a-b05e-3999af5d4159\") " pod="openstack/dnsmasq-dns-675f4bcbfc-qwcd8" Dec 11 02:19:55 crc kubenswrapper[4824]: I1211 02:19:55.932638 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1081fac2-0512-4e5a-b05e-3999af5d4159-config\") pod \"dnsmasq-dns-675f4bcbfc-qwcd8\" (UID: \"1081fac2-0512-4e5a-b05e-3999af5d4159\") " pod="openstack/dnsmasq-dns-675f4bcbfc-qwcd8" Dec 11 02:19:55 crc kubenswrapper[4824]: I1211 02:19:55.997267 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-xhg5m"] Dec 11 02:19:55 crc kubenswrapper[4824]: I1211 02:19:55.998544 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-xhg5m" Dec 11 02:19:56 crc kubenswrapper[4824]: I1211 02:19:56.002041 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Dec 11 02:19:56 crc kubenswrapper[4824]: I1211 02:19:56.017407 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-xhg5m"] Dec 11 02:19:56 crc kubenswrapper[4824]: I1211 02:19:56.034250 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hmc5m\" (UniqueName: \"kubernetes.io/projected/e319794c-b4f6-4ae6-b4fa-746aa4028edd-kube-api-access-hmc5m\") pod \"dnsmasq-dns-78dd6ddcc-xhg5m\" (UID: \"e319794c-b4f6-4ae6-b4fa-746aa4028edd\") " pod="openstack/dnsmasq-dns-78dd6ddcc-xhg5m" Dec 11 02:19:56 crc kubenswrapper[4824]: I1211 02:19:56.034496 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e319794c-b4f6-4ae6-b4fa-746aa4028edd-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-xhg5m\" (UID: \"e319794c-b4f6-4ae6-b4fa-746aa4028edd\") " pod="openstack/dnsmasq-dns-78dd6ddcc-xhg5m" Dec 11 02:19:56 crc kubenswrapper[4824]: I1211 02:19:56.034589 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e319794c-b4f6-4ae6-b4fa-746aa4028edd-config\") pod \"dnsmasq-dns-78dd6ddcc-xhg5m\" (UID: \"e319794c-b4f6-4ae6-b4fa-746aa4028edd\") " pod="openstack/dnsmasq-dns-78dd6ddcc-xhg5m" Dec 11 02:19:56 crc kubenswrapper[4824]: I1211 02:19:56.034685 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tvn4n\" (UniqueName: \"kubernetes.io/projected/1081fac2-0512-4e5a-b05e-3999af5d4159-kube-api-access-tvn4n\") pod \"dnsmasq-dns-675f4bcbfc-qwcd8\" (UID: \"1081fac2-0512-4e5a-b05e-3999af5d4159\") " pod="openstack/dnsmasq-dns-675f4bcbfc-qwcd8" Dec 11 02:19:56 crc kubenswrapper[4824]: I1211 02:19:56.034761 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1081fac2-0512-4e5a-b05e-3999af5d4159-config\") pod \"dnsmasq-dns-675f4bcbfc-qwcd8\" (UID: \"1081fac2-0512-4e5a-b05e-3999af5d4159\") " pod="openstack/dnsmasq-dns-675f4bcbfc-qwcd8" Dec 11 02:19:56 crc kubenswrapper[4824]: I1211 02:19:56.035615 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1081fac2-0512-4e5a-b05e-3999af5d4159-config\") pod \"dnsmasq-dns-675f4bcbfc-qwcd8\" (UID: \"1081fac2-0512-4e5a-b05e-3999af5d4159\") " pod="openstack/dnsmasq-dns-675f4bcbfc-qwcd8" Dec 11 02:19:56 crc kubenswrapper[4824]: I1211 02:19:56.058959 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tvn4n\" (UniqueName: \"kubernetes.io/projected/1081fac2-0512-4e5a-b05e-3999af5d4159-kube-api-access-tvn4n\") pod \"dnsmasq-dns-675f4bcbfc-qwcd8\" (UID: \"1081fac2-0512-4e5a-b05e-3999af5d4159\") " pod="openstack/dnsmasq-dns-675f4bcbfc-qwcd8" Dec 11 02:19:56 crc kubenswrapper[4824]: I1211 02:19:56.136018 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hmc5m\" (UniqueName: \"kubernetes.io/projected/e319794c-b4f6-4ae6-b4fa-746aa4028edd-kube-api-access-hmc5m\") pod \"dnsmasq-dns-78dd6ddcc-xhg5m\" (UID: \"e319794c-b4f6-4ae6-b4fa-746aa4028edd\") " pod="openstack/dnsmasq-dns-78dd6ddcc-xhg5m" Dec 11 02:19:56 crc kubenswrapper[4824]: I1211 02:19:56.136101 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e319794c-b4f6-4ae6-b4fa-746aa4028edd-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-xhg5m\" (UID: \"e319794c-b4f6-4ae6-b4fa-746aa4028edd\") " pod="openstack/dnsmasq-dns-78dd6ddcc-xhg5m" Dec 11 02:19:56 crc kubenswrapper[4824]: I1211 02:19:56.136207 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e319794c-b4f6-4ae6-b4fa-746aa4028edd-config\") pod \"dnsmasq-dns-78dd6ddcc-xhg5m\" (UID: \"e319794c-b4f6-4ae6-b4fa-746aa4028edd\") " pod="openstack/dnsmasq-dns-78dd6ddcc-xhg5m" Dec 11 02:19:56 crc kubenswrapper[4824]: I1211 02:19:56.137726 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e319794c-b4f6-4ae6-b4fa-746aa4028edd-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-xhg5m\" (UID: \"e319794c-b4f6-4ae6-b4fa-746aa4028edd\") " pod="openstack/dnsmasq-dns-78dd6ddcc-xhg5m" Dec 11 02:19:56 crc kubenswrapper[4824]: I1211 02:19:56.140540 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e319794c-b4f6-4ae6-b4fa-746aa4028edd-config\") pod \"dnsmasq-dns-78dd6ddcc-xhg5m\" (UID: \"e319794c-b4f6-4ae6-b4fa-746aa4028edd\") " pod="openstack/dnsmasq-dns-78dd6ddcc-xhg5m" Dec 11 02:19:56 crc kubenswrapper[4824]: I1211 02:19:56.153846 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hmc5m\" (UniqueName: \"kubernetes.io/projected/e319794c-b4f6-4ae6-b4fa-746aa4028edd-kube-api-access-hmc5m\") pod \"dnsmasq-dns-78dd6ddcc-xhg5m\" (UID: \"e319794c-b4f6-4ae6-b4fa-746aa4028edd\") " pod="openstack/dnsmasq-dns-78dd6ddcc-xhg5m" Dec 11 02:19:56 crc kubenswrapper[4824]: I1211 02:19:56.523002 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-xhg5m" Dec 11 02:19:56 crc kubenswrapper[4824]: I1211 02:19:56.523962 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-qwcd8" Dec 11 02:19:57 crc kubenswrapper[4824]: I1211 02:19:57.070638 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-qwcd8"] Dec 11 02:19:57 crc kubenswrapper[4824]: I1211 02:19:57.123834 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-xhg5m"] Dec 11 02:19:57 crc kubenswrapper[4824]: W1211 02:19:57.131276 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode319794c_b4f6_4ae6_b4fa_746aa4028edd.slice/crio-e5f48fa2892da280d618b5b769e86b9593ae9b364c473dae883691151a5d57ab WatchSource:0}: Error finding container e5f48fa2892da280d618b5b769e86b9593ae9b364c473dae883691151a5d57ab: Status 404 returned error can't find the container with id e5f48fa2892da280d618b5b769e86b9593ae9b364c473dae883691151a5d57ab Dec 11 02:19:57 crc kubenswrapper[4824]: I1211 02:19:57.870910 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-qwcd8" event={"ID":"1081fac2-0512-4e5a-b05e-3999af5d4159","Type":"ContainerStarted","Data":"16e94f0a27abffff6837104803bf49a14f2d372f10aa85ba1b8ffc0b35b9c964"} Dec 11 02:19:57 crc kubenswrapper[4824]: I1211 02:19:57.872362 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-xhg5m" event={"ID":"e319794c-b4f6-4ae6-b4fa-746aa4028edd","Type":"ContainerStarted","Data":"e5f48fa2892da280d618b5b769e86b9593ae9b364c473dae883691151a5d57ab"} Dec 11 02:19:58 crc kubenswrapper[4824]: I1211 02:19:58.936855 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-qwcd8"] Dec 11 02:19:58 crc kubenswrapper[4824]: I1211 02:19:58.981777 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-r8gdk"] Dec 11 02:19:58 crc kubenswrapper[4824]: I1211 02:19:58.982994 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-r8gdk" Dec 11 02:19:58 crc kubenswrapper[4824]: I1211 02:19:58.997130 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-r8gdk"] Dec 11 02:19:59 crc kubenswrapper[4824]: I1211 02:19:59.089130 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0b407c6b-fc23-459c-8c39-5a13c9889de5-dns-svc\") pod \"dnsmasq-dns-666b6646f7-r8gdk\" (UID: \"0b407c6b-fc23-459c-8c39-5a13c9889de5\") " pod="openstack/dnsmasq-dns-666b6646f7-r8gdk" Dec 11 02:19:59 crc kubenswrapper[4824]: I1211 02:19:59.089231 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b407c6b-fc23-459c-8c39-5a13c9889de5-config\") pod \"dnsmasq-dns-666b6646f7-r8gdk\" (UID: \"0b407c6b-fc23-459c-8c39-5a13c9889de5\") " pod="openstack/dnsmasq-dns-666b6646f7-r8gdk" Dec 11 02:19:59 crc kubenswrapper[4824]: I1211 02:19:59.089265 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-766fx\" (UniqueName: \"kubernetes.io/projected/0b407c6b-fc23-459c-8c39-5a13c9889de5-kube-api-access-766fx\") pod \"dnsmasq-dns-666b6646f7-r8gdk\" (UID: \"0b407c6b-fc23-459c-8c39-5a13c9889de5\") " pod="openstack/dnsmasq-dns-666b6646f7-r8gdk" Dec 11 02:19:59 crc kubenswrapper[4824]: I1211 02:19:59.191001 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0b407c6b-fc23-459c-8c39-5a13c9889de5-dns-svc\") pod \"dnsmasq-dns-666b6646f7-r8gdk\" (UID: \"0b407c6b-fc23-459c-8c39-5a13c9889de5\") " pod="openstack/dnsmasq-dns-666b6646f7-r8gdk" Dec 11 02:19:59 crc kubenswrapper[4824]: I1211 02:19:59.191099 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b407c6b-fc23-459c-8c39-5a13c9889de5-config\") pod \"dnsmasq-dns-666b6646f7-r8gdk\" (UID: \"0b407c6b-fc23-459c-8c39-5a13c9889de5\") " pod="openstack/dnsmasq-dns-666b6646f7-r8gdk" Dec 11 02:19:59 crc kubenswrapper[4824]: I1211 02:19:59.191155 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-766fx\" (UniqueName: \"kubernetes.io/projected/0b407c6b-fc23-459c-8c39-5a13c9889de5-kube-api-access-766fx\") pod \"dnsmasq-dns-666b6646f7-r8gdk\" (UID: \"0b407c6b-fc23-459c-8c39-5a13c9889de5\") " pod="openstack/dnsmasq-dns-666b6646f7-r8gdk" Dec 11 02:19:59 crc kubenswrapper[4824]: I1211 02:19:59.192087 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0b407c6b-fc23-459c-8c39-5a13c9889de5-dns-svc\") pod \"dnsmasq-dns-666b6646f7-r8gdk\" (UID: \"0b407c6b-fc23-459c-8c39-5a13c9889de5\") " pod="openstack/dnsmasq-dns-666b6646f7-r8gdk" Dec 11 02:19:59 crc kubenswrapper[4824]: I1211 02:19:59.193200 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b407c6b-fc23-459c-8c39-5a13c9889de5-config\") pod \"dnsmasq-dns-666b6646f7-r8gdk\" (UID: \"0b407c6b-fc23-459c-8c39-5a13c9889de5\") " pod="openstack/dnsmasq-dns-666b6646f7-r8gdk" Dec 11 02:19:59 crc kubenswrapper[4824]: I1211 02:19:59.195226 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-xhg5m"] Dec 11 02:19:59 crc kubenswrapper[4824]: I1211 02:19:59.220364 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-766fx\" (UniqueName: \"kubernetes.io/projected/0b407c6b-fc23-459c-8c39-5a13c9889de5-kube-api-access-766fx\") pod \"dnsmasq-dns-666b6646f7-r8gdk\" (UID: \"0b407c6b-fc23-459c-8c39-5a13c9889de5\") " pod="openstack/dnsmasq-dns-666b6646f7-r8gdk" Dec 11 02:19:59 crc kubenswrapper[4824]: I1211 02:19:59.234449 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-8bq9w"] Dec 11 02:19:59 crc kubenswrapper[4824]: I1211 02:19:59.235902 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-8bq9w" Dec 11 02:19:59 crc kubenswrapper[4824]: I1211 02:19:59.242962 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-8bq9w"] Dec 11 02:19:59 crc kubenswrapper[4824]: I1211 02:19:59.292065 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d72d278a-3895-4b90-abed-0905680540c8-config\") pod \"dnsmasq-dns-57d769cc4f-8bq9w\" (UID: \"d72d278a-3895-4b90-abed-0905680540c8\") " pod="openstack/dnsmasq-dns-57d769cc4f-8bq9w" Dec 11 02:19:59 crc kubenswrapper[4824]: I1211 02:19:59.292133 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z45c9\" (UniqueName: \"kubernetes.io/projected/d72d278a-3895-4b90-abed-0905680540c8-kube-api-access-z45c9\") pod \"dnsmasq-dns-57d769cc4f-8bq9w\" (UID: \"d72d278a-3895-4b90-abed-0905680540c8\") " pod="openstack/dnsmasq-dns-57d769cc4f-8bq9w" Dec 11 02:19:59 crc kubenswrapper[4824]: I1211 02:19:59.292175 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d72d278a-3895-4b90-abed-0905680540c8-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-8bq9w\" (UID: \"d72d278a-3895-4b90-abed-0905680540c8\") " pod="openstack/dnsmasq-dns-57d769cc4f-8bq9w" Dec 11 02:19:59 crc kubenswrapper[4824]: I1211 02:19:59.310792 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-r8gdk" Dec 11 02:19:59 crc kubenswrapper[4824]: I1211 02:19:59.393936 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d72d278a-3895-4b90-abed-0905680540c8-config\") pod \"dnsmasq-dns-57d769cc4f-8bq9w\" (UID: \"d72d278a-3895-4b90-abed-0905680540c8\") " pod="openstack/dnsmasq-dns-57d769cc4f-8bq9w" Dec 11 02:19:59 crc kubenswrapper[4824]: I1211 02:19:59.393980 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z45c9\" (UniqueName: \"kubernetes.io/projected/d72d278a-3895-4b90-abed-0905680540c8-kube-api-access-z45c9\") pod \"dnsmasq-dns-57d769cc4f-8bq9w\" (UID: \"d72d278a-3895-4b90-abed-0905680540c8\") " pod="openstack/dnsmasq-dns-57d769cc4f-8bq9w" Dec 11 02:19:59 crc kubenswrapper[4824]: I1211 02:19:59.394024 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d72d278a-3895-4b90-abed-0905680540c8-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-8bq9w\" (UID: \"d72d278a-3895-4b90-abed-0905680540c8\") " pod="openstack/dnsmasq-dns-57d769cc4f-8bq9w" Dec 11 02:19:59 crc kubenswrapper[4824]: I1211 02:19:59.394852 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d72d278a-3895-4b90-abed-0905680540c8-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-8bq9w\" (UID: \"d72d278a-3895-4b90-abed-0905680540c8\") " pod="openstack/dnsmasq-dns-57d769cc4f-8bq9w" Dec 11 02:19:59 crc kubenswrapper[4824]: I1211 02:19:59.395315 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d72d278a-3895-4b90-abed-0905680540c8-config\") pod \"dnsmasq-dns-57d769cc4f-8bq9w\" (UID: \"d72d278a-3895-4b90-abed-0905680540c8\") " pod="openstack/dnsmasq-dns-57d769cc4f-8bq9w" Dec 11 02:19:59 crc kubenswrapper[4824]: I1211 02:19:59.420453 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z45c9\" (UniqueName: \"kubernetes.io/projected/d72d278a-3895-4b90-abed-0905680540c8-kube-api-access-z45c9\") pod \"dnsmasq-dns-57d769cc4f-8bq9w\" (UID: \"d72d278a-3895-4b90-abed-0905680540c8\") " pod="openstack/dnsmasq-dns-57d769cc4f-8bq9w" Dec 11 02:19:59 crc kubenswrapper[4824]: I1211 02:19:59.564981 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-8bq9w" Dec 11 02:19:59 crc kubenswrapper[4824]: I1211 02:19:59.809842 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-r8gdk"] Dec 11 02:19:59 crc kubenswrapper[4824]: W1211 02:19:59.822847 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0b407c6b_fc23_459c_8c39_5a13c9889de5.slice/crio-742e56327af98155873a8a08f312d8b5917a1c51667dc827b44e6569bc8299ee WatchSource:0}: Error finding container 742e56327af98155873a8a08f312d8b5917a1c51667dc827b44e6569bc8299ee: Status 404 returned error can't find the container with id 742e56327af98155873a8a08f312d8b5917a1c51667dc827b44e6569bc8299ee Dec 11 02:19:59 crc kubenswrapper[4824]: I1211 02:19:59.905316 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-r8gdk" event={"ID":"0b407c6b-fc23-459c-8c39-5a13c9889de5","Type":"ContainerStarted","Data":"742e56327af98155873a8a08f312d8b5917a1c51667dc827b44e6569bc8299ee"} Dec 11 02:20:00 crc kubenswrapper[4824]: W1211 02:20:00.008148 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd72d278a_3895_4b90_abed_0905680540c8.slice/crio-864eba193e2630c696553c6d5038d28695e3aaf1224f61c024446584b949fbb9 WatchSource:0}: Error finding container 864eba193e2630c696553c6d5038d28695e3aaf1224f61c024446584b949fbb9: Status 404 returned error can't find the container with id 864eba193e2630c696553c6d5038d28695e3aaf1224f61c024446584b949fbb9 Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.009878 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-8bq9w"] Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.106596 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.110371 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.113367 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.114460 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.119447 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.120163 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.120294 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.120314 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.120461 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.121767 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-vtbb9" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.308054 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1fd840b6-4ace-45b8-a8d2-38a041110095-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"1fd840b6-4ace-45b8-a8d2-38a041110095\") " pod="openstack/rabbitmq-server-0" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.308164 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"1fd840b6-4ace-45b8-a8d2-38a041110095\") " pod="openstack/rabbitmq-server-0" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.308195 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1fd840b6-4ace-45b8-a8d2-38a041110095-server-conf\") pod \"rabbitmq-server-0\" (UID: \"1fd840b6-4ace-45b8-a8d2-38a041110095\") " pod="openstack/rabbitmq-server-0" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.308236 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1fd840b6-4ace-45b8-a8d2-38a041110095-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"1fd840b6-4ace-45b8-a8d2-38a041110095\") " pod="openstack/rabbitmq-server-0" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.308261 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1fd840b6-4ace-45b8-a8d2-38a041110095-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"1fd840b6-4ace-45b8-a8d2-38a041110095\") " pod="openstack/rabbitmq-server-0" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.308285 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1fd840b6-4ace-45b8-a8d2-38a041110095-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"1fd840b6-4ace-45b8-a8d2-38a041110095\") " pod="openstack/rabbitmq-server-0" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.308353 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1fd840b6-4ace-45b8-a8d2-38a041110095-config-data\") pod \"rabbitmq-server-0\" (UID: \"1fd840b6-4ace-45b8-a8d2-38a041110095\") " pod="openstack/rabbitmq-server-0" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.308369 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dnhpc\" (UniqueName: \"kubernetes.io/projected/1fd840b6-4ace-45b8-a8d2-38a041110095-kube-api-access-dnhpc\") pod \"rabbitmq-server-0\" (UID: \"1fd840b6-4ace-45b8-a8d2-38a041110095\") " pod="openstack/rabbitmq-server-0" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.308396 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1fd840b6-4ace-45b8-a8d2-38a041110095-pod-info\") pod \"rabbitmq-server-0\" (UID: \"1fd840b6-4ace-45b8-a8d2-38a041110095\") " pod="openstack/rabbitmq-server-0" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.308421 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1fd840b6-4ace-45b8-a8d2-38a041110095-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"1fd840b6-4ace-45b8-a8d2-38a041110095\") " pod="openstack/rabbitmq-server-0" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.309193 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1fd840b6-4ace-45b8-a8d2-38a041110095-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"1fd840b6-4ace-45b8-a8d2-38a041110095\") " pod="openstack/rabbitmq-server-0" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.414429 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1fd840b6-4ace-45b8-a8d2-38a041110095-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"1fd840b6-4ace-45b8-a8d2-38a041110095\") " pod="openstack/rabbitmq-server-0" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.414719 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"1fd840b6-4ace-45b8-a8d2-38a041110095\") " pod="openstack/rabbitmq-server-0" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.414747 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1fd840b6-4ace-45b8-a8d2-38a041110095-server-conf\") pod \"rabbitmq-server-0\" (UID: \"1fd840b6-4ace-45b8-a8d2-38a041110095\") " pod="openstack/rabbitmq-server-0" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.414765 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1fd840b6-4ace-45b8-a8d2-38a041110095-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"1fd840b6-4ace-45b8-a8d2-38a041110095\") " pod="openstack/rabbitmq-server-0" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.414791 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1fd840b6-4ace-45b8-a8d2-38a041110095-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"1fd840b6-4ace-45b8-a8d2-38a041110095\") " pod="openstack/rabbitmq-server-0" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.414813 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1fd840b6-4ace-45b8-a8d2-38a041110095-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"1fd840b6-4ace-45b8-a8d2-38a041110095\") " pod="openstack/rabbitmq-server-0" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.414831 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1fd840b6-4ace-45b8-a8d2-38a041110095-config-data\") pod \"rabbitmq-server-0\" (UID: \"1fd840b6-4ace-45b8-a8d2-38a041110095\") " pod="openstack/rabbitmq-server-0" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.414846 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dnhpc\" (UniqueName: \"kubernetes.io/projected/1fd840b6-4ace-45b8-a8d2-38a041110095-kube-api-access-dnhpc\") pod \"rabbitmq-server-0\" (UID: \"1fd840b6-4ace-45b8-a8d2-38a041110095\") " pod="openstack/rabbitmq-server-0" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.414873 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1fd840b6-4ace-45b8-a8d2-38a041110095-pod-info\") pod \"rabbitmq-server-0\" (UID: \"1fd840b6-4ace-45b8-a8d2-38a041110095\") " pod="openstack/rabbitmq-server-0" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.414895 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1fd840b6-4ace-45b8-a8d2-38a041110095-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"1fd840b6-4ace-45b8-a8d2-38a041110095\") " pod="openstack/rabbitmq-server-0" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.414917 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1fd840b6-4ace-45b8-a8d2-38a041110095-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"1fd840b6-4ace-45b8-a8d2-38a041110095\") " pod="openstack/rabbitmq-server-0" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.415897 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.416130 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"1fd840b6-4ace-45b8-a8d2-38a041110095\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/rabbitmq-server-0" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.416837 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1fd840b6-4ace-45b8-a8d2-38a041110095-config-data\") pod \"rabbitmq-server-0\" (UID: \"1fd840b6-4ace-45b8-a8d2-38a041110095\") " pod="openstack/rabbitmq-server-0" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.417474 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.417567 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1fd840b6-4ace-45b8-a8d2-38a041110095-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"1fd840b6-4ace-45b8-a8d2-38a041110095\") " pod="openstack/rabbitmq-server-0" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.419273 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1fd840b6-4ace-45b8-a8d2-38a041110095-server-conf\") pod \"rabbitmq-server-0\" (UID: \"1fd840b6-4ace-45b8-a8d2-38a041110095\") " pod="openstack/rabbitmq-server-0" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.419800 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.420205 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.420998 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1fd840b6-4ace-45b8-a8d2-38a041110095-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"1fd840b6-4ace-45b8-a8d2-38a041110095\") " pod="openstack/rabbitmq-server-0" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.421066 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1fd840b6-4ace-45b8-a8d2-38a041110095-pod-info\") pod \"rabbitmq-server-0\" (UID: \"1fd840b6-4ace-45b8-a8d2-38a041110095\") " pod="openstack/rabbitmq-server-0" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.421208 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.421208 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.421573 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1fd840b6-4ace-45b8-a8d2-38a041110095-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"1fd840b6-4ace-45b8-a8d2-38a041110095\") " pod="openstack/rabbitmq-server-0" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.421802 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-8mmtq" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.421999 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.422392 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.422839 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1fd840b6-4ace-45b8-a8d2-38a041110095-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"1fd840b6-4ace-45b8-a8d2-38a041110095\") " pod="openstack/rabbitmq-server-0" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.423676 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1fd840b6-4ace-45b8-a8d2-38a041110095-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"1fd840b6-4ace-45b8-a8d2-38a041110095\") " pod="openstack/rabbitmq-server-0" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.431867 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.433752 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1fd840b6-4ace-45b8-a8d2-38a041110095-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"1fd840b6-4ace-45b8-a8d2-38a041110095\") " pod="openstack/rabbitmq-server-0" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.457522 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dnhpc\" (UniqueName: \"kubernetes.io/projected/1fd840b6-4ace-45b8-a8d2-38a041110095-kube-api-access-dnhpc\") pod \"rabbitmq-server-0\" (UID: \"1fd840b6-4ace-45b8-a8d2-38a041110095\") " pod="openstack/rabbitmq-server-0" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.476661 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"1fd840b6-4ace-45b8-a8d2-38a041110095\") " pod="openstack/rabbitmq-server-0" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.621759 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ff37fa2f-dc06-430a-a4fe-af2b379fe563-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"ff37fa2f-dc06-430a-a4fe-af2b379fe563\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.621812 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"ff37fa2f-dc06-430a-a4fe-af2b379fe563\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.621834 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ff37fa2f-dc06-430a-a4fe-af2b379fe563-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"ff37fa2f-dc06-430a-a4fe-af2b379fe563\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.621852 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ff37fa2f-dc06-430a-a4fe-af2b379fe563-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"ff37fa2f-dc06-430a-a4fe-af2b379fe563\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.621888 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ff37fa2f-dc06-430a-a4fe-af2b379fe563-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"ff37fa2f-dc06-430a-a4fe-af2b379fe563\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.621911 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ff37fa2f-dc06-430a-a4fe-af2b379fe563-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"ff37fa2f-dc06-430a-a4fe-af2b379fe563\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.621930 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ff37fa2f-dc06-430a-a4fe-af2b379fe563-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ff37fa2f-dc06-430a-a4fe-af2b379fe563\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.621950 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ff37fa2f-dc06-430a-a4fe-af2b379fe563-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"ff37fa2f-dc06-430a-a4fe-af2b379fe563\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.621970 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ff37fa2f-dc06-430a-a4fe-af2b379fe563-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"ff37fa2f-dc06-430a-a4fe-af2b379fe563\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.621991 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7gjz6\" (UniqueName: \"kubernetes.io/projected/ff37fa2f-dc06-430a-a4fe-af2b379fe563-kube-api-access-7gjz6\") pod \"rabbitmq-cell1-server-0\" (UID: \"ff37fa2f-dc06-430a-a4fe-af2b379fe563\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.622031 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ff37fa2f-dc06-430a-a4fe-af2b379fe563-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ff37fa2f-dc06-430a-a4fe-af2b379fe563\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.724268 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ff37fa2f-dc06-430a-a4fe-af2b379fe563-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"ff37fa2f-dc06-430a-a4fe-af2b379fe563\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.724330 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ff37fa2f-dc06-430a-a4fe-af2b379fe563-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"ff37fa2f-dc06-430a-a4fe-af2b379fe563\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.724363 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7gjz6\" (UniqueName: \"kubernetes.io/projected/ff37fa2f-dc06-430a-a4fe-af2b379fe563-kube-api-access-7gjz6\") pod \"rabbitmq-cell1-server-0\" (UID: \"ff37fa2f-dc06-430a-a4fe-af2b379fe563\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.724465 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ff37fa2f-dc06-430a-a4fe-af2b379fe563-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ff37fa2f-dc06-430a-a4fe-af2b379fe563\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.724495 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ff37fa2f-dc06-430a-a4fe-af2b379fe563-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"ff37fa2f-dc06-430a-a4fe-af2b379fe563\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.724520 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"ff37fa2f-dc06-430a-a4fe-af2b379fe563\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.724542 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ff37fa2f-dc06-430a-a4fe-af2b379fe563-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"ff37fa2f-dc06-430a-a4fe-af2b379fe563\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.724566 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ff37fa2f-dc06-430a-a4fe-af2b379fe563-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"ff37fa2f-dc06-430a-a4fe-af2b379fe563\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.724605 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ff37fa2f-dc06-430a-a4fe-af2b379fe563-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"ff37fa2f-dc06-430a-a4fe-af2b379fe563\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.724636 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ff37fa2f-dc06-430a-a4fe-af2b379fe563-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"ff37fa2f-dc06-430a-a4fe-af2b379fe563\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.724660 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ff37fa2f-dc06-430a-a4fe-af2b379fe563-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ff37fa2f-dc06-430a-a4fe-af2b379fe563\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.725206 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"ff37fa2f-dc06-430a-a4fe-af2b379fe563\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-cell1-server-0" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.728590 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ff37fa2f-dc06-430a-a4fe-af2b379fe563-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"ff37fa2f-dc06-430a-a4fe-af2b379fe563\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.729320 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ff37fa2f-dc06-430a-a4fe-af2b379fe563-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"ff37fa2f-dc06-430a-a4fe-af2b379fe563\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.730562 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ff37fa2f-dc06-430a-a4fe-af2b379fe563-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"ff37fa2f-dc06-430a-a4fe-af2b379fe563\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.730838 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ff37fa2f-dc06-430a-a4fe-af2b379fe563-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ff37fa2f-dc06-430a-a4fe-af2b379fe563\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.733586 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ff37fa2f-dc06-430a-a4fe-af2b379fe563-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"ff37fa2f-dc06-430a-a4fe-af2b379fe563\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.734513 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ff37fa2f-dc06-430a-a4fe-af2b379fe563-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ff37fa2f-dc06-430a-a4fe-af2b379fe563\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.735518 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ff37fa2f-dc06-430a-a4fe-af2b379fe563-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"ff37fa2f-dc06-430a-a4fe-af2b379fe563\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.738510 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ff37fa2f-dc06-430a-a4fe-af2b379fe563-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"ff37fa2f-dc06-430a-a4fe-af2b379fe563\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.739148 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ff37fa2f-dc06-430a-a4fe-af2b379fe563-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"ff37fa2f-dc06-430a-a4fe-af2b379fe563\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.744941 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.745484 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7gjz6\" (UniqueName: \"kubernetes.io/projected/ff37fa2f-dc06-430a-a4fe-af2b379fe563-kube-api-access-7gjz6\") pod \"rabbitmq-cell1-server-0\" (UID: \"ff37fa2f-dc06-430a-a4fe-af2b379fe563\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.767601 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"ff37fa2f-dc06-430a-a4fe-af2b379fe563\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.819942 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 11 02:20:00 crc kubenswrapper[4824]: I1211 02:20:00.926856 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-8bq9w" event={"ID":"d72d278a-3895-4b90-abed-0905680540c8","Type":"ContainerStarted","Data":"864eba193e2630c696553c6d5038d28695e3aaf1224f61c024446584b949fbb9"} Dec 11 02:20:01 crc kubenswrapper[4824]: I1211 02:20:01.934093 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Dec 11 02:20:01 crc kubenswrapper[4824]: I1211 02:20:01.938207 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 11 02:20:01 crc kubenswrapper[4824]: I1211 02:20:01.944609 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Dec 11 02:20:01 crc kubenswrapper[4824]: I1211 02:20:01.944725 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Dec 11 02:20:01 crc kubenswrapper[4824]: I1211 02:20:01.945404 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-b4sph" Dec 11 02:20:01 crc kubenswrapper[4824]: I1211 02:20:01.945488 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Dec 11 02:20:01 crc kubenswrapper[4824]: I1211 02:20:01.946359 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 11 02:20:01 crc kubenswrapper[4824]: I1211 02:20:01.949867 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Dec 11 02:20:02 crc kubenswrapper[4824]: I1211 02:20:02.046560 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e5994f4-2c9e-4e8c-b7d3-84a15af846cd-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"6e5994f4-2c9e-4e8c-b7d3-84a15af846cd\") " pod="openstack/openstack-galera-0" Dec 11 02:20:02 crc kubenswrapper[4824]: I1211 02:20:02.046617 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-galera-0\" (UID: \"6e5994f4-2c9e-4e8c-b7d3-84a15af846cd\") " pod="openstack/openstack-galera-0" Dec 11 02:20:02 crc kubenswrapper[4824]: I1211 02:20:02.046633 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/6e5994f4-2c9e-4e8c-b7d3-84a15af846cd-kolla-config\") pod \"openstack-galera-0\" (UID: \"6e5994f4-2c9e-4e8c-b7d3-84a15af846cd\") " pod="openstack/openstack-galera-0" Dec 11 02:20:02 crc kubenswrapper[4824]: I1211 02:20:02.046647 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6e5994f4-2c9e-4e8c-b7d3-84a15af846cd-operator-scripts\") pod \"openstack-galera-0\" (UID: \"6e5994f4-2c9e-4e8c-b7d3-84a15af846cd\") " pod="openstack/openstack-galera-0" Dec 11 02:20:02 crc kubenswrapper[4824]: I1211 02:20:02.046673 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/6e5994f4-2c9e-4e8c-b7d3-84a15af846cd-config-data-generated\") pod \"openstack-galera-0\" (UID: \"6e5994f4-2c9e-4e8c-b7d3-84a15af846cd\") " pod="openstack/openstack-galera-0" Dec 11 02:20:02 crc kubenswrapper[4824]: I1211 02:20:02.046695 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/6e5994f4-2c9e-4e8c-b7d3-84a15af846cd-config-data-default\") pod \"openstack-galera-0\" (UID: \"6e5994f4-2c9e-4e8c-b7d3-84a15af846cd\") " pod="openstack/openstack-galera-0" Dec 11 02:20:02 crc kubenswrapper[4824]: I1211 02:20:02.046722 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pp6cb\" (UniqueName: \"kubernetes.io/projected/6e5994f4-2c9e-4e8c-b7d3-84a15af846cd-kube-api-access-pp6cb\") pod \"openstack-galera-0\" (UID: \"6e5994f4-2c9e-4e8c-b7d3-84a15af846cd\") " pod="openstack/openstack-galera-0" Dec 11 02:20:02 crc kubenswrapper[4824]: I1211 02:20:02.046756 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e5994f4-2c9e-4e8c-b7d3-84a15af846cd-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"6e5994f4-2c9e-4e8c-b7d3-84a15af846cd\") " pod="openstack/openstack-galera-0" Dec 11 02:20:02 crc kubenswrapper[4824]: I1211 02:20:02.148646 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e5994f4-2c9e-4e8c-b7d3-84a15af846cd-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"6e5994f4-2c9e-4e8c-b7d3-84a15af846cd\") " pod="openstack/openstack-galera-0" Dec 11 02:20:02 crc kubenswrapper[4824]: I1211 02:20:02.150246 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-galera-0\" (UID: \"6e5994f4-2c9e-4e8c-b7d3-84a15af846cd\") " pod="openstack/openstack-galera-0" Dec 11 02:20:02 crc kubenswrapper[4824]: I1211 02:20:02.150267 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/6e5994f4-2c9e-4e8c-b7d3-84a15af846cd-kolla-config\") pod \"openstack-galera-0\" (UID: \"6e5994f4-2c9e-4e8c-b7d3-84a15af846cd\") " pod="openstack/openstack-galera-0" Dec 11 02:20:02 crc kubenswrapper[4824]: I1211 02:20:02.150306 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6e5994f4-2c9e-4e8c-b7d3-84a15af846cd-operator-scripts\") pod \"openstack-galera-0\" (UID: \"6e5994f4-2c9e-4e8c-b7d3-84a15af846cd\") " pod="openstack/openstack-galera-0" Dec 11 02:20:02 crc kubenswrapper[4824]: I1211 02:20:02.150337 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/6e5994f4-2c9e-4e8c-b7d3-84a15af846cd-config-data-generated\") pod \"openstack-galera-0\" (UID: \"6e5994f4-2c9e-4e8c-b7d3-84a15af846cd\") " pod="openstack/openstack-galera-0" Dec 11 02:20:02 crc kubenswrapper[4824]: I1211 02:20:02.150355 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/6e5994f4-2c9e-4e8c-b7d3-84a15af846cd-config-data-default\") pod \"openstack-galera-0\" (UID: \"6e5994f4-2c9e-4e8c-b7d3-84a15af846cd\") " pod="openstack/openstack-galera-0" Dec 11 02:20:02 crc kubenswrapper[4824]: I1211 02:20:02.150405 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pp6cb\" (UniqueName: \"kubernetes.io/projected/6e5994f4-2c9e-4e8c-b7d3-84a15af846cd-kube-api-access-pp6cb\") pod \"openstack-galera-0\" (UID: \"6e5994f4-2c9e-4e8c-b7d3-84a15af846cd\") " pod="openstack/openstack-galera-0" Dec 11 02:20:02 crc kubenswrapper[4824]: I1211 02:20:02.150441 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e5994f4-2c9e-4e8c-b7d3-84a15af846cd-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"6e5994f4-2c9e-4e8c-b7d3-84a15af846cd\") " pod="openstack/openstack-galera-0" Dec 11 02:20:02 crc kubenswrapper[4824]: I1211 02:20:02.150546 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-galera-0\" (UID: \"6e5994f4-2c9e-4e8c-b7d3-84a15af846cd\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/openstack-galera-0" Dec 11 02:20:02 crc kubenswrapper[4824]: I1211 02:20:02.150877 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/6e5994f4-2c9e-4e8c-b7d3-84a15af846cd-config-data-generated\") pod \"openstack-galera-0\" (UID: \"6e5994f4-2c9e-4e8c-b7d3-84a15af846cd\") " pod="openstack/openstack-galera-0" Dec 11 02:20:02 crc kubenswrapper[4824]: I1211 02:20:02.151542 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/6e5994f4-2c9e-4e8c-b7d3-84a15af846cd-config-data-default\") pod \"openstack-galera-0\" (UID: \"6e5994f4-2c9e-4e8c-b7d3-84a15af846cd\") " pod="openstack/openstack-galera-0" Dec 11 02:20:02 crc kubenswrapper[4824]: I1211 02:20:02.152150 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6e5994f4-2c9e-4e8c-b7d3-84a15af846cd-operator-scripts\") pod \"openstack-galera-0\" (UID: \"6e5994f4-2c9e-4e8c-b7d3-84a15af846cd\") " pod="openstack/openstack-galera-0" Dec 11 02:20:02 crc kubenswrapper[4824]: I1211 02:20:02.153140 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/6e5994f4-2c9e-4e8c-b7d3-84a15af846cd-kolla-config\") pod \"openstack-galera-0\" (UID: \"6e5994f4-2c9e-4e8c-b7d3-84a15af846cd\") " pod="openstack/openstack-galera-0" Dec 11 02:20:02 crc kubenswrapper[4824]: I1211 02:20:02.163898 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e5994f4-2c9e-4e8c-b7d3-84a15af846cd-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"6e5994f4-2c9e-4e8c-b7d3-84a15af846cd\") " pod="openstack/openstack-galera-0" Dec 11 02:20:02 crc kubenswrapper[4824]: I1211 02:20:02.167945 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e5994f4-2c9e-4e8c-b7d3-84a15af846cd-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"6e5994f4-2c9e-4e8c-b7d3-84a15af846cd\") " pod="openstack/openstack-galera-0" Dec 11 02:20:02 crc kubenswrapper[4824]: I1211 02:20:02.169576 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pp6cb\" (UniqueName: \"kubernetes.io/projected/6e5994f4-2c9e-4e8c-b7d3-84a15af846cd-kube-api-access-pp6cb\") pod \"openstack-galera-0\" (UID: \"6e5994f4-2c9e-4e8c-b7d3-84a15af846cd\") " pod="openstack/openstack-galera-0" Dec 11 02:20:02 crc kubenswrapper[4824]: I1211 02:20:02.197721 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-galera-0\" (UID: \"6e5994f4-2c9e-4e8c-b7d3-84a15af846cd\") " pod="openstack/openstack-galera-0" Dec 11 02:20:02 crc kubenswrapper[4824]: I1211 02:20:02.268095 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 11 02:20:03 crc kubenswrapper[4824]: I1211 02:20:03.248170 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 11 02:20:03 crc kubenswrapper[4824]: I1211 02:20:03.249814 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 11 02:20:03 crc kubenswrapper[4824]: I1211 02:20:03.252315 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-gxfwx" Dec 11 02:20:03 crc kubenswrapper[4824]: I1211 02:20:03.252349 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Dec 11 02:20:03 crc kubenswrapper[4824]: I1211 02:20:03.252471 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Dec 11 02:20:03 crc kubenswrapper[4824]: I1211 02:20:03.252632 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Dec 11 02:20:03 crc kubenswrapper[4824]: I1211 02:20:03.257951 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 11 02:20:03 crc kubenswrapper[4824]: I1211 02:20:03.370958 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/f35efb32-a031-4c75-9358-3f5143335131-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"f35efb32-a031-4c75-9358-3f5143335131\") " pod="openstack/openstack-cell1-galera-0" Dec 11 02:20:03 crc kubenswrapper[4824]: I1211 02:20:03.371007 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-cell1-galera-0\" (UID: \"f35efb32-a031-4c75-9358-3f5143335131\") " pod="openstack/openstack-cell1-galera-0" Dec 11 02:20:03 crc kubenswrapper[4824]: I1211 02:20:03.371025 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f35efb32-a031-4c75-9358-3f5143335131-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"f35efb32-a031-4c75-9358-3f5143335131\") " pod="openstack/openstack-cell1-galera-0" Dec 11 02:20:03 crc kubenswrapper[4824]: I1211 02:20:03.371659 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f35efb32-a031-4c75-9358-3f5143335131-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"f35efb32-a031-4c75-9358-3f5143335131\") " pod="openstack/openstack-cell1-galera-0" Dec 11 02:20:03 crc kubenswrapper[4824]: I1211 02:20:03.371746 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-clzgk\" (UniqueName: \"kubernetes.io/projected/f35efb32-a031-4c75-9358-3f5143335131-kube-api-access-clzgk\") pod \"openstack-cell1-galera-0\" (UID: \"f35efb32-a031-4c75-9358-3f5143335131\") " pod="openstack/openstack-cell1-galera-0" Dec 11 02:20:03 crc kubenswrapper[4824]: I1211 02:20:03.371837 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f35efb32-a031-4c75-9358-3f5143335131-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"f35efb32-a031-4c75-9358-3f5143335131\") " pod="openstack/openstack-cell1-galera-0" Dec 11 02:20:03 crc kubenswrapper[4824]: I1211 02:20:03.371865 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f35efb32-a031-4c75-9358-3f5143335131-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"f35efb32-a031-4c75-9358-3f5143335131\") " pod="openstack/openstack-cell1-galera-0" Dec 11 02:20:03 crc kubenswrapper[4824]: I1211 02:20:03.371963 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f35efb32-a031-4c75-9358-3f5143335131-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"f35efb32-a031-4c75-9358-3f5143335131\") " pod="openstack/openstack-cell1-galera-0" Dec 11 02:20:03 crc kubenswrapper[4824]: I1211 02:20:03.472889 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f35efb32-a031-4c75-9358-3f5143335131-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"f35efb32-a031-4c75-9358-3f5143335131\") " pod="openstack/openstack-cell1-galera-0" Dec 11 02:20:03 crc kubenswrapper[4824]: I1211 02:20:03.472923 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f35efb32-a031-4c75-9358-3f5143335131-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"f35efb32-a031-4c75-9358-3f5143335131\") " pod="openstack/openstack-cell1-galera-0" Dec 11 02:20:03 crc kubenswrapper[4824]: I1211 02:20:03.472961 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f35efb32-a031-4c75-9358-3f5143335131-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"f35efb32-a031-4c75-9358-3f5143335131\") " pod="openstack/openstack-cell1-galera-0" Dec 11 02:20:03 crc kubenswrapper[4824]: I1211 02:20:03.473003 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/f35efb32-a031-4c75-9358-3f5143335131-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"f35efb32-a031-4c75-9358-3f5143335131\") " pod="openstack/openstack-cell1-galera-0" Dec 11 02:20:03 crc kubenswrapper[4824]: I1211 02:20:03.473021 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-cell1-galera-0\" (UID: \"f35efb32-a031-4c75-9358-3f5143335131\") " pod="openstack/openstack-cell1-galera-0" Dec 11 02:20:03 crc kubenswrapper[4824]: I1211 02:20:03.473038 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f35efb32-a031-4c75-9358-3f5143335131-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"f35efb32-a031-4c75-9358-3f5143335131\") " pod="openstack/openstack-cell1-galera-0" Dec 11 02:20:03 crc kubenswrapper[4824]: I1211 02:20:03.473105 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f35efb32-a031-4c75-9358-3f5143335131-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"f35efb32-a031-4c75-9358-3f5143335131\") " pod="openstack/openstack-cell1-galera-0" Dec 11 02:20:03 crc kubenswrapper[4824]: I1211 02:20:03.473147 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-clzgk\" (UniqueName: \"kubernetes.io/projected/f35efb32-a031-4c75-9358-3f5143335131-kube-api-access-clzgk\") pod \"openstack-cell1-galera-0\" (UID: \"f35efb32-a031-4c75-9358-3f5143335131\") " pod="openstack/openstack-cell1-galera-0" Dec 11 02:20:03 crc kubenswrapper[4824]: I1211 02:20:03.473786 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f35efb32-a031-4c75-9358-3f5143335131-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"f35efb32-a031-4c75-9358-3f5143335131\") " pod="openstack/openstack-cell1-galera-0" Dec 11 02:20:03 crc kubenswrapper[4824]: I1211 02:20:03.474021 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f35efb32-a031-4c75-9358-3f5143335131-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"f35efb32-a031-4c75-9358-3f5143335131\") " pod="openstack/openstack-cell1-galera-0" Dec 11 02:20:03 crc kubenswrapper[4824]: I1211 02:20:03.474019 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-cell1-galera-0\" (UID: \"f35efb32-a031-4c75-9358-3f5143335131\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/openstack-cell1-galera-0" Dec 11 02:20:03 crc kubenswrapper[4824]: I1211 02:20:03.478987 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f35efb32-a031-4c75-9358-3f5143335131-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"f35efb32-a031-4c75-9358-3f5143335131\") " pod="openstack/openstack-cell1-galera-0" Dec 11 02:20:03 crc kubenswrapper[4824]: I1211 02:20:03.479634 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/f35efb32-a031-4c75-9358-3f5143335131-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"f35efb32-a031-4c75-9358-3f5143335131\") " pod="openstack/openstack-cell1-galera-0" Dec 11 02:20:03 crc kubenswrapper[4824]: I1211 02:20:03.480186 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f35efb32-a031-4c75-9358-3f5143335131-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"f35efb32-a031-4c75-9358-3f5143335131\") " pod="openstack/openstack-cell1-galera-0" Dec 11 02:20:03 crc kubenswrapper[4824]: I1211 02:20:03.480470 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f35efb32-a031-4c75-9358-3f5143335131-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"f35efb32-a031-4c75-9358-3f5143335131\") " pod="openstack/openstack-cell1-galera-0" Dec 11 02:20:03 crc kubenswrapper[4824]: I1211 02:20:03.497093 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-clzgk\" (UniqueName: \"kubernetes.io/projected/f35efb32-a031-4c75-9358-3f5143335131-kube-api-access-clzgk\") pod \"openstack-cell1-galera-0\" (UID: \"f35efb32-a031-4c75-9358-3f5143335131\") " pod="openstack/openstack-cell1-galera-0" Dec 11 02:20:03 crc kubenswrapper[4824]: I1211 02:20:03.502969 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-cell1-galera-0\" (UID: \"f35efb32-a031-4c75-9358-3f5143335131\") " pod="openstack/openstack-cell1-galera-0" Dec 11 02:20:03 crc kubenswrapper[4824]: I1211 02:20:03.534722 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Dec 11 02:20:03 crc kubenswrapper[4824]: I1211 02:20:03.535611 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 11 02:20:03 crc kubenswrapper[4824]: I1211 02:20:03.538837 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Dec 11 02:20:03 crc kubenswrapper[4824]: I1211 02:20:03.539386 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Dec 11 02:20:03 crc kubenswrapper[4824]: I1211 02:20:03.539469 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-dmwpb" Dec 11 02:20:03 crc kubenswrapper[4824]: I1211 02:20:03.549242 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 11 02:20:03 crc kubenswrapper[4824]: I1211 02:20:03.575920 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 11 02:20:03 crc kubenswrapper[4824]: I1211 02:20:03.677766 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/709f3d23-a475-443a-8787-7f018c486b56-config-data\") pod \"memcached-0\" (UID: \"709f3d23-a475-443a-8787-7f018c486b56\") " pod="openstack/memcached-0" Dec 11 02:20:03 crc kubenswrapper[4824]: I1211 02:20:03.677810 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/709f3d23-a475-443a-8787-7f018c486b56-combined-ca-bundle\") pod \"memcached-0\" (UID: \"709f3d23-a475-443a-8787-7f018c486b56\") " pod="openstack/memcached-0" Dec 11 02:20:03 crc kubenswrapper[4824]: I1211 02:20:03.677854 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/709f3d23-a475-443a-8787-7f018c486b56-kolla-config\") pod \"memcached-0\" (UID: \"709f3d23-a475-443a-8787-7f018c486b56\") " pod="openstack/memcached-0" Dec 11 02:20:03 crc kubenswrapper[4824]: I1211 02:20:03.677899 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/709f3d23-a475-443a-8787-7f018c486b56-memcached-tls-certs\") pod \"memcached-0\" (UID: \"709f3d23-a475-443a-8787-7f018c486b56\") " pod="openstack/memcached-0" Dec 11 02:20:03 crc kubenswrapper[4824]: I1211 02:20:03.678024 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kczhb\" (UniqueName: \"kubernetes.io/projected/709f3d23-a475-443a-8787-7f018c486b56-kube-api-access-kczhb\") pod \"memcached-0\" (UID: \"709f3d23-a475-443a-8787-7f018c486b56\") " pod="openstack/memcached-0" Dec 11 02:20:03 crc kubenswrapper[4824]: I1211 02:20:03.779154 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/709f3d23-a475-443a-8787-7f018c486b56-memcached-tls-certs\") pod \"memcached-0\" (UID: \"709f3d23-a475-443a-8787-7f018c486b56\") " pod="openstack/memcached-0" Dec 11 02:20:03 crc kubenswrapper[4824]: I1211 02:20:03.779210 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kczhb\" (UniqueName: \"kubernetes.io/projected/709f3d23-a475-443a-8787-7f018c486b56-kube-api-access-kczhb\") pod \"memcached-0\" (UID: \"709f3d23-a475-443a-8787-7f018c486b56\") " pod="openstack/memcached-0" Dec 11 02:20:03 crc kubenswrapper[4824]: I1211 02:20:03.779307 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/709f3d23-a475-443a-8787-7f018c486b56-config-data\") pod \"memcached-0\" (UID: \"709f3d23-a475-443a-8787-7f018c486b56\") " pod="openstack/memcached-0" Dec 11 02:20:03 crc kubenswrapper[4824]: I1211 02:20:03.779324 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/709f3d23-a475-443a-8787-7f018c486b56-combined-ca-bundle\") pod \"memcached-0\" (UID: \"709f3d23-a475-443a-8787-7f018c486b56\") " pod="openstack/memcached-0" Dec 11 02:20:03 crc kubenswrapper[4824]: I1211 02:20:03.780179 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/709f3d23-a475-443a-8787-7f018c486b56-kolla-config\") pod \"memcached-0\" (UID: \"709f3d23-a475-443a-8787-7f018c486b56\") " pod="openstack/memcached-0" Dec 11 02:20:03 crc kubenswrapper[4824]: I1211 02:20:03.780775 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/709f3d23-a475-443a-8787-7f018c486b56-config-data\") pod \"memcached-0\" (UID: \"709f3d23-a475-443a-8787-7f018c486b56\") " pod="openstack/memcached-0" Dec 11 02:20:03 crc kubenswrapper[4824]: I1211 02:20:03.780865 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/709f3d23-a475-443a-8787-7f018c486b56-kolla-config\") pod \"memcached-0\" (UID: \"709f3d23-a475-443a-8787-7f018c486b56\") " pod="openstack/memcached-0" Dec 11 02:20:03 crc kubenswrapper[4824]: I1211 02:20:03.783380 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/709f3d23-a475-443a-8787-7f018c486b56-combined-ca-bundle\") pod \"memcached-0\" (UID: \"709f3d23-a475-443a-8787-7f018c486b56\") " pod="openstack/memcached-0" Dec 11 02:20:03 crc kubenswrapper[4824]: I1211 02:20:03.783590 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/709f3d23-a475-443a-8787-7f018c486b56-memcached-tls-certs\") pod \"memcached-0\" (UID: \"709f3d23-a475-443a-8787-7f018c486b56\") " pod="openstack/memcached-0" Dec 11 02:20:03 crc kubenswrapper[4824]: I1211 02:20:03.793567 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kczhb\" (UniqueName: \"kubernetes.io/projected/709f3d23-a475-443a-8787-7f018c486b56-kube-api-access-kczhb\") pod \"memcached-0\" (UID: \"709f3d23-a475-443a-8787-7f018c486b56\") " pod="openstack/memcached-0" Dec 11 02:20:03 crc kubenswrapper[4824]: I1211 02:20:03.880490 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 11 02:20:05 crc kubenswrapper[4824]: I1211 02:20:05.559912 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 11 02:20:05 crc kubenswrapper[4824]: I1211 02:20:05.569195 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 11 02:20:05 crc kubenswrapper[4824]: I1211 02:20:05.572454 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-pqr2g" Dec 11 02:20:05 crc kubenswrapper[4824]: I1211 02:20:05.602613 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 11 02:20:05 crc kubenswrapper[4824]: I1211 02:20:05.715895 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mk6xs\" (UniqueName: \"kubernetes.io/projected/5b375617-b0a6-42cf-be23-ef7f755f9eff-kube-api-access-mk6xs\") pod \"kube-state-metrics-0\" (UID: \"5b375617-b0a6-42cf-be23-ef7f755f9eff\") " pod="openstack/kube-state-metrics-0" Dec 11 02:20:05 crc kubenswrapper[4824]: I1211 02:20:05.816831 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mk6xs\" (UniqueName: \"kubernetes.io/projected/5b375617-b0a6-42cf-be23-ef7f755f9eff-kube-api-access-mk6xs\") pod \"kube-state-metrics-0\" (UID: \"5b375617-b0a6-42cf-be23-ef7f755f9eff\") " pod="openstack/kube-state-metrics-0" Dec 11 02:20:05 crc kubenswrapper[4824]: I1211 02:20:05.861203 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mk6xs\" (UniqueName: \"kubernetes.io/projected/5b375617-b0a6-42cf-be23-ef7f755f9eff-kube-api-access-mk6xs\") pod \"kube-state-metrics-0\" (UID: \"5b375617-b0a6-42cf-be23-ef7f755f9eff\") " pod="openstack/kube-state-metrics-0" Dec 11 02:20:05 crc kubenswrapper[4824]: I1211 02:20:05.931435 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 11 02:20:07 crc kubenswrapper[4824]: I1211 02:20:07.250997 4824 patch_prober.go:28] interesting pod/machine-config-daemon-gx6xt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 02:20:07 crc kubenswrapper[4824]: I1211 02:20:07.251057 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 02:20:09 crc kubenswrapper[4824]: I1211 02:20:09.239583 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-779x4"] Dec 11 02:20:09 crc kubenswrapper[4824]: I1211 02:20:09.240767 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-779x4" Dec 11 02:20:09 crc kubenswrapper[4824]: I1211 02:20:09.242934 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Dec 11 02:20:09 crc kubenswrapper[4824]: I1211 02:20:09.243247 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-99xvr" Dec 11 02:20:09 crc kubenswrapper[4824]: I1211 02:20:09.247590 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Dec 11 02:20:09 crc kubenswrapper[4824]: I1211 02:20:09.253840 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-779x4"] Dec 11 02:20:09 crc kubenswrapper[4824]: I1211 02:20:09.310803 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-xzkzd"] Dec 11 02:20:09 crc kubenswrapper[4824]: I1211 02:20:09.312450 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-xzkzd" Dec 11 02:20:09 crc kubenswrapper[4824]: I1211 02:20:09.316700 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-xzkzd"] Dec 11 02:20:09 crc kubenswrapper[4824]: I1211 02:20:09.373800 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/148c69f8-3121-4a55-b3a6-be44b816f643-var-run-ovn\") pod \"ovn-controller-779x4\" (UID: \"148c69f8-3121-4a55-b3a6-be44b816f643\") " pod="openstack/ovn-controller-779x4" Dec 11 02:20:09 crc kubenswrapper[4824]: I1211 02:20:09.373879 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/148c69f8-3121-4a55-b3a6-be44b816f643-ovn-controller-tls-certs\") pod \"ovn-controller-779x4\" (UID: \"148c69f8-3121-4a55-b3a6-be44b816f643\") " pod="openstack/ovn-controller-779x4" Dec 11 02:20:09 crc kubenswrapper[4824]: I1211 02:20:09.373940 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/17cc80fc-a51f-41e4-be0f-f593fc23476e-var-log\") pod \"ovn-controller-ovs-xzkzd\" (UID: \"17cc80fc-a51f-41e4-be0f-f593fc23476e\") " pod="openstack/ovn-controller-ovs-xzkzd" Dec 11 02:20:09 crc kubenswrapper[4824]: I1211 02:20:09.373966 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/148c69f8-3121-4a55-b3a6-be44b816f643-var-log-ovn\") pod \"ovn-controller-779x4\" (UID: \"148c69f8-3121-4a55-b3a6-be44b816f643\") " pod="openstack/ovn-controller-779x4" Dec 11 02:20:09 crc kubenswrapper[4824]: I1211 02:20:09.374008 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/17cc80fc-a51f-41e4-be0f-f593fc23476e-var-lib\") pod \"ovn-controller-ovs-xzkzd\" (UID: \"17cc80fc-a51f-41e4-be0f-f593fc23476e\") " pod="openstack/ovn-controller-ovs-xzkzd" Dec 11 02:20:09 crc kubenswrapper[4824]: I1211 02:20:09.374087 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l5jtp\" (UniqueName: \"kubernetes.io/projected/17cc80fc-a51f-41e4-be0f-f593fc23476e-kube-api-access-l5jtp\") pod \"ovn-controller-ovs-xzkzd\" (UID: \"17cc80fc-a51f-41e4-be0f-f593fc23476e\") " pod="openstack/ovn-controller-ovs-xzkzd" Dec 11 02:20:09 crc kubenswrapper[4824]: I1211 02:20:09.374140 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/17cc80fc-a51f-41e4-be0f-f593fc23476e-var-run\") pod \"ovn-controller-ovs-xzkzd\" (UID: \"17cc80fc-a51f-41e4-be0f-f593fc23476e\") " pod="openstack/ovn-controller-ovs-xzkzd" Dec 11 02:20:09 crc kubenswrapper[4824]: I1211 02:20:09.374185 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/148c69f8-3121-4a55-b3a6-be44b816f643-var-run\") pod \"ovn-controller-779x4\" (UID: \"148c69f8-3121-4a55-b3a6-be44b816f643\") " pod="openstack/ovn-controller-779x4" Dec 11 02:20:09 crc kubenswrapper[4824]: I1211 02:20:09.374228 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ltj7n\" (UniqueName: \"kubernetes.io/projected/148c69f8-3121-4a55-b3a6-be44b816f643-kube-api-access-ltj7n\") pod \"ovn-controller-779x4\" (UID: \"148c69f8-3121-4a55-b3a6-be44b816f643\") " pod="openstack/ovn-controller-779x4" Dec 11 02:20:09 crc kubenswrapper[4824]: I1211 02:20:09.374258 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/17cc80fc-a51f-41e4-be0f-f593fc23476e-etc-ovs\") pod \"ovn-controller-ovs-xzkzd\" (UID: \"17cc80fc-a51f-41e4-be0f-f593fc23476e\") " pod="openstack/ovn-controller-ovs-xzkzd" Dec 11 02:20:09 crc kubenswrapper[4824]: I1211 02:20:09.374314 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/148c69f8-3121-4a55-b3a6-be44b816f643-scripts\") pod \"ovn-controller-779x4\" (UID: \"148c69f8-3121-4a55-b3a6-be44b816f643\") " pod="openstack/ovn-controller-779x4" Dec 11 02:20:09 crc kubenswrapper[4824]: I1211 02:20:09.374382 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/148c69f8-3121-4a55-b3a6-be44b816f643-combined-ca-bundle\") pod \"ovn-controller-779x4\" (UID: \"148c69f8-3121-4a55-b3a6-be44b816f643\") " pod="openstack/ovn-controller-779x4" Dec 11 02:20:09 crc kubenswrapper[4824]: I1211 02:20:09.374415 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/17cc80fc-a51f-41e4-be0f-f593fc23476e-scripts\") pod \"ovn-controller-ovs-xzkzd\" (UID: \"17cc80fc-a51f-41e4-be0f-f593fc23476e\") " pod="openstack/ovn-controller-ovs-xzkzd" Dec 11 02:20:09 crc kubenswrapper[4824]: I1211 02:20:09.475312 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/148c69f8-3121-4a55-b3a6-be44b816f643-combined-ca-bundle\") pod \"ovn-controller-779x4\" (UID: \"148c69f8-3121-4a55-b3a6-be44b816f643\") " pod="openstack/ovn-controller-779x4" Dec 11 02:20:09 crc kubenswrapper[4824]: I1211 02:20:09.475368 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/17cc80fc-a51f-41e4-be0f-f593fc23476e-scripts\") pod \"ovn-controller-ovs-xzkzd\" (UID: \"17cc80fc-a51f-41e4-be0f-f593fc23476e\") " pod="openstack/ovn-controller-ovs-xzkzd" Dec 11 02:20:09 crc kubenswrapper[4824]: I1211 02:20:09.475410 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/148c69f8-3121-4a55-b3a6-be44b816f643-var-run-ovn\") pod \"ovn-controller-779x4\" (UID: \"148c69f8-3121-4a55-b3a6-be44b816f643\") " pod="openstack/ovn-controller-779x4" Dec 11 02:20:09 crc kubenswrapper[4824]: I1211 02:20:09.475444 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/148c69f8-3121-4a55-b3a6-be44b816f643-ovn-controller-tls-certs\") pod \"ovn-controller-779x4\" (UID: \"148c69f8-3121-4a55-b3a6-be44b816f643\") " pod="openstack/ovn-controller-779x4" Dec 11 02:20:09 crc kubenswrapper[4824]: I1211 02:20:09.475473 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/17cc80fc-a51f-41e4-be0f-f593fc23476e-var-log\") pod \"ovn-controller-ovs-xzkzd\" (UID: \"17cc80fc-a51f-41e4-be0f-f593fc23476e\") " pod="openstack/ovn-controller-ovs-xzkzd" Dec 11 02:20:09 crc kubenswrapper[4824]: I1211 02:20:09.475498 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/148c69f8-3121-4a55-b3a6-be44b816f643-var-log-ovn\") pod \"ovn-controller-779x4\" (UID: \"148c69f8-3121-4a55-b3a6-be44b816f643\") " pod="openstack/ovn-controller-779x4" Dec 11 02:20:09 crc kubenswrapper[4824]: I1211 02:20:09.475526 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/17cc80fc-a51f-41e4-be0f-f593fc23476e-var-lib\") pod \"ovn-controller-ovs-xzkzd\" (UID: \"17cc80fc-a51f-41e4-be0f-f593fc23476e\") " pod="openstack/ovn-controller-ovs-xzkzd" Dec 11 02:20:09 crc kubenswrapper[4824]: I1211 02:20:09.475561 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l5jtp\" (UniqueName: \"kubernetes.io/projected/17cc80fc-a51f-41e4-be0f-f593fc23476e-kube-api-access-l5jtp\") pod \"ovn-controller-ovs-xzkzd\" (UID: \"17cc80fc-a51f-41e4-be0f-f593fc23476e\") " pod="openstack/ovn-controller-ovs-xzkzd" Dec 11 02:20:09 crc kubenswrapper[4824]: I1211 02:20:09.475586 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/17cc80fc-a51f-41e4-be0f-f593fc23476e-var-run\") pod \"ovn-controller-ovs-xzkzd\" (UID: \"17cc80fc-a51f-41e4-be0f-f593fc23476e\") " pod="openstack/ovn-controller-ovs-xzkzd" Dec 11 02:20:09 crc kubenswrapper[4824]: I1211 02:20:09.475637 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/148c69f8-3121-4a55-b3a6-be44b816f643-var-run\") pod \"ovn-controller-779x4\" (UID: \"148c69f8-3121-4a55-b3a6-be44b816f643\") " pod="openstack/ovn-controller-779x4" Dec 11 02:20:09 crc kubenswrapper[4824]: I1211 02:20:09.475658 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ltj7n\" (UniqueName: \"kubernetes.io/projected/148c69f8-3121-4a55-b3a6-be44b816f643-kube-api-access-ltj7n\") pod \"ovn-controller-779x4\" (UID: \"148c69f8-3121-4a55-b3a6-be44b816f643\") " pod="openstack/ovn-controller-779x4" Dec 11 02:20:09 crc kubenswrapper[4824]: I1211 02:20:09.475695 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/17cc80fc-a51f-41e4-be0f-f593fc23476e-etc-ovs\") pod \"ovn-controller-ovs-xzkzd\" (UID: \"17cc80fc-a51f-41e4-be0f-f593fc23476e\") " pod="openstack/ovn-controller-ovs-xzkzd" Dec 11 02:20:09 crc kubenswrapper[4824]: I1211 02:20:09.475715 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/148c69f8-3121-4a55-b3a6-be44b816f643-scripts\") pod \"ovn-controller-779x4\" (UID: \"148c69f8-3121-4a55-b3a6-be44b816f643\") " pod="openstack/ovn-controller-779x4" Dec 11 02:20:09 crc kubenswrapper[4824]: I1211 02:20:09.476431 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/148c69f8-3121-4a55-b3a6-be44b816f643-var-run-ovn\") pod \"ovn-controller-779x4\" (UID: \"148c69f8-3121-4a55-b3a6-be44b816f643\") " pod="openstack/ovn-controller-779x4" Dec 11 02:20:09 crc kubenswrapper[4824]: I1211 02:20:09.476560 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/148c69f8-3121-4a55-b3a6-be44b816f643-var-log-ovn\") pod \"ovn-controller-779x4\" (UID: \"148c69f8-3121-4a55-b3a6-be44b816f643\") " pod="openstack/ovn-controller-779x4" Dec 11 02:20:09 crc kubenswrapper[4824]: I1211 02:20:09.476647 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/148c69f8-3121-4a55-b3a6-be44b816f643-var-run\") pod \"ovn-controller-779x4\" (UID: \"148c69f8-3121-4a55-b3a6-be44b816f643\") " pod="openstack/ovn-controller-779x4" Dec 11 02:20:09 crc kubenswrapper[4824]: I1211 02:20:09.476724 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/17cc80fc-a51f-41e4-be0f-f593fc23476e-var-run\") pod \"ovn-controller-ovs-xzkzd\" (UID: \"17cc80fc-a51f-41e4-be0f-f593fc23476e\") " pod="openstack/ovn-controller-ovs-xzkzd" Dec 11 02:20:09 crc kubenswrapper[4824]: I1211 02:20:09.476899 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/17cc80fc-a51f-41e4-be0f-f593fc23476e-var-lib\") pod \"ovn-controller-ovs-xzkzd\" (UID: \"17cc80fc-a51f-41e4-be0f-f593fc23476e\") " pod="openstack/ovn-controller-ovs-xzkzd" Dec 11 02:20:09 crc kubenswrapper[4824]: I1211 02:20:09.477013 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/17cc80fc-a51f-41e4-be0f-f593fc23476e-etc-ovs\") pod \"ovn-controller-ovs-xzkzd\" (UID: \"17cc80fc-a51f-41e4-be0f-f593fc23476e\") " pod="openstack/ovn-controller-ovs-xzkzd" Dec 11 02:20:09 crc kubenswrapper[4824]: I1211 02:20:09.477262 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/17cc80fc-a51f-41e4-be0f-f593fc23476e-var-log\") pod \"ovn-controller-ovs-xzkzd\" (UID: \"17cc80fc-a51f-41e4-be0f-f593fc23476e\") " pod="openstack/ovn-controller-ovs-xzkzd" Dec 11 02:20:09 crc kubenswrapper[4824]: I1211 02:20:09.478497 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/148c69f8-3121-4a55-b3a6-be44b816f643-scripts\") pod \"ovn-controller-779x4\" (UID: \"148c69f8-3121-4a55-b3a6-be44b816f643\") " pod="openstack/ovn-controller-779x4" Dec 11 02:20:09 crc kubenswrapper[4824]: I1211 02:20:09.479417 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/17cc80fc-a51f-41e4-be0f-f593fc23476e-scripts\") pod \"ovn-controller-ovs-xzkzd\" (UID: \"17cc80fc-a51f-41e4-be0f-f593fc23476e\") " pod="openstack/ovn-controller-ovs-xzkzd" Dec 11 02:20:09 crc kubenswrapper[4824]: I1211 02:20:09.483691 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/148c69f8-3121-4a55-b3a6-be44b816f643-combined-ca-bundle\") pod \"ovn-controller-779x4\" (UID: \"148c69f8-3121-4a55-b3a6-be44b816f643\") " pod="openstack/ovn-controller-779x4" Dec 11 02:20:09 crc kubenswrapper[4824]: I1211 02:20:09.486136 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/148c69f8-3121-4a55-b3a6-be44b816f643-ovn-controller-tls-certs\") pod \"ovn-controller-779x4\" (UID: \"148c69f8-3121-4a55-b3a6-be44b816f643\") " pod="openstack/ovn-controller-779x4" Dec 11 02:20:09 crc kubenswrapper[4824]: I1211 02:20:09.497721 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l5jtp\" (UniqueName: \"kubernetes.io/projected/17cc80fc-a51f-41e4-be0f-f593fc23476e-kube-api-access-l5jtp\") pod \"ovn-controller-ovs-xzkzd\" (UID: \"17cc80fc-a51f-41e4-be0f-f593fc23476e\") " pod="openstack/ovn-controller-ovs-xzkzd" Dec 11 02:20:09 crc kubenswrapper[4824]: I1211 02:20:09.509719 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ltj7n\" (UniqueName: \"kubernetes.io/projected/148c69f8-3121-4a55-b3a6-be44b816f643-kube-api-access-ltj7n\") pod \"ovn-controller-779x4\" (UID: \"148c69f8-3121-4a55-b3a6-be44b816f643\") " pod="openstack/ovn-controller-779x4" Dec 11 02:20:09 crc kubenswrapper[4824]: I1211 02:20:09.570717 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-779x4" Dec 11 02:20:09 crc kubenswrapper[4824]: I1211 02:20:09.637518 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-xzkzd" Dec 11 02:20:11 crc kubenswrapper[4824]: I1211 02:20:11.230138 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 11 02:20:11 crc kubenswrapper[4824]: I1211 02:20:11.231631 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 11 02:20:11 crc kubenswrapper[4824]: I1211 02:20:11.234844 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Dec 11 02:20:11 crc kubenswrapper[4824]: I1211 02:20:11.235242 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Dec 11 02:20:11 crc kubenswrapper[4824]: I1211 02:20:11.235530 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Dec 11 02:20:11 crc kubenswrapper[4824]: I1211 02:20:11.235759 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Dec 11 02:20:11 crc kubenswrapper[4824]: I1211 02:20:11.235972 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-hpqxw" Dec 11 02:20:11 crc kubenswrapper[4824]: I1211 02:20:11.236264 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 11 02:20:11 crc kubenswrapper[4824]: I1211 02:20:11.315796 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/402ac70d-73e7-4697-ac46-a92bbbaf1aac-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"402ac70d-73e7-4697-ac46-a92bbbaf1aac\") " pod="openstack/ovsdbserver-sb-0" Dec 11 02:20:11 crc kubenswrapper[4824]: I1211 02:20:11.315859 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/402ac70d-73e7-4697-ac46-a92bbbaf1aac-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"402ac70d-73e7-4697-ac46-a92bbbaf1aac\") " pod="openstack/ovsdbserver-sb-0" Dec 11 02:20:11 crc kubenswrapper[4824]: I1211 02:20:11.315948 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/402ac70d-73e7-4697-ac46-a92bbbaf1aac-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"402ac70d-73e7-4697-ac46-a92bbbaf1aac\") " pod="openstack/ovsdbserver-sb-0" Dec 11 02:20:11 crc kubenswrapper[4824]: I1211 02:20:11.315987 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8zb9j\" (UniqueName: \"kubernetes.io/projected/402ac70d-73e7-4697-ac46-a92bbbaf1aac-kube-api-access-8zb9j\") pod \"ovsdbserver-sb-0\" (UID: \"402ac70d-73e7-4697-ac46-a92bbbaf1aac\") " pod="openstack/ovsdbserver-sb-0" Dec 11 02:20:11 crc kubenswrapper[4824]: I1211 02:20:11.316018 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/402ac70d-73e7-4697-ac46-a92bbbaf1aac-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"402ac70d-73e7-4697-ac46-a92bbbaf1aac\") " pod="openstack/ovsdbserver-sb-0" Dec 11 02:20:11 crc kubenswrapper[4824]: I1211 02:20:11.316212 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"402ac70d-73e7-4697-ac46-a92bbbaf1aac\") " pod="openstack/ovsdbserver-sb-0" Dec 11 02:20:11 crc kubenswrapper[4824]: I1211 02:20:11.316415 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/402ac70d-73e7-4697-ac46-a92bbbaf1aac-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"402ac70d-73e7-4697-ac46-a92bbbaf1aac\") " pod="openstack/ovsdbserver-sb-0" Dec 11 02:20:11 crc kubenswrapper[4824]: I1211 02:20:11.316519 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/402ac70d-73e7-4697-ac46-a92bbbaf1aac-config\") pod \"ovsdbserver-sb-0\" (UID: \"402ac70d-73e7-4697-ac46-a92bbbaf1aac\") " pod="openstack/ovsdbserver-sb-0" Dec 11 02:20:11 crc kubenswrapper[4824]: I1211 02:20:11.419989 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/402ac70d-73e7-4697-ac46-a92bbbaf1aac-config\") pod \"ovsdbserver-sb-0\" (UID: \"402ac70d-73e7-4697-ac46-a92bbbaf1aac\") " pod="openstack/ovsdbserver-sb-0" Dec 11 02:20:11 crc kubenswrapper[4824]: I1211 02:20:11.420077 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/402ac70d-73e7-4697-ac46-a92bbbaf1aac-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"402ac70d-73e7-4697-ac46-a92bbbaf1aac\") " pod="openstack/ovsdbserver-sb-0" Dec 11 02:20:11 crc kubenswrapper[4824]: I1211 02:20:11.420103 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/402ac70d-73e7-4697-ac46-a92bbbaf1aac-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"402ac70d-73e7-4697-ac46-a92bbbaf1aac\") " pod="openstack/ovsdbserver-sb-0" Dec 11 02:20:11 crc kubenswrapper[4824]: I1211 02:20:11.420666 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/402ac70d-73e7-4697-ac46-a92bbbaf1aac-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"402ac70d-73e7-4697-ac46-a92bbbaf1aac\") " pod="openstack/ovsdbserver-sb-0" Dec 11 02:20:11 crc kubenswrapper[4824]: I1211 02:20:11.420711 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8zb9j\" (UniqueName: \"kubernetes.io/projected/402ac70d-73e7-4697-ac46-a92bbbaf1aac-kube-api-access-8zb9j\") pod \"ovsdbserver-sb-0\" (UID: \"402ac70d-73e7-4697-ac46-a92bbbaf1aac\") " pod="openstack/ovsdbserver-sb-0" Dec 11 02:20:11 crc kubenswrapper[4824]: I1211 02:20:11.420740 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/402ac70d-73e7-4697-ac46-a92bbbaf1aac-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"402ac70d-73e7-4697-ac46-a92bbbaf1aac\") " pod="openstack/ovsdbserver-sb-0" Dec 11 02:20:11 crc kubenswrapper[4824]: I1211 02:20:11.420809 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"402ac70d-73e7-4697-ac46-a92bbbaf1aac\") " pod="openstack/ovsdbserver-sb-0" Dec 11 02:20:11 crc kubenswrapper[4824]: I1211 02:20:11.420886 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/402ac70d-73e7-4697-ac46-a92bbbaf1aac-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"402ac70d-73e7-4697-ac46-a92bbbaf1aac\") " pod="openstack/ovsdbserver-sb-0" Dec 11 02:20:11 crc kubenswrapper[4824]: I1211 02:20:11.421579 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/402ac70d-73e7-4697-ac46-a92bbbaf1aac-config\") pod \"ovsdbserver-sb-0\" (UID: \"402ac70d-73e7-4697-ac46-a92bbbaf1aac\") " pod="openstack/ovsdbserver-sb-0" Dec 11 02:20:11 crc kubenswrapper[4824]: I1211 02:20:11.423150 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/402ac70d-73e7-4697-ac46-a92bbbaf1aac-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"402ac70d-73e7-4697-ac46-a92bbbaf1aac\") " pod="openstack/ovsdbserver-sb-0" Dec 11 02:20:11 crc kubenswrapper[4824]: I1211 02:20:11.423761 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"402ac70d-73e7-4697-ac46-a92bbbaf1aac\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/ovsdbserver-sb-0" Dec 11 02:20:11 crc kubenswrapper[4824]: I1211 02:20:11.423773 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/402ac70d-73e7-4697-ac46-a92bbbaf1aac-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"402ac70d-73e7-4697-ac46-a92bbbaf1aac\") " pod="openstack/ovsdbserver-sb-0" Dec 11 02:20:11 crc kubenswrapper[4824]: I1211 02:20:11.430743 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/402ac70d-73e7-4697-ac46-a92bbbaf1aac-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"402ac70d-73e7-4697-ac46-a92bbbaf1aac\") " pod="openstack/ovsdbserver-sb-0" Dec 11 02:20:11 crc kubenswrapper[4824]: I1211 02:20:11.441823 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/402ac70d-73e7-4697-ac46-a92bbbaf1aac-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"402ac70d-73e7-4697-ac46-a92bbbaf1aac\") " pod="openstack/ovsdbserver-sb-0" Dec 11 02:20:11 crc kubenswrapper[4824]: I1211 02:20:11.445019 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/402ac70d-73e7-4697-ac46-a92bbbaf1aac-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"402ac70d-73e7-4697-ac46-a92bbbaf1aac\") " pod="openstack/ovsdbserver-sb-0" Dec 11 02:20:11 crc kubenswrapper[4824]: I1211 02:20:11.450383 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8zb9j\" (UniqueName: \"kubernetes.io/projected/402ac70d-73e7-4697-ac46-a92bbbaf1aac-kube-api-access-8zb9j\") pod \"ovsdbserver-sb-0\" (UID: \"402ac70d-73e7-4697-ac46-a92bbbaf1aac\") " pod="openstack/ovsdbserver-sb-0" Dec 11 02:20:11 crc kubenswrapper[4824]: I1211 02:20:11.454478 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"402ac70d-73e7-4697-ac46-a92bbbaf1aac\") " pod="openstack/ovsdbserver-sb-0" Dec 11 02:20:11 crc kubenswrapper[4824]: I1211 02:20:11.578381 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 11 02:20:12 crc kubenswrapper[4824]: I1211 02:20:12.442000 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 11 02:20:12 crc kubenswrapper[4824]: I1211 02:20:12.443533 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 11 02:20:12 crc kubenswrapper[4824]: I1211 02:20:12.450850 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Dec 11 02:20:12 crc kubenswrapper[4824]: I1211 02:20:12.451078 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Dec 11 02:20:12 crc kubenswrapper[4824]: I1211 02:20:12.451319 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Dec 11 02:20:12 crc kubenswrapper[4824]: I1211 02:20:12.452003 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-xchlw" Dec 11 02:20:12 crc kubenswrapper[4824]: I1211 02:20:12.462531 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 11 02:20:12 crc kubenswrapper[4824]: I1211 02:20:12.548634 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/068df61c-d193-4911-b077-fd80d25fa4de-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"068df61c-d193-4911-b077-fd80d25fa4de\") " pod="openstack/ovsdbserver-nb-0" Dec 11 02:20:12 crc kubenswrapper[4824]: I1211 02:20:12.548705 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-nb-0\" (UID: \"068df61c-d193-4911-b077-fd80d25fa4de\") " pod="openstack/ovsdbserver-nb-0" Dec 11 02:20:12 crc kubenswrapper[4824]: I1211 02:20:12.548729 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/068df61c-d193-4911-b077-fd80d25fa4de-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"068df61c-d193-4911-b077-fd80d25fa4de\") " pod="openstack/ovsdbserver-nb-0" Dec 11 02:20:12 crc kubenswrapper[4824]: I1211 02:20:12.548769 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/068df61c-d193-4911-b077-fd80d25fa4de-config\") pod \"ovsdbserver-nb-0\" (UID: \"068df61c-d193-4911-b077-fd80d25fa4de\") " pod="openstack/ovsdbserver-nb-0" Dec 11 02:20:12 crc kubenswrapper[4824]: I1211 02:20:12.548797 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/068df61c-d193-4911-b077-fd80d25fa4de-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"068df61c-d193-4911-b077-fd80d25fa4de\") " pod="openstack/ovsdbserver-nb-0" Dec 11 02:20:12 crc kubenswrapper[4824]: I1211 02:20:12.548817 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/068df61c-d193-4911-b077-fd80d25fa4de-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"068df61c-d193-4911-b077-fd80d25fa4de\") " pod="openstack/ovsdbserver-nb-0" Dec 11 02:20:12 crc kubenswrapper[4824]: I1211 02:20:12.548835 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/068df61c-d193-4911-b077-fd80d25fa4de-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"068df61c-d193-4911-b077-fd80d25fa4de\") " pod="openstack/ovsdbserver-nb-0" Dec 11 02:20:12 crc kubenswrapper[4824]: I1211 02:20:12.548856 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8c96v\" (UniqueName: \"kubernetes.io/projected/068df61c-d193-4911-b077-fd80d25fa4de-kube-api-access-8c96v\") pod \"ovsdbserver-nb-0\" (UID: \"068df61c-d193-4911-b077-fd80d25fa4de\") " pod="openstack/ovsdbserver-nb-0" Dec 11 02:20:12 crc kubenswrapper[4824]: I1211 02:20:12.649753 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/068df61c-d193-4911-b077-fd80d25fa4de-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"068df61c-d193-4911-b077-fd80d25fa4de\") " pod="openstack/ovsdbserver-nb-0" Dec 11 02:20:12 crc kubenswrapper[4824]: I1211 02:20:12.649804 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/068df61c-d193-4911-b077-fd80d25fa4de-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"068df61c-d193-4911-b077-fd80d25fa4de\") " pod="openstack/ovsdbserver-nb-0" Dec 11 02:20:12 crc kubenswrapper[4824]: I1211 02:20:12.649829 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/068df61c-d193-4911-b077-fd80d25fa4de-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"068df61c-d193-4911-b077-fd80d25fa4de\") " pod="openstack/ovsdbserver-nb-0" Dec 11 02:20:12 crc kubenswrapper[4824]: I1211 02:20:12.649854 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8c96v\" (UniqueName: \"kubernetes.io/projected/068df61c-d193-4911-b077-fd80d25fa4de-kube-api-access-8c96v\") pod \"ovsdbserver-nb-0\" (UID: \"068df61c-d193-4911-b077-fd80d25fa4de\") " pod="openstack/ovsdbserver-nb-0" Dec 11 02:20:12 crc kubenswrapper[4824]: I1211 02:20:12.649940 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/068df61c-d193-4911-b077-fd80d25fa4de-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"068df61c-d193-4911-b077-fd80d25fa4de\") " pod="openstack/ovsdbserver-nb-0" Dec 11 02:20:12 crc kubenswrapper[4824]: I1211 02:20:12.650358 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-nb-0\" (UID: \"068df61c-d193-4911-b077-fd80d25fa4de\") " pod="openstack/ovsdbserver-nb-0" Dec 11 02:20:12 crc kubenswrapper[4824]: I1211 02:20:12.650381 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/068df61c-d193-4911-b077-fd80d25fa4de-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"068df61c-d193-4911-b077-fd80d25fa4de\") " pod="openstack/ovsdbserver-nb-0" Dec 11 02:20:12 crc kubenswrapper[4824]: I1211 02:20:12.650444 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/068df61c-d193-4911-b077-fd80d25fa4de-config\") pod \"ovsdbserver-nb-0\" (UID: \"068df61c-d193-4911-b077-fd80d25fa4de\") " pod="openstack/ovsdbserver-nb-0" Dec 11 02:20:12 crc kubenswrapper[4824]: I1211 02:20:12.650571 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-nb-0\" (UID: \"068df61c-d193-4911-b077-fd80d25fa4de\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/ovsdbserver-nb-0" Dec 11 02:20:12 crc kubenswrapper[4824]: I1211 02:20:12.650587 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/068df61c-d193-4911-b077-fd80d25fa4de-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"068df61c-d193-4911-b077-fd80d25fa4de\") " pod="openstack/ovsdbserver-nb-0" Dec 11 02:20:12 crc kubenswrapper[4824]: I1211 02:20:12.651592 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/068df61c-d193-4911-b077-fd80d25fa4de-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"068df61c-d193-4911-b077-fd80d25fa4de\") " pod="openstack/ovsdbserver-nb-0" Dec 11 02:20:12 crc kubenswrapper[4824]: I1211 02:20:12.652625 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/068df61c-d193-4911-b077-fd80d25fa4de-config\") pod \"ovsdbserver-nb-0\" (UID: \"068df61c-d193-4911-b077-fd80d25fa4de\") " pod="openstack/ovsdbserver-nb-0" Dec 11 02:20:12 crc kubenswrapper[4824]: I1211 02:20:12.658458 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/068df61c-d193-4911-b077-fd80d25fa4de-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"068df61c-d193-4911-b077-fd80d25fa4de\") " pod="openstack/ovsdbserver-nb-0" Dec 11 02:20:12 crc kubenswrapper[4824]: I1211 02:20:12.664705 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/068df61c-d193-4911-b077-fd80d25fa4de-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"068df61c-d193-4911-b077-fd80d25fa4de\") " pod="openstack/ovsdbserver-nb-0" Dec 11 02:20:12 crc kubenswrapper[4824]: I1211 02:20:12.667612 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/068df61c-d193-4911-b077-fd80d25fa4de-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"068df61c-d193-4911-b077-fd80d25fa4de\") " pod="openstack/ovsdbserver-nb-0" Dec 11 02:20:12 crc kubenswrapper[4824]: I1211 02:20:12.670309 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8c96v\" (UniqueName: \"kubernetes.io/projected/068df61c-d193-4911-b077-fd80d25fa4de-kube-api-access-8c96v\") pod \"ovsdbserver-nb-0\" (UID: \"068df61c-d193-4911-b077-fd80d25fa4de\") " pod="openstack/ovsdbserver-nb-0" Dec 11 02:20:12 crc kubenswrapper[4824]: I1211 02:20:12.680279 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-nb-0\" (UID: \"068df61c-d193-4911-b077-fd80d25fa4de\") " pod="openstack/ovsdbserver-nb-0" Dec 11 02:20:12 crc kubenswrapper[4824]: I1211 02:20:12.774441 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 11 02:20:18 crc kubenswrapper[4824]: E1211 02:20:18.566839 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 11 02:20:18 crc kubenswrapper[4824]: E1211 02:20:18.567554 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-z45c9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-57d769cc4f-8bq9w_openstack(d72d278a-3895-4b90-abed-0905680540c8): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 11 02:20:18 crc kubenswrapper[4824]: E1211 02:20:18.568743 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-57d769cc4f-8bq9w" podUID="d72d278a-3895-4b90-abed-0905680540c8" Dec 11 02:20:18 crc kubenswrapper[4824]: E1211 02:20:18.590264 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 11 02:20:18 crc kubenswrapper[4824]: E1211 02:20:18.590459 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tvn4n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-qwcd8_openstack(1081fac2-0512-4e5a-b05e-3999af5d4159): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 11 02:20:18 crc kubenswrapper[4824]: E1211 02:20:18.591627 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-qwcd8" podUID="1081fac2-0512-4e5a-b05e-3999af5d4159" Dec 11 02:20:18 crc kubenswrapper[4824]: E1211 02:20:18.604531 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 11 02:20:18 crc kubenswrapper[4824]: E1211 02:20:18.604702 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hmc5m,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-xhg5m_openstack(e319794c-b4f6-4ae6-b4fa-746aa4028edd): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 11 02:20:18 crc kubenswrapper[4824]: E1211 02:20:18.605895 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-xhg5m" podUID="e319794c-b4f6-4ae6-b4fa-746aa4028edd" Dec 11 02:20:18 crc kubenswrapper[4824]: E1211 02:20:18.617791 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 11 02:20:18 crc kubenswrapper[4824]: E1211 02:20:18.618063 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-766fx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-666b6646f7-r8gdk_openstack(0b407c6b-fc23-459c-8c39-5a13c9889de5): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 11 02:20:18 crc kubenswrapper[4824]: E1211 02:20:18.619316 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-666b6646f7-r8gdk" podUID="0b407c6b-fc23-459c-8c39-5a13c9889de5" Dec 11 02:20:19 crc kubenswrapper[4824]: I1211 02:20:19.088767 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 11 02:20:19 crc kubenswrapper[4824]: I1211 02:20:19.129923 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"709f3d23-a475-443a-8787-7f018c486b56","Type":"ContainerStarted","Data":"54be2f976c76c5ada5b8524d1da7b48c325497a23b57283ea82e9ad0893bba0f"} Dec 11 02:20:19 crc kubenswrapper[4824]: E1211 02:20:19.130094 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-57d769cc4f-8bq9w" podUID="d72d278a-3895-4b90-abed-0905680540c8" Dec 11 02:20:19 crc kubenswrapper[4824]: E1211 02:20:19.133387 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-666b6646f7-r8gdk" podUID="0b407c6b-fc23-459c-8c39-5a13c9889de5" Dec 11 02:20:19 crc kubenswrapper[4824]: I1211 02:20:19.409360 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 11 02:20:19 crc kubenswrapper[4824]: I1211 02:20:19.421071 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 11 02:20:19 crc kubenswrapper[4824]: I1211 02:20:19.430759 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 11 02:20:19 crc kubenswrapper[4824]: I1211 02:20:19.441956 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 11 02:20:19 crc kubenswrapper[4824]: I1211 02:20:19.450405 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-779x4"] Dec 11 02:20:19 crc kubenswrapper[4824]: I1211 02:20:19.455476 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 11 02:20:19 crc kubenswrapper[4824]: I1211 02:20:19.538502 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-xhg5m" Dec 11 02:20:19 crc kubenswrapper[4824]: I1211 02:20:19.590277 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 11 02:20:19 crc kubenswrapper[4824]: W1211 02:20:19.593052 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod068df61c_d193_4911_b077_fd80d25fa4de.slice/crio-4e0655fbf18ebc1b71e76cbd3042954717883c95460ff7e0b230488318ba472e WatchSource:0}: Error finding container 4e0655fbf18ebc1b71e76cbd3042954717883c95460ff7e0b230488318ba472e: Status 404 returned error can't find the container with id 4e0655fbf18ebc1b71e76cbd3042954717883c95460ff7e0b230488318ba472e Dec 11 02:20:19 crc kubenswrapper[4824]: I1211 02:20:19.603770 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-qwcd8" Dec 11 02:20:19 crc kubenswrapper[4824]: I1211 02:20:19.682047 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e319794c-b4f6-4ae6-b4fa-746aa4028edd-config\") pod \"e319794c-b4f6-4ae6-b4fa-746aa4028edd\" (UID: \"e319794c-b4f6-4ae6-b4fa-746aa4028edd\") " Dec 11 02:20:19 crc kubenswrapper[4824]: I1211 02:20:19.682092 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hmc5m\" (UniqueName: \"kubernetes.io/projected/e319794c-b4f6-4ae6-b4fa-746aa4028edd-kube-api-access-hmc5m\") pod \"e319794c-b4f6-4ae6-b4fa-746aa4028edd\" (UID: \"e319794c-b4f6-4ae6-b4fa-746aa4028edd\") " Dec 11 02:20:19 crc kubenswrapper[4824]: I1211 02:20:19.682143 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e319794c-b4f6-4ae6-b4fa-746aa4028edd-dns-svc\") pod \"e319794c-b4f6-4ae6-b4fa-746aa4028edd\" (UID: \"e319794c-b4f6-4ae6-b4fa-746aa4028edd\") " Dec 11 02:20:19 crc kubenswrapper[4824]: I1211 02:20:19.682634 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e319794c-b4f6-4ae6-b4fa-746aa4028edd-config" (OuterVolumeSpecName: "config") pod "e319794c-b4f6-4ae6-b4fa-746aa4028edd" (UID: "e319794c-b4f6-4ae6-b4fa-746aa4028edd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:20:19 crc kubenswrapper[4824]: I1211 02:20:19.682643 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e319794c-b4f6-4ae6-b4fa-746aa4028edd-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e319794c-b4f6-4ae6-b4fa-746aa4028edd" (UID: "e319794c-b4f6-4ae6-b4fa-746aa4028edd"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:20:19 crc kubenswrapper[4824]: I1211 02:20:19.683631 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-xzkzd"] Dec 11 02:20:19 crc kubenswrapper[4824]: I1211 02:20:19.688608 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e319794c-b4f6-4ae6-b4fa-746aa4028edd-kube-api-access-hmc5m" (OuterVolumeSpecName: "kube-api-access-hmc5m") pod "e319794c-b4f6-4ae6-b4fa-746aa4028edd" (UID: "e319794c-b4f6-4ae6-b4fa-746aa4028edd"). InnerVolumeSpecName "kube-api-access-hmc5m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:20:19 crc kubenswrapper[4824]: W1211 02:20:19.689017 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod17cc80fc_a51f_41e4_be0f_f593fc23476e.slice/crio-93beb9cb8fbbf8251fc51c114a62169462ad759db1472452fef97f2b4e46deed WatchSource:0}: Error finding container 93beb9cb8fbbf8251fc51c114a62169462ad759db1472452fef97f2b4e46deed: Status 404 returned error can't find the container with id 93beb9cb8fbbf8251fc51c114a62169462ad759db1472452fef97f2b4e46deed Dec 11 02:20:19 crc kubenswrapper[4824]: I1211 02:20:19.784428 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tvn4n\" (UniqueName: \"kubernetes.io/projected/1081fac2-0512-4e5a-b05e-3999af5d4159-kube-api-access-tvn4n\") pod \"1081fac2-0512-4e5a-b05e-3999af5d4159\" (UID: \"1081fac2-0512-4e5a-b05e-3999af5d4159\") " Dec 11 02:20:19 crc kubenswrapper[4824]: I1211 02:20:19.784621 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1081fac2-0512-4e5a-b05e-3999af5d4159-config\") pod \"1081fac2-0512-4e5a-b05e-3999af5d4159\" (UID: \"1081fac2-0512-4e5a-b05e-3999af5d4159\") " Dec 11 02:20:19 crc kubenswrapper[4824]: I1211 02:20:19.785089 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e319794c-b4f6-4ae6-b4fa-746aa4028edd-config\") on node \"crc\" DevicePath \"\"" Dec 11 02:20:19 crc kubenswrapper[4824]: I1211 02:20:19.785105 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hmc5m\" (UniqueName: \"kubernetes.io/projected/e319794c-b4f6-4ae6-b4fa-746aa4028edd-kube-api-access-hmc5m\") on node \"crc\" DevicePath \"\"" Dec 11 02:20:19 crc kubenswrapper[4824]: I1211 02:20:19.785182 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e319794c-b4f6-4ae6-b4fa-746aa4028edd-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 11 02:20:19 crc kubenswrapper[4824]: I1211 02:20:19.786722 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1081fac2-0512-4e5a-b05e-3999af5d4159-config" (OuterVolumeSpecName: "config") pod "1081fac2-0512-4e5a-b05e-3999af5d4159" (UID: "1081fac2-0512-4e5a-b05e-3999af5d4159"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:20:19 crc kubenswrapper[4824]: I1211 02:20:19.792391 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1081fac2-0512-4e5a-b05e-3999af5d4159-kube-api-access-tvn4n" (OuterVolumeSpecName: "kube-api-access-tvn4n") pod "1081fac2-0512-4e5a-b05e-3999af5d4159" (UID: "1081fac2-0512-4e5a-b05e-3999af5d4159"). InnerVolumeSpecName "kube-api-access-tvn4n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:20:19 crc kubenswrapper[4824]: I1211 02:20:19.886730 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tvn4n\" (UniqueName: \"kubernetes.io/projected/1081fac2-0512-4e5a-b05e-3999af5d4159-kube-api-access-tvn4n\") on node \"crc\" DevicePath \"\"" Dec 11 02:20:19 crc kubenswrapper[4824]: I1211 02:20:19.886759 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1081fac2-0512-4e5a-b05e-3999af5d4159-config\") on node \"crc\" DevicePath \"\"" Dec 11 02:20:20 crc kubenswrapper[4824]: I1211 02:20:20.139071 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-xhg5m" event={"ID":"e319794c-b4f6-4ae6-b4fa-746aa4028edd","Type":"ContainerDied","Data":"e5f48fa2892da280d618b5b769e86b9593ae9b364c473dae883691151a5d57ab"} Dec 11 02:20:20 crc kubenswrapper[4824]: I1211 02:20:20.139107 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-xhg5m" Dec 11 02:20:20 crc kubenswrapper[4824]: I1211 02:20:20.140226 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-xzkzd" event={"ID":"17cc80fc-a51f-41e4-be0f-f593fc23476e","Type":"ContainerStarted","Data":"93beb9cb8fbbf8251fc51c114a62169462ad759db1472452fef97f2b4e46deed"} Dec 11 02:20:20 crc kubenswrapper[4824]: I1211 02:20:20.141428 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-qwcd8" event={"ID":"1081fac2-0512-4e5a-b05e-3999af5d4159","Type":"ContainerDied","Data":"16e94f0a27abffff6837104803bf49a14f2d372f10aa85ba1b8ffc0b35b9c964"} Dec 11 02:20:20 crc kubenswrapper[4824]: I1211 02:20:20.141433 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-qwcd8" Dec 11 02:20:20 crc kubenswrapper[4824]: I1211 02:20:20.143018 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"f35efb32-a031-4c75-9358-3f5143335131","Type":"ContainerStarted","Data":"4a332c8f472db58852b476859f91ed98d3ad8b65e016b93fb69a82f6bb4e3b83"} Dec 11 02:20:20 crc kubenswrapper[4824]: I1211 02:20:20.144478 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"068df61c-d193-4911-b077-fd80d25fa4de","Type":"ContainerStarted","Data":"4e0655fbf18ebc1b71e76cbd3042954717883c95460ff7e0b230488318ba472e"} Dec 11 02:20:20 crc kubenswrapper[4824]: I1211 02:20:20.145699 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"6e5994f4-2c9e-4e8c-b7d3-84a15af846cd","Type":"ContainerStarted","Data":"89be154ebedec532d6462eeeeb10d02ae428876373d471a7e8dcf7c7d53e877f"} Dec 11 02:20:20 crc kubenswrapper[4824]: I1211 02:20:20.146821 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ff37fa2f-dc06-430a-a4fe-af2b379fe563","Type":"ContainerStarted","Data":"80a3ece43e3b693e934325a4e77de52fbe6239b597684802908dd120ff83d31c"} Dec 11 02:20:20 crc kubenswrapper[4824]: I1211 02:20:20.149250 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1fd840b6-4ace-45b8-a8d2-38a041110095","Type":"ContainerStarted","Data":"7e13edf4baa20214b9db9d0a84b13be73c96e113fe3b4a0da3a5bcd9654b8599"} Dec 11 02:20:20 crc kubenswrapper[4824]: I1211 02:20:20.150301 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"5b375617-b0a6-42cf-be23-ef7f755f9eff","Type":"ContainerStarted","Data":"07ee1898b339c95b30e9f829c3cb73411381c4743aba8b18ad5e7b8d5c5a3aa6"} Dec 11 02:20:20 crc kubenswrapper[4824]: I1211 02:20:20.151567 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-779x4" event={"ID":"148c69f8-3121-4a55-b3a6-be44b816f643","Type":"ContainerStarted","Data":"42ea4e6464ed721152e8efbe085bf1bae1a8995eb476f05c0264054908c7ad9c"} Dec 11 02:20:20 crc kubenswrapper[4824]: I1211 02:20:20.200667 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-qwcd8"] Dec 11 02:20:20 crc kubenswrapper[4824]: I1211 02:20:20.207699 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-qwcd8"] Dec 11 02:20:20 crc kubenswrapper[4824]: I1211 02:20:20.246396 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-xhg5m"] Dec 11 02:20:20 crc kubenswrapper[4824]: I1211 02:20:20.269911 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-xhg5m"] Dec 11 02:20:20 crc kubenswrapper[4824]: I1211 02:20:20.311869 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 11 02:20:20 crc kubenswrapper[4824]: I1211 02:20:20.643218 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1081fac2-0512-4e5a-b05e-3999af5d4159" path="/var/lib/kubelet/pods/1081fac2-0512-4e5a-b05e-3999af5d4159/volumes" Dec 11 02:20:20 crc kubenswrapper[4824]: I1211 02:20:20.643584 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e319794c-b4f6-4ae6-b4fa-746aa4028edd" path="/var/lib/kubelet/pods/e319794c-b4f6-4ae6-b4fa-746aa4028edd/volumes" Dec 11 02:20:20 crc kubenswrapper[4824]: I1211 02:20:20.865828 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-9fj44"] Dec 11 02:20:20 crc kubenswrapper[4824]: I1211 02:20:20.866802 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-9fj44" Dec 11 02:20:20 crc kubenswrapper[4824]: I1211 02:20:20.869122 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Dec 11 02:20:20 crc kubenswrapper[4824]: I1211 02:20:20.876219 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-9fj44"] Dec 11 02:20:20 crc kubenswrapper[4824]: I1211 02:20:20.911801 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/d4fa84ff-af7c-4d69-9dfd-e4c1f20b0ce9-ovn-rundir\") pod \"ovn-controller-metrics-9fj44\" (UID: \"d4fa84ff-af7c-4d69-9dfd-e4c1f20b0ce9\") " pod="openstack/ovn-controller-metrics-9fj44" Dec 11 02:20:20 crc kubenswrapper[4824]: I1211 02:20:20.911857 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4fa84ff-af7c-4d69-9dfd-e4c1f20b0ce9-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-9fj44\" (UID: \"d4fa84ff-af7c-4d69-9dfd-e4c1f20b0ce9\") " pod="openstack/ovn-controller-metrics-9fj44" Dec 11 02:20:20 crc kubenswrapper[4824]: I1211 02:20:20.911967 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4fa84ff-af7c-4d69-9dfd-e4c1f20b0ce9-config\") pod \"ovn-controller-metrics-9fj44\" (UID: \"d4fa84ff-af7c-4d69-9dfd-e4c1f20b0ce9\") " pod="openstack/ovn-controller-metrics-9fj44" Dec 11 02:20:20 crc kubenswrapper[4824]: I1211 02:20:20.911988 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j495q\" (UniqueName: \"kubernetes.io/projected/d4fa84ff-af7c-4d69-9dfd-e4c1f20b0ce9-kube-api-access-j495q\") pod \"ovn-controller-metrics-9fj44\" (UID: \"d4fa84ff-af7c-4d69-9dfd-e4c1f20b0ce9\") " pod="openstack/ovn-controller-metrics-9fj44" Dec 11 02:20:20 crc kubenswrapper[4824]: I1211 02:20:20.912008 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4fa84ff-af7c-4d69-9dfd-e4c1f20b0ce9-combined-ca-bundle\") pod \"ovn-controller-metrics-9fj44\" (UID: \"d4fa84ff-af7c-4d69-9dfd-e4c1f20b0ce9\") " pod="openstack/ovn-controller-metrics-9fj44" Dec 11 02:20:20 crc kubenswrapper[4824]: I1211 02:20:20.912045 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/d4fa84ff-af7c-4d69-9dfd-e4c1f20b0ce9-ovs-rundir\") pod \"ovn-controller-metrics-9fj44\" (UID: \"d4fa84ff-af7c-4d69-9dfd-e4c1f20b0ce9\") " pod="openstack/ovn-controller-metrics-9fj44" Dec 11 02:20:20 crc kubenswrapper[4824]: I1211 02:20:20.994050 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-r8gdk"] Dec 11 02:20:21 crc kubenswrapper[4824]: I1211 02:20:21.015500 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j495q\" (UniqueName: \"kubernetes.io/projected/d4fa84ff-af7c-4d69-9dfd-e4c1f20b0ce9-kube-api-access-j495q\") pod \"ovn-controller-metrics-9fj44\" (UID: \"d4fa84ff-af7c-4d69-9dfd-e4c1f20b0ce9\") " pod="openstack/ovn-controller-metrics-9fj44" Dec 11 02:20:21 crc kubenswrapper[4824]: I1211 02:20:21.015567 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4fa84ff-af7c-4d69-9dfd-e4c1f20b0ce9-combined-ca-bundle\") pod \"ovn-controller-metrics-9fj44\" (UID: \"d4fa84ff-af7c-4d69-9dfd-e4c1f20b0ce9\") " pod="openstack/ovn-controller-metrics-9fj44" Dec 11 02:20:21 crc kubenswrapper[4824]: I1211 02:20:21.015616 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/d4fa84ff-af7c-4d69-9dfd-e4c1f20b0ce9-ovs-rundir\") pod \"ovn-controller-metrics-9fj44\" (UID: \"d4fa84ff-af7c-4d69-9dfd-e4c1f20b0ce9\") " pod="openstack/ovn-controller-metrics-9fj44" Dec 11 02:20:21 crc kubenswrapper[4824]: I1211 02:20:21.015665 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/d4fa84ff-af7c-4d69-9dfd-e4c1f20b0ce9-ovn-rundir\") pod \"ovn-controller-metrics-9fj44\" (UID: \"d4fa84ff-af7c-4d69-9dfd-e4c1f20b0ce9\") " pod="openstack/ovn-controller-metrics-9fj44" Dec 11 02:20:21 crc kubenswrapper[4824]: I1211 02:20:21.015688 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4fa84ff-af7c-4d69-9dfd-e4c1f20b0ce9-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-9fj44\" (UID: \"d4fa84ff-af7c-4d69-9dfd-e4c1f20b0ce9\") " pod="openstack/ovn-controller-metrics-9fj44" Dec 11 02:20:21 crc kubenswrapper[4824]: I1211 02:20:21.015755 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4fa84ff-af7c-4d69-9dfd-e4c1f20b0ce9-config\") pod \"ovn-controller-metrics-9fj44\" (UID: \"d4fa84ff-af7c-4d69-9dfd-e4c1f20b0ce9\") " pod="openstack/ovn-controller-metrics-9fj44" Dec 11 02:20:21 crc kubenswrapper[4824]: I1211 02:20:21.016260 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/d4fa84ff-af7c-4d69-9dfd-e4c1f20b0ce9-ovs-rundir\") pod \"ovn-controller-metrics-9fj44\" (UID: \"d4fa84ff-af7c-4d69-9dfd-e4c1f20b0ce9\") " pod="openstack/ovn-controller-metrics-9fj44" Dec 11 02:20:21 crc kubenswrapper[4824]: I1211 02:20:21.016512 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4fa84ff-af7c-4d69-9dfd-e4c1f20b0ce9-config\") pod \"ovn-controller-metrics-9fj44\" (UID: \"d4fa84ff-af7c-4d69-9dfd-e4c1f20b0ce9\") " pod="openstack/ovn-controller-metrics-9fj44" Dec 11 02:20:21 crc kubenswrapper[4824]: I1211 02:20:21.016602 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/d4fa84ff-af7c-4d69-9dfd-e4c1f20b0ce9-ovn-rundir\") pod \"ovn-controller-metrics-9fj44\" (UID: \"d4fa84ff-af7c-4d69-9dfd-e4c1f20b0ce9\") " pod="openstack/ovn-controller-metrics-9fj44" Dec 11 02:20:21 crc kubenswrapper[4824]: I1211 02:20:21.030953 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-c7jtf"] Dec 11 02:20:21 crc kubenswrapper[4824]: I1211 02:20:21.032330 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4fa84ff-af7c-4d69-9dfd-e4c1f20b0ce9-combined-ca-bundle\") pod \"ovn-controller-metrics-9fj44\" (UID: \"d4fa84ff-af7c-4d69-9dfd-e4c1f20b0ce9\") " pod="openstack/ovn-controller-metrics-9fj44" Dec 11 02:20:21 crc kubenswrapper[4824]: I1211 02:20:21.047136 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j495q\" (UniqueName: \"kubernetes.io/projected/d4fa84ff-af7c-4d69-9dfd-e4c1f20b0ce9-kube-api-access-j495q\") pod \"ovn-controller-metrics-9fj44\" (UID: \"d4fa84ff-af7c-4d69-9dfd-e4c1f20b0ce9\") " pod="openstack/ovn-controller-metrics-9fj44" Dec 11 02:20:21 crc kubenswrapper[4824]: I1211 02:20:21.047611 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f896c8c65-c7jtf" Dec 11 02:20:21 crc kubenswrapper[4824]: I1211 02:20:21.050492 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4fa84ff-af7c-4d69-9dfd-e4c1f20b0ce9-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-9fj44\" (UID: \"d4fa84ff-af7c-4d69-9dfd-e4c1f20b0ce9\") " pod="openstack/ovn-controller-metrics-9fj44" Dec 11 02:20:21 crc kubenswrapper[4824]: I1211 02:20:21.060505 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Dec 11 02:20:21 crc kubenswrapper[4824]: I1211 02:20:21.072168 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-c7jtf"] Dec 11 02:20:21 crc kubenswrapper[4824]: I1211 02:20:21.140691 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-8bq9w"] Dec 11 02:20:21 crc kubenswrapper[4824]: I1211 02:20:21.175465 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"402ac70d-73e7-4697-ac46-a92bbbaf1aac","Type":"ContainerStarted","Data":"9ee7057d3b65c15f2671a8aa7b9d0fb710901703ff489c8f7d6703d9e5f3dbf0"} Dec 11 02:20:21 crc kubenswrapper[4824]: I1211 02:20:21.199360 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-9fj44" Dec 11 02:20:21 crc kubenswrapper[4824]: I1211 02:20:21.204413 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-2cvmc"] Dec 11 02:20:21 crc kubenswrapper[4824]: I1211 02:20:21.206313 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-2cvmc" Dec 11 02:20:21 crc kubenswrapper[4824]: I1211 02:20:21.211779 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-2cvmc"] Dec 11 02:20:21 crc kubenswrapper[4824]: I1211 02:20:21.212737 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Dec 11 02:20:21 crc kubenswrapper[4824]: I1211 02:20:21.219181 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35d304da-8076-40eb-8314-f4ff5405b6f4-config\") pod \"dnsmasq-dns-7f896c8c65-c7jtf\" (UID: \"35d304da-8076-40eb-8314-f4ff5405b6f4\") " pod="openstack/dnsmasq-dns-7f896c8c65-c7jtf" Dec 11 02:20:21 crc kubenswrapper[4824]: I1211 02:20:21.219251 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/35d304da-8076-40eb-8314-f4ff5405b6f4-ovsdbserver-sb\") pod \"dnsmasq-dns-7f896c8c65-c7jtf\" (UID: \"35d304da-8076-40eb-8314-f4ff5405b6f4\") " pod="openstack/dnsmasq-dns-7f896c8c65-c7jtf" Dec 11 02:20:21 crc kubenswrapper[4824]: I1211 02:20:21.219324 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/35d304da-8076-40eb-8314-f4ff5405b6f4-dns-svc\") pod \"dnsmasq-dns-7f896c8c65-c7jtf\" (UID: \"35d304da-8076-40eb-8314-f4ff5405b6f4\") " pod="openstack/dnsmasq-dns-7f896c8c65-c7jtf" Dec 11 02:20:21 crc kubenswrapper[4824]: I1211 02:20:21.219542 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxzb9\" (UniqueName: \"kubernetes.io/projected/35d304da-8076-40eb-8314-f4ff5405b6f4-kube-api-access-qxzb9\") pod \"dnsmasq-dns-7f896c8c65-c7jtf\" (UID: \"35d304da-8076-40eb-8314-f4ff5405b6f4\") " pod="openstack/dnsmasq-dns-7f896c8c65-c7jtf" Dec 11 02:20:21 crc kubenswrapper[4824]: I1211 02:20:21.321194 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/35d304da-8076-40eb-8314-f4ff5405b6f4-ovsdbserver-sb\") pod \"dnsmasq-dns-7f896c8c65-c7jtf\" (UID: \"35d304da-8076-40eb-8314-f4ff5405b6f4\") " pod="openstack/dnsmasq-dns-7f896c8c65-c7jtf" Dec 11 02:20:21 crc kubenswrapper[4824]: I1211 02:20:21.321248 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fdb2591b-b56c-48f0-82a4-c5752e9107b0-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-2cvmc\" (UID: \"fdb2591b-b56c-48f0-82a4-c5752e9107b0\") " pod="openstack/dnsmasq-dns-86db49b7ff-2cvmc" Dec 11 02:20:21 crc kubenswrapper[4824]: I1211 02:20:21.321468 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fdb2591b-b56c-48f0-82a4-c5752e9107b0-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-2cvmc\" (UID: \"fdb2591b-b56c-48f0-82a4-c5752e9107b0\") " pod="openstack/dnsmasq-dns-86db49b7ff-2cvmc" Dec 11 02:20:21 crc kubenswrapper[4824]: I1211 02:20:21.321504 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/35d304da-8076-40eb-8314-f4ff5405b6f4-dns-svc\") pod \"dnsmasq-dns-7f896c8c65-c7jtf\" (UID: \"35d304da-8076-40eb-8314-f4ff5405b6f4\") " pod="openstack/dnsmasq-dns-7f896c8c65-c7jtf" Dec 11 02:20:21 crc kubenswrapper[4824]: I1211 02:20:21.321529 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fdb2591b-b56c-48f0-82a4-c5752e9107b0-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-2cvmc\" (UID: \"fdb2591b-b56c-48f0-82a4-c5752e9107b0\") " pod="openstack/dnsmasq-dns-86db49b7ff-2cvmc" Dec 11 02:20:21 crc kubenswrapper[4824]: I1211 02:20:21.321583 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qxzb9\" (UniqueName: \"kubernetes.io/projected/35d304da-8076-40eb-8314-f4ff5405b6f4-kube-api-access-qxzb9\") pod \"dnsmasq-dns-7f896c8c65-c7jtf\" (UID: \"35d304da-8076-40eb-8314-f4ff5405b6f4\") " pod="openstack/dnsmasq-dns-7f896c8c65-c7jtf" Dec 11 02:20:21 crc kubenswrapper[4824]: I1211 02:20:21.321602 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l8jjl\" (UniqueName: \"kubernetes.io/projected/fdb2591b-b56c-48f0-82a4-c5752e9107b0-kube-api-access-l8jjl\") pod \"dnsmasq-dns-86db49b7ff-2cvmc\" (UID: \"fdb2591b-b56c-48f0-82a4-c5752e9107b0\") " pod="openstack/dnsmasq-dns-86db49b7ff-2cvmc" Dec 11 02:20:21 crc kubenswrapper[4824]: I1211 02:20:21.321638 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35d304da-8076-40eb-8314-f4ff5405b6f4-config\") pod \"dnsmasq-dns-7f896c8c65-c7jtf\" (UID: \"35d304da-8076-40eb-8314-f4ff5405b6f4\") " pod="openstack/dnsmasq-dns-7f896c8c65-c7jtf" Dec 11 02:20:21 crc kubenswrapper[4824]: I1211 02:20:21.321660 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fdb2591b-b56c-48f0-82a4-c5752e9107b0-config\") pod \"dnsmasq-dns-86db49b7ff-2cvmc\" (UID: \"fdb2591b-b56c-48f0-82a4-c5752e9107b0\") " pod="openstack/dnsmasq-dns-86db49b7ff-2cvmc" Dec 11 02:20:21 crc kubenswrapper[4824]: I1211 02:20:21.323245 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/35d304da-8076-40eb-8314-f4ff5405b6f4-ovsdbserver-sb\") pod \"dnsmasq-dns-7f896c8c65-c7jtf\" (UID: \"35d304da-8076-40eb-8314-f4ff5405b6f4\") " pod="openstack/dnsmasq-dns-7f896c8c65-c7jtf" Dec 11 02:20:21 crc kubenswrapper[4824]: I1211 02:20:21.323522 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/35d304da-8076-40eb-8314-f4ff5405b6f4-dns-svc\") pod \"dnsmasq-dns-7f896c8c65-c7jtf\" (UID: \"35d304da-8076-40eb-8314-f4ff5405b6f4\") " pod="openstack/dnsmasq-dns-7f896c8c65-c7jtf" Dec 11 02:20:21 crc kubenswrapper[4824]: I1211 02:20:21.324233 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35d304da-8076-40eb-8314-f4ff5405b6f4-config\") pod \"dnsmasq-dns-7f896c8c65-c7jtf\" (UID: \"35d304da-8076-40eb-8314-f4ff5405b6f4\") " pod="openstack/dnsmasq-dns-7f896c8c65-c7jtf" Dec 11 02:20:21 crc kubenswrapper[4824]: I1211 02:20:21.353645 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxzb9\" (UniqueName: \"kubernetes.io/projected/35d304da-8076-40eb-8314-f4ff5405b6f4-kube-api-access-qxzb9\") pod \"dnsmasq-dns-7f896c8c65-c7jtf\" (UID: \"35d304da-8076-40eb-8314-f4ff5405b6f4\") " pod="openstack/dnsmasq-dns-7f896c8c65-c7jtf" Dec 11 02:20:21 crc kubenswrapper[4824]: I1211 02:20:21.423341 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fdb2591b-b56c-48f0-82a4-c5752e9107b0-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-2cvmc\" (UID: \"fdb2591b-b56c-48f0-82a4-c5752e9107b0\") " pod="openstack/dnsmasq-dns-86db49b7ff-2cvmc" Dec 11 02:20:21 crc kubenswrapper[4824]: I1211 02:20:21.423397 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fdb2591b-b56c-48f0-82a4-c5752e9107b0-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-2cvmc\" (UID: \"fdb2591b-b56c-48f0-82a4-c5752e9107b0\") " pod="openstack/dnsmasq-dns-86db49b7ff-2cvmc" Dec 11 02:20:21 crc kubenswrapper[4824]: I1211 02:20:21.423438 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fdb2591b-b56c-48f0-82a4-c5752e9107b0-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-2cvmc\" (UID: \"fdb2591b-b56c-48f0-82a4-c5752e9107b0\") " pod="openstack/dnsmasq-dns-86db49b7ff-2cvmc" Dec 11 02:20:21 crc kubenswrapper[4824]: I1211 02:20:21.423509 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l8jjl\" (UniqueName: \"kubernetes.io/projected/fdb2591b-b56c-48f0-82a4-c5752e9107b0-kube-api-access-l8jjl\") pod \"dnsmasq-dns-86db49b7ff-2cvmc\" (UID: \"fdb2591b-b56c-48f0-82a4-c5752e9107b0\") " pod="openstack/dnsmasq-dns-86db49b7ff-2cvmc" Dec 11 02:20:21 crc kubenswrapper[4824]: I1211 02:20:21.423555 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fdb2591b-b56c-48f0-82a4-c5752e9107b0-config\") pod \"dnsmasq-dns-86db49b7ff-2cvmc\" (UID: \"fdb2591b-b56c-48f0-82a4-c5752e9107b0\") " pod="openstack/dnsmasq-dns-86db49b7ff-2cvmc" Dec 11 02:20:21 crc kubenswrapper[4824]: I1211 02:20:21.424452 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fdb2591b-b56c-48f0-82a4-c5752e9107b0-config\") pod \"dnsmasq-dns-86db49b7ff-2cvmc\" (UID: \"fdb2591b-b56c-48f0-82a4-c5752e9107b0\") " pod="openstack/dnsmasq-dns-86db49b7ff-2cvmc" Dec 11 02:20:21 crc kubenswrapper[4824]: I1211 02:20:21.425018 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fdb2591b-b56c-48f0-82a4-c5752e9107b0-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-2cvmc\" (UID: \"fdb2591b-b56c-48f0-82a4-c5752e9107b0\") " pod="openstack/dnsmasq-dns-86db49b7ff-2cvmc" Dec 11 02:20:21 crc kubenswrapper[4824]: I1211 02:20:21.425619 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fdb2591b-b56c-48f0-82a4-c5752e9107b0-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-2cvmc\" (UID: \"fdb2591b-b56c-48f0-82a4-c5752e9107b0\") " pod="openstack/dnsmasq-dns-86db49b7ff-2cvmc" Dec 11 02:20:21 crc kubenswrapper[4824]: I1211 02:20:21.426123 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fdb2591b-b56c-48f0-82a4-c5752e9107b0-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-2cvmc\" (UID: \"fdb2591b-b56c-48f0-82a4-c5752e9107b0\") " pod="openstack/dnsmasq-dns-86db49b7ff-2cvmc" Dec 11 02:20:21 crc kubenswrapper[4824]: I1211 02:20:21.433572 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f896c8c65-c7jtf" Dec 11 02:20:21 crc kubenswrapper[4824]: I1211 02:20:21.441650 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l8jjl\" (UniqueName: \"kubernetes.io/projected/fdb2591b-b56c-48f0-82a4-c5752e9107b0-kube-api-access-l8jjl\") pod \"dnsmasq-dns-86db49b7ff-2cvmc\" (UID: \"fdb2591b-b56c-48f0-82a4-c5752e9107b0\") " pod="openstack/dnsmasq-dns-86db49b7ff-2cvmc" Dec 11 02:20:21 crc kubenswrapper[4824]: I1211 02:20:21.528722 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-2cvmc" Dec 11 02:20:23 crc kubenswrapper[4824]: I1211 02:20:23.160615 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-8bq9w" Dec 11 02:20:23 crc kubenswrapper[4824]: I1211 02:20:23.198939 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-8bq9w" event={"ID":"d72d278a-3895-4b90-abed-0905680540c8","Type":"ContainerDied","Data":"864eba193e2630c696553c6d5038d28695e3aaf1224f61c024446584b949fbb9"} Dec 11 02:20:23 crc kubenswrapper[4824]: I1211 02:20:23.198992 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-8bq9w" Dec 11 02:20:23 crc kubenswrapper[4824]: I1211 02:20:23.354429 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d72d278a-3895-4b90-abed-0905680540c8-dns-svc\") pod \"d72d278a-3895-4b90-abed-0905680540c8\" (UID: \"d72d278a-3895-4b90-abed-0905680540c8\") " Dec 11 02:20:23 crc kubenswrapper[4824]: I1211 02:20:23.354639 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d72d278a-3895-4b90-abed-0905680540c8-config\") pod \"d72d278a-3895-4b90-abed-0905680540c8\" (UID: \"d72d278a-3895-4b90-abed-0905680540c8\") " Dec 11 02:20:23 crc kubenswrapper[4824]: I1211 02:20:23.354694 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z45c9\" (UniqueName: \"kubernetes.io/projected/d72d278a-3895-4b90-abed-0905680540c8-kube-api-access-z45c9\") pod \"d72d278a-3895-4b90-abed-0905680540c8\" (UID: \"d72d278a-3895-4b90-abed-0905680540c8\") " Dec 11 02:20:23 crc kubenswrapper[4824]: I1211 02:20:23.355266 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d72d278a-3895-4b90-abed-0905680540c8-config" (OuterVolumeSpecName: "config") pod "d72d278a-3895-4b90-abed-0905680540c8" (UID: "d72d278a-3895-4b90-abed-0905680540c8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:20:23 crc kubenswrapper[4824]: I1211 02:20:23.356176 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d72d278a-3895-4b90-abed-0905680540c8-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d72d278a-3895-4b90-abed-0905680540c8" (UID: "d72d278a-3895-4b90-abed-0905680540c8"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:20:23 crc kubenswrapper[4824]: I1211 02:20:23.363063 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d72d278a-3895-4b90-abed-0905680540c8-kube-api-access-z45c9" (OuterVolumeSpecName: "kube-api-access-z45c9") pod "d72d278a-3895-4b90-abed-0905680540c8" (UID: "d72d278a-3895-4b90-abed-0905680540c8"). InnerVolumeSpecName "kube-api-access-z45c9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:20:23 crc kubenswrapper[4824]: I1211 02:20:23.457438 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d72d278a-3895-4b90-abed-0905680540c8-config\") on node \"crc\" DevicePath \"\"" Dec 11 02:20:23 crc kubenswrapper[4824]: I1211 02:20:23.457486 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z45c9\" (UniqueName: \"kubernetes.io/projected/d72d278a-3895-4b90-abed-0905680540c8-kube-api-access-z45c9\") on node \"crc\" DevicePath \"\"" Dec 11 02:20:23 crc kubenswrapper[4824]: I1211 02:20:23.457508 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d72d278a-3895-4b90-abed-0905680540c8-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 11 02:20:23 crc kubenswrapper[4824]: I1211 02:20:23.577472 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-8bq9w"] Dec 11 02:20:23 crc kubenswrapper[4824]: I1211 02:20:23.585422 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-8bq9w"] Dec 11 02:20:24 crc kubenswrapper[4824]: I1211 02:20:24.550863 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-r8gdk" Dec 11 02:20:24 crc kubenswrapper[4824]: I1211 02:20:24.647669 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d72d278a-3895-4b90-abed-0905680540c8" path="/var/lib/kubelet/pods/d72d278a-3895-4b90-abed-0905680540c8/volumes" Dec 11 02:20:24 crc kubenswrapper[4824]: I1211 02:20:24.675544 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b407c6b-fc23-459c-8c39-5a13c9889de5-config\") pod \"0b407c6b-fc23-459c-8c39-5a13c9889de5\" (UID: \"0b407c6b-fc23-459c-8c39-5a13c9889de5\") " Dec 11 02:20:24 crc kubenswrapper[4824]: I1211 02:20:24.676422 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b407c6b-fc23-459c-8c39-5a13c9889de5-config" (OuterVolumeSpecName: "config") pod "0b407c6b-fc23-459c-8c39-5a13c9889de5" (UID: "0b407c6b-fc23-459c-8c39-5a13c9889de5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:20:24 crc kubenswrapper[4824]: I1211 02:20:24.676734 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0b407c6b-fc23-459c-8c39-5a13c9889de5-dns-svc\") pod \"0b407c6b-fc23-459c-8c39-5a13c9889de5\" (UID: \"0b407c6b-fc23-459c-8c39-5a13c9889de5\") " Dec 11 02:20:24 crc kubenswrapper[4824]: I1211 02:20:24.677309 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b407c6b-fc23-459c-8c39-5a13c9889de5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0b407c6b-fc23-459c-8c39-5a13c9889de5" (UID: "0b407c6b-fc23-459c-8c39-5a13c9889de5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:20:24 crc kubenswrapper[4824]: I1211 02:20:24.677533 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-766fx\" (UniqueName: \"kubernetes.io/projected/0b407c6b-fc23-459c-8c39-5a13c9889de5-kube-api-access-766fx\") pod \"0b407c6b-fc23-459c-8c39-5a13c9889de5\" (UID: \"0b407c6b-fc23-459c-8c39-5a13c9889de5\") " Dec 11 02:20:24 crc kubenswrapper[4824]: I1211 02:20:24.678572 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0b407c6b-fc23-459c-8c39-5a13c9889de5-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 11 02:20:24 crc kubenswrapper[4824]: I1211 02:20:24.678763 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b407c6b-fc23-459c-8c39-5a13c9889de5-config\") on node \"crc\" DevicePath \"\"" Dec 11 02:20:24 crc kubenswrapper[4824]: I1211 02:20:24.681634 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b407c6b-fc23-459c-8c39-5a13c9889de5-kube-api-access-766fx" (OuterVolumeSpecName: "kube-api-access-766fx") pod "0b407c6b-fc23-459c-8c39-5a13c9889de5" (UID: "0b407c6b-fc23-459c-8c39-5a13c9889de5"). InnerVolumeSpecName "kube-api-access-766fx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:20:24 crc kubenswrapper[4824]: I1211 02:20:24.782805 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-766fx\" (UniqueName: \"kubernetes.io/projected/0b407c6b-fc23-459c-8c39-5a13c9889de5-kube-api-access-766fx\") on node \"crc\" DevicePath \"\"" Dec 11 02:20:25 crc kubenswrapper[4824]: I1211 02:20:25.216678 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-r8gdk" event={"ID":"0b407c6b-fc23-459c-8c39-5a13c9889de5","Type":"ContainerDied","Data":"742e56327af98155873a8a08f312d8b5917a1c51667dc827b44e6569bc8299ee"} Dec 11 02:20:25 crc kubenswrapper[4824]: I1211 02:20:25.216802 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-r8gdk" Dec 11 02:20:25 crc kubenswrapper[4824]: I1211 02:20:25.265991 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-r8gdk"] Dec 11 02:20:25 crc kubenswrapper[4824]: I1211 02:20:25.276228 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-r8gdk"] Dec 11 02:20:26 crc kubenswrapper[4824]: I1211 02:20:26.645976 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b407c6b-fc23-459c-8c39-5a13c9889de5" path="/var/lib/kubelet/pods/0b407c6b-fc23-459c-8c39-5a13c9889de5/volumes" Dec 11 02:20:27 crc kubenswrapper[4824]: I1211 02:20:27.809453 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-c7jtf"] Dec 11 02:20:28 crc kubenswrapper[4824]: I1211 02:20:28.105296 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-9fj44"] Dec 11 02:20:28 crc kubenswrapper[4824]: I1211 02:20:28.111296 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-2cvmc"] Dec 11 02:20:28 crc kubenswrapper[4824]: W1211 02:20:28.166738 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd4fa84ff_af7c_4d69_9dfd_e4c1f20b0ce9.slice/crio-155230e54851ea73d4a9f748e7b6f37e5774ab93fb0e8ff3b2297336950dbf06 WatchSource:0}: Error finding container 155230e54851ea73d4a9f748e7b6f37e5774ab93fb0e8ff3b2297336950dbf06: Status 404 returned error can't find the container with id 155230e54851ea73d4a9f748e7b6f37e5774ab93fb0e8ff3b2297336950dbf06 Dec 11 02:20:28 crc kubenswrapper[4824]: I1211 02:20:28.245213 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-9fj44" event={"ID":"d4fa84ff-af7c-4d69-9dfd-e4c1f20b0ce9","Type":"ContainerStarted","Data":"155230e54851ea73d4a9f748e7b6f37e5774ab93fb0e8ff3b2297336950dbf06"} Dec 11 02:20:28 crc kubenswrapper[4824]: I1211 02:20:28.246372 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-2cvmc" event={"ID":"fdb2591b-b56c-48f0-82a4-c5752e9107b0","Type":"ContainerStarted","Data":"e0b8f12355030c117508163acfd8d5ca4c6c33999e3771b455c8e77865647a1d"} Dec 11 02:20:28 crc kubenswrapper[4824]: I1211 02:20:28.248284 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f896c8c65-c7jtf" event={"ID":"35d304da-8076-40eb-8314-f4ff5405b6f4","Type":"ContainerStarted","Data":"17695eb12243c44967c773d72597a8dd8e1c14f36ed21c57a368474369ccf1fd"} Dec 11 02:20:29 crc kubenswrapper[4824]: I1211 02:20:29.260403 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"709f3d23-a475-443a-8787-7f018c486b56","Type":"ContainerStarted","Data":"cb6564f8c56a42e0113ce2b764f7930f24b08c26044bc6c4fd4ce7ffa08b652a"} Dec 11 02:20:29 crc kubenswrapper[4824]: I1211 02:20:29.262485 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Dec 11 02:20:29 crc kubenswrapper[4824]: I1211 02:20:29.266117 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"f35efb32-a031-4c75-9358-3f5143335131","Type":"ContainerStarted","Data":"f42863989f38164739df4031a6b1a3b20510960a4faebf5c0e8f1893ad80f3f2"} Dec 11 02:20:29 crc kubenswrapper[4824]: I1211 02:20:29.288821 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=18.27676611 podStartE2EDuration="26.288795753s" podCreationTimestamp="2025-12-11 02:20:03 +0000 UTC" firstStartedPulling="2025-12-11 02:20:19.086828167 +0000 UTC m=+1160.775865546" lastFinishedPulling="2025-12-11 02:20:27.09885776 +0000 UTC m=+1168.787895189" observedRunningTime="2025-12-11 02:20:29.281328929 +0000 UTC m=+1170.970366338" watchObservedRunningTime="2025-12-11 02:20:29.288795753 +0000 UTC m=+1170.977833142" Dec 11 02:20:30 crc kubenswrapper[4824]: I1211 02:20:30.276933 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1fd840b6-4ace-45b8-a8d2-38a041110095","Type":"ContainerStarted","Data":"690854c03401211031fd0831cdfab50c3cc32ac8bb1ccb2f7ab9e7897c3a6efe"} Dec 11 02:20:30 crc kubenswrapper[4824]: I1211 02:20:30.282775 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"5b375617-b0a6-42cf-be23-ef7f755f9eff","Type":"ContainerStarted","Data":"552c5666299888465ed8c969e420b63ce34abef304f83696d7c2337684b297b3"} Dec 11 02:20:30 crc kubenswrapper[4824]: I1211 02:20:30.283291 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 11 02:20:30 crc kubenswrapper[4824]: I1211 02:20:30.285853 4824 generic.go:334] "Generic (PLEG): container finished" podID="35d304da-8076-40eb-8314-f4ff5405b6f4" containerID="e6b504b5d603835a10b87caa4f8c6f591f8d57fb3917df1c9d68b01dc53242cd" exitCode=0 Dec 11 02:20:30 crc kubenswrapper[4824]: I1211 02:20:30.285975 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f896c8c65-c7jtf" event={"ID":"35d304da-8076-40eb-8314-f4ff5405b6f4","Type":"ContainerDied","Data":"e6b504b5d603835a10b87caa4f8c6f591f8d57fb3917df1c9d68b01dc53242cd"} Dec 11 02:20:30 crc kubenswrapper[4824]: I1211 02:20:30.288867 4824 generic.go:334] "Generic (PLEG): container finished" podID="17cc80fc-a51f-41e4-be0f-f593fc23476e" containerID="89ac95b7b600490e0fbaeb25083be9a86cd46d0138176b078f17ed3e03225982" exitCode=0 Dec 11 02:20:30 crc kubenswrapper[4824]: I1211 02:20:30.289153 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-xzkzd" event={"ID":"17cc80fc-a51f-41e4-be0f-f593fc23476e","Type":"ContainerDied","Data":"89ac95b7b600490e0fbaeb25083be9a86cd46d0138176b078f17ed3e03225982"} Dec 11 02:20:30 crc kubenswrapper[4824]: I1211 02:20:30.290983 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"402ac70d-73e7-4697-ac46-a92bbbaf1aac","Type":"ContainerStarted","Data":"b291dfd92f68ebd55f735105586d328ac25ad7f054996f4a4bb6ef87bcaa122d"} Dec 11 02:20:30 crc kubenswrapper[4824]: I1211 02:20:30.292219 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"6e5994f4-2c9e-4e8c-b7d3-84a15af846cd","Type":"ContainerStarted","Data":"3b99562ce83856431e7d4a515eadf3e8d75bc6c1b5714952876d95921f26ce6e"} Dec 11 02:20:30 crc kubenswrapper[4824]: I1211 02:20:30.294322 4824 generic.go:334] "Generic (PLEG): container finished" podID="fdb2591b-b56c-48f0-82a4-c5752e9107b0" containerID="d023802bf723a4402fc9ac5bb3d51be9aae723d79401b98aa6f5c63fb8ff67f1" exitCode=0 Dec 11 02:20:30 crc kubenswrapper[4824]: I1211 02:20:30.294398 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-2cvmc" event={"ID":"fdb2591b-b56c-48f0-82a4-c5752e9107b0","Type":"ContainerDied","Data":"d023802bf723a4402fc9ac5bb3d51be9aae723d79401b98aa6f5c63fb8ff67f1"} Dec 11 02:20:30 crc kubenswrapper[4824]: I1211 02:20:30.298145 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"068df61c-d193-4911-b077-fd80d25fa4de","Type":"ContainerStarted","Data":"161bd14a12acba3198e7ad147a9c02c349dd8f1ee7e17ba609847dc0914f3ac1"} Dec 11 02:20:30 crc kubenswrapper[4824]: I1211 02:20:30.300872 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-779x4" event={"ID":"148c69f8-3121-4a55-b3a6-be44b816f643","Type":"ContainerStarted","Data":"1c6c1b8396f80c7d818f3844597562fa9d054b1aba2db81573b4ce0450e9bdbc"} Dec 11 02:20:30 crc kubenswrapper[4824]: I1211 02:20:30.380327 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=15.714058632 podStartE2EDuration="25.380306978s" podCreationTimestamp="2025-12-11 02:20:05 +0000 UTC" firstStartedPulling="2025-12-11 02:20:19.42866637 +0000 UTC m=+1161.117703749" lastFinishedPulling="2025-12-11 02:20:29.094914706 +0000 UTC m=+1170.783952095" observedRunningTime="2025-12-11 02:20:30.370411816 +0000 UTC m=+1172.059449225" watchObservedRunningTime="2025-12-11 02:20:30.380306978 +0000 UTC m=+1172.069344357" Dec 11 02:20:30 crc kubenswrapper[4824]: I1211 02:20:30.423553 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-779x4" podStartSLOduration=13.104968286 podStartE2EDuration="21.423489292s" podCreationTimestamp="2025-12-11 02:20:09 +0000 UTC" firstStartedPulling="2025-12-11 02:20:19.480723616 +0000 UTC m=+1161.169760995" lastFinishedPulling="2025-12-11 02:20:27.799244602 +0000 UTC m=+1169.488282001" observedRunningTime="2025-12-11 02:20:30.421879278 +0000 UTC m=+1172.110916837" watchObservedRunningTime="2025-12-11 02:20:30.423489292 +0000 UTC m=+1172.112526671" Dec 11 02:20:31 crc kubenswrapper[4824]: I1211 02:20:31.316678 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ff37fa2f-dc06-430a-a4fe-af2b379fe563","Type":"ContainerStarted","Data":"1d02f24dc933f44c12ca63c4374960b6bf5f14ea8005dd772dbd07f54f6c8f51"} Dec 11 02:20:31 crc kubenswrapper[4824]: I1211 02:20:31.317034 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-779x4" Dec 11 02:20:33 crc kubenswrapper[4824]: I1211 02:20:33.335988 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-xzkzd" event={"ID":"17cc80fc-a51f-41e4-be0f-f593fc23476e","Type":"ContainerStarted","Data":"8dabb38a32c27a077918098b26d3ab2dfff27ea48ada6ae29c5d4db8b947c8eb"} Dec 11 02:20:33 crc kubenswrapper[4824]: I1211 02:20:33.336624 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-xzkzd" event={"ID":"17cc80fc-a51f-41e4-be0f-f593fc23476e","Type":"ContainerStarted","Data":"f5441f8c2515dead3879d849a13fcc79eb2a6d9a4508f301be46819aacc5922d"} Dec 11 02:20:33 crc kubenswrapper[4824]: I1211 02:20:33.336655 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-xzkzd" Dec 11 02:20:33 crc kubenswrapper[4824]: I1211 02:20:33.336678 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-xzkzd" Dec 11 02:20:33 crc kubenswrapper[4824]: I1211 02:20:33.339878 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"402ac70d-73e7-4697-ac46-a92bbbaf1aac","Type":"ContainerStarted","Data":"f7636d8483b5c004e0b07dd67cd362f5a1c7b35d80a3542df5f6956ad6bdbbdb"} Dec 11 02:20:33 crc kubenswrapper[4824]: I1211 02:20:33.343181 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-2cvmc" event={"ID":"fdb2591b-b56c-48f0-82a4-c5752e9107b0","Type":"ContainerStarted","Data":"7e10ab73d64d3d111fbfbd8c0b836b6565fe58bf39f1c603e715c1d9ca2abab1"} Dec 11 02:20:33 crc kubenswrapper[4824]: I1211 02:20:33.343348 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-86db49b7ff-2cvmc" Dec 11 02:20:33 crc kubenswrapper[4824]: I1211 02:20:33.345893 4824 generic.go:334] "Generic (PLEG): container finished" podID="f35efb32-a031-4c75-9358-3f5143335131" containerID="f42863989f38164739df4031a6b1a3b20510960a4faebf5c0e8f1893ad80f3f2" exitCode=0 Dec 11 02:20:33 crc kubenswrapper[4824]: I1211 02:20:33.345973 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"f35efb32-a031-4c75-9358-3f5143335131","Type":"ContainerDied","Data":"f42863989f38164739df4031a6b1a3b20510960a4faebf5c0e8f1893ad80f3f2"} Dec 11 02:20:33 crc kubenswrapper[4824]: I1211 02:20:33.349191 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"068df61c-d193-4911-b077-fd80d25fa4de","Type":"ContainerStarted","Data":"ff001d2cd0eb79e1a4c69e0fef1a91dad20e4deba67733292ab060c2edd3bfbd"} Dec 11 02:20:33 crc kubenswrapper[4824]: I1211 02:20:33.353080 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f896c8c65-c7jtf" event={"ID":"35d304da-8076-40eb-8314-f4ff5405b6f4","Type":"ContainerStarted","Data":"226339d7735db5963c656beebd3d1b62220a6c88f98dac9e9fa12270179191c7"} Dec 11 02:20:33 crc kubenswrapper[4824]: I1211 02:20:33.353411 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7f896c8c65-c7jtf" Dec 11 02:20:33 crc kubenswrapper[4824]: I1211 02:20:33.356234 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-9fj44" event={"ID":"d4fa84ff-af7c-4d69-9dfd-e4c1f20b0ce9","Type":"ContainerStarted","Data":"2c8e97acd4630d5e53f6f5906e5ee60cf91c111e87c7c2e20eb18600253c5590"} Dec 11 02:20:33 crc kubenswrapper[4824]: I1211 02:20:33.370016 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-xzkzd" podStartSLOduration=16.612778033 podStartE2EDuration="24.370000381s" podCreationTimestamp="2025-12-11 02:20:09 +0000 UTC" firstStartedPulling="2025-12-11 02:20:19.691709405 +0000 UTC m=+1161.380746784" lastFinishedPulling="2025-12-11 02:20:27.448931723 +0000 UTC m=+1169.137969132" observedRunningTime="2025-12-11 02:20:33.369696592 +0000 UTC m=+1175.058734031" watchObservedRunningTime="2025-12-11 02:20:33.370000381 +0000 UTC m=+1175.059037760" Dec 11 02:20:33 crc kubenswrapper[4824]: I1211 02:20:33.402142 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-86db49b7ff-2cvmc" podStartSLOduration=11.391506025 podStartE2EDuration="12.402128662s" podCreationTimestamp="2025-12-11 02:20:21 +0000 UTC" firstStartedPulling="2025-12-11 02:20:28.159470301 +0000 UTC m=+1169.848507680" lastFinishedPulling="2025-12-11 02:20:29.170092898 +0000 UTC m=+1170.859130317" observedRunningTime="2025-12-11 02:20:33.401004121 +0000 UTC m=+1175.090041500" watchObservedRunningTime="2025-12-11 02:20:33.402128662 +0000 UTC m=+1175.091166041" Dec 11 02:20:33 crc kubenswrapper[4824]: I1211 02:20:33.437565 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=9.764190674 podStartE2EDuration="22.437531892s" podCreationTimestamp="2025-12-11 02:20:11 +0000 UTC" firstStartedPulling="2025-12-11 02:20:19.595086782 +0000 UTC m=+1161.284124161" lastFinishedPulling="2025-12-11 02:20:32.268428 +0000 UTC m=+1173.957465379" observedRunningTime="2025-12-11 02:20:33.425975476 +0000 UTC m=+1175.115012855" watchObservedRunningTime="2025-12-11 02:20:33.437531892 +0000 UTC m=+1175.126569321" Dec 11 02:20:33 crc kubenswrapper[4824]: I1211 02:20:33.483981 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-9fj44" podStartSLOduration=9.403592241 podStartE2EDuration="13.483956156s" podCreationTimestamp="2025-12-11 02:20:20 +0000 UTC" firstStartedPulling="2025-12-11 02:20:28.17001966 +0000 UTC m=+1169.859057069" lastFinishedPulling="2025-12-11 02:20:32.250383605 +0000 UTC m=+1173.939420984" observedRunningTime="2025-12-11 02:20:33.464999346 +0000 UTC m=+1175.154036755" watchObservedRunningTime="2025-12-11 02:20:33.483956156 +0000 UTC m=+1175.172993545" Dec 11 02:20:33 crc kubenswrapper[4824]: I1211 02:20:33.498242 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=11.735491756 podStartE2EDuration="23.498224877s" podCreationTimestamp="2025-12-11 02:20:10 +0000 UTC" firstStartedPulling="2025-12-11 02:20:20.511465917 +0000 UTC m=+1162.200503296" lastFinishedPulling="2025-12-11 02:20:32.274199038 +0000 UTC m=+1173.963236417" observedRunningTime="2025-12-11 02:20:33.48956817 +0000 UTC m=+1175.178605549" watchObservedRunningTime="2025-12-11 02:20:33.498224877 +0000 UTC m=+1175.187262256" Dec 11 02:20:33 crc kubenswrapper[4824]: I1211 02:20:33.519220 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7f896c8c65-c7jtf" podStartSLOduration=12.231665887 podStartE2EDuration="13.519204882s" podCreationTimestamp="2025-12-11 02:20:20 +0000 UTC" firstStartedPulling="2025-12-11 02:20:27.878989295 +0000 UTC m=+1169.568026664" lastFinishedPulling="2025-12-11 02:20:29.16652825 +0000 UTC m=+1170.855565659" observedRunningTime="2025-12-11 02:20:33.516878949 +0000 UTC m=+1175.205916338" watchObservedRunningTime="2025-12-11 02:20:33.519204882 +0000 UTC m=+1175.208242251" Dec 11 02:20:33 crc kubenswrapper[4824]: I1211 02:20:33.775537 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Dec 11 02:20:33 crc kubenswrapper[4824]: I1211 02:20:33.809406 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Dec 11 02:20:33 crc kubenswrapper[4824]: I1211 02:20:33.883323 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Dec 11 02:20:34 crc kubenswrapper[4824]: I1211 02:20:34.370201 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"f35efb32-a031-4c75-9358-3f5143335131","Type":"ContainerStarted","Data":"ef08382c7c3d08de1b4c0a5c9e73cef4cc770cea77cca4d3fc81dccc4f6d87e7"} Dec 11 02:20:34 crc kubenswrapper[4824]: I1211 02:20:34.372585 4824 generic.go:334] "Generic (PLEG): container finished" podID="6e5994f4-2c9e-4e8c-b7d3-84a15af846cd" containerID="3b99562ce83856431e7d4a515eadf3e8d75bc6c1b5714952876d95921f26ce6e" exitCode=0 Dec 11 02:20:34 crc kubenswrapper[4824]: I1211 02:20:34.373675 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"6e5994f4-2c9e-4e8c-b7d3-84a15af846cd","Type":"ContainerDied","Data":"3b99562ce83856431e7d4a515eadf3e8d75bc6c1b5714952876d95921f26ce6e"} Dec 11 02:20:34 crc kubenswrapper[4824]: I1211 02:20:34.375791 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Dec 11 02:20:34 crc kubenswrapper[4824]: I1211 02:20:34.416671 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=24.094711349 podStartE2EDuration="32.416640975s" podCreationTimestamp="2025-12-11 02:20:02 +0000 UTC" firstStartedPulling="2025-12-11 02:20:19.476805542 +0000 UTC m=+1161.165842921" lastFinishedPulling="2025-12-11 02:20:27.798735168 +0000 UTC m=+1169.487772547" observedRunningTime="2025-12-11 02:20:34.408507712 +0000 UTC m=+1176.097545131" watchObservedRunningTime="2025-12-11 02:20:34.416640975 +0000 UTC m=+1176.105678384" Dec 11 02:20:34 crc kubenswrapper[4824]: I1211 02:20:34.447238 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Dec 11 02:20:35 crc kubenswrapper[4824]: I1211 02:20:35.386860 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"6e5994f4-2c9e-4e8c-b7d3-84a15af846cd","Type":"ContainerStarted","Data":"b54050286c626acd3c3b12429cdeeedb181e75e0b3de3d328634c2dee2f634b4"} Dec 11 02:20:35 crc kubenswrapper[4824]: I1211 02:20:35.431204 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=26.898416241 podStartE2EDuration="35.431177499s" podCreationTimestamp="2025-12-11 02:20:00 +0000 UTC" firstStartedPulling="2025-12-11 02:20:19.477514701 +0000 UTC m=+1161.166552070" lastFinishedPulling="2025-12-11 02:20:28.010275949 +0000 UTC m=+1169.699313328" observedRunningTime="2025-12-11 02:20:35.418580503 +0000 UTC m=+1177.107617872" watchObservedRunningTime="2025-12-11 02:20:35.431177499 +0000 UTC m=+1177.120214918" Dec 11 02:20:35 crc kubenswrapper[4824]: I1211 02:20:35.579003 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Dec 11 02:20:35 crc kubenswrapper[4824]: I1211 02:20:35.659841 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Dec 11 02:20:35 crc kubenswrapper[4824]: I1211 02:20:35.916292 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-c7jtf"] Dec 11 02:20:35 crc kubenswrapper[4824]: I1211 02:20:35.916660 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7f896c8c65-c7jtf" podUID="35d304da-8076-40eb-8314-f4ff5405b6f4" containerName="dnsmasq-dns" containerID="cri-o://226339d7735db5963c656beebd3d1b62220a6c88f98dac9e9fa12270179191c7" gracePeriod=10 Dec 11 02:20:35 crc kubenswrapper[4824]: I1211 02:20:35.944800 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 11 02:20:36 crc kubenswrapper[4824]: I1211 02:20:36.016812 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-698758b865-5j6cw"] Dec 11 02:20:36 crc kubenswrapper[4824]: I1211 02:20:36.018513 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-5j6cw" Dec 11 02:20:36 crc kubenswrapper[4824]: I1211 02:20:36.030411 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-5j6cw"] Dec 11 02:20:36 crc kubenswrapper[4824]: E1211 02:20:36.136488 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod35d304da_8076_40eb_8314_f4ff5405b6f4.slice/crio-conmon-226339d7735db5963c656beebd3d1b62220a6c88f98dac9e9fa12270179191c7.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod35d304da_8076_40eb_8314_f4ff5405b6f4.slice/crio-226339d7735db5963c656beebd3d1b62220a6c88f98dac9e9fa12270179191c7.scope\": RecentStats: unable to find data in memory cache]" Dec 11 02:20:36 crc kubenswrapper[4824]: I1211 02:20:36.190200 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/342737ab-7fa3-4654-9493-e826dacbaa6f-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-5j6cw\" (UID: \"342737ab-7fa3-4654-9493-e826dacbaa6f\") " pod="openstack/dnsmasq-dns-698758b865-5j6cw" Dec 11 02:20:36 crc kubenswrapper[4824]: I1211 02:20:36.190538 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/342737ab-7fa3-4654-9493-e826dacbaa6f-config\") pod \"dnsmasq-dns-698758b865-5j6cw\" (UID: \"342737ab-7fa3-4654-9493-e826dacbaa6f\") " pod="openstack/dnsmasq-dns-698758b865-5j6cw" Dec 11 02:20:36 crc kubenswrapper[4824]: I1211 02:20:36.190636 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/342737ab-7fa3-4654-9493-e826dacbaa6f-dns-svc\") pod \"dnsmasq-dns-698758b865-5j6cw\" (UID: \"342737ab-7fa3-4654-9493-e826dacbaa6f\") " pod="openstack/dnsmasq-dns-698758b865-5j6cw" Dec 11 02:20:36 crc kubenswrapper[4824]: I1211 02:20:36.190665 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/342737ab-7fa3-4654-9493-e826dacbaa6f-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-5j6cw\" (UID: \"342737ab-7fa3-4654-9493-e826dacbaa6f\") " pod="openstack/dnsmasq-dns-698758b865-5j6cw" Dec 11 02:20:36 crc kubenswrapper[4824]: I1211 02:20:36.190698 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zkx8k\" (UniqueName: \"kubernetes.io/projected/342737ab-7fa3-4654-9493-e826dacbaa6f-kube-api-access-zkx8k\") pod \"dnsmasq-dns-698758b865-5j6cw\" (UID: \"342737ab-7fa3-4654-9493-e826dacbaa6f\") " pod="openstack/dnsmasq-dns-698758b865-5j6cw" Dec 11 02:20:36 crc kubenswrapper[4824]: I1211 02:20:36.292502 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/342737ab-7fa3-4654-9493-e826dacbaa6f-config\") pod \"dnsmasq-dns-698758b865-5j6cw\" (UID: \"342737ab-7fa3-4654-9493-e826dacbaa6f\") " pod="openstack/dnsmasq-dns-698758b865-5j6cw" Dec 11 02:20:36 crc kubenswrapper[4824]: I1211 02:20:36.292596 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/342737ab-7fa3-4654-9493-e826dacbaa6f-dns-svc\") pod \"dnsmasq-dns-698758b865-5j6cw\" (UID: \"342737ab-7fa3-4654-9493-e826dacbaa6f\") " pod="openstack/dnsmasq-dns-698758b865-5j6cw" Dec 11 02:20:36 crc kubenswrapper[4824]: I1211 02:20:36.292626 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/342737ab-7fa3-4654-9493-e826dacbaa6f-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-5j6cw\" (UID: \"342737ab-7fa3-4654-9493-e826dacbaa6f\") " pod="openstack/dnsmasq-dns-698758b865-5j6cw" Dec 11 02:20:36 crc kubenswrapper[4824]: I1211 02:20:36.292650 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zkx8k\" (UniqueName: \"kubernetes.io/projected/342737ab-7fa3-4654-9493-e826dacbaa6f-kube-api-access-zkx8k\") pod \"dnsmasq-dns-698758b865-5j6cw\" (UID: \"342737ab-7fa3-4654-9493-e826dacbaa6f\") " pod="openstack/dnsmasq-dns-698758b865-5j6cw" Dec 11 02:20:36 crc kubenswrapper[4824]: I1211 02:20:36.292681 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/342737ab-7fa3-4654-9493-e826dacbaa6f-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-5j6cw\" (UID: \"342737ab-7fa3-4654-9493-e826dacbaa6f\") " pod="openstack/dnsmasq-dns-698758b865-5j6cw" Dec 11 02:20:36 crc kubenswrapper[4824]: I1211 02:20:36.293610 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/342737ab-7fa3-4654-9493-e826dacbaa6f-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-5j6cw\" (UID: \"342737ab-7fa3-4654-9493-e826dacbaa6f\") " pod="openstack/dnsmasq-dns-698758b865-5j6cw" Dec 11 02:20:36 crc kubenswrapper[4824]: I1211 02:20:36.293667 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/342737ab-7fa3-4654-9493-e826dacbaa6f-config\") pod \"dnsmasq-dns-698758b865-5j6cw\" (UID: \"342737ab-7fa3-4654-9493-e826dacbaa6f\") " pod="openstack/dnsmasq-dns-698758b865-5j6cw" Dec 11 02:20:36 crc kubenswrapper[4824]: I1211 02:20:36.293804 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/342737ab-7fa3-4654-9493-e826dacbaa6f-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-5j6cw\" (UID: \"342737ab-7fa3-4654-9493-e826dacbaa6f\") " pod="openstack/dnsmasq-dns-698758b865-5j6cw" Dec 11 02:20:36 crc kubenswrapper[4824]: I1211 02:20:36.294794 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/342737ab-7fa3-4654-9493-e826dacbaa6f-dns-svc\") pod \"dnsmasq-dns-698758b865-5j6cw\" (UID: \"342737ab-7fa3-4654-9493-e826dacbaa6f\") " pod="openstack/dnsmasq-dns-698758b865-5j6cw" Dec 11 02:20:36 crc kubenswrapper[4824]: I1211 02:20:36.314424 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zkx8k\" (UniqueName: \"kubernetes.io/projected/342737ab-7fa3-4654-9493-e826dacbaa6f-kube-api-access-zkx8k\") pod \"dnsmasq-dns-698758b865-5j6cw\" (UID: \"342737ab-7fa3-4654-9493-e826dacbaa6f\") " pod="openstack/dnsmasq-dns-698758b865-5j6cw" Dec 11 02:20:36 crc kubenswrapper[4824]: I1211 02:20:36.378247 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f896c8c65-c7jtf" Dec 11 02:20:36 crc kubenswrapper[4824]: I1211 02:20:36.401892 4824 generic.go:334] "Generic (PLEG): container finished" podID="35d304da-8076-40eb-8314-f4ff5405b6f4" containerID="226339d7735db5963c656beebd3d1b62220a6c88f98dac9e9fa12270179191c7" exitCode=0 Dec 11 02:20:36 crc kubenswrapper[4824]: I1211 02:20:36.402272 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f896c8c65-c7jtf" event={"ID":"35d304da-8076-40eb-8314-f4ff5405b6f4","Type":"ContainerDied","Data":"226339d7735db5963c656beebd3d1b62220a6c88f98dac9e9fa12270179191c7"} Dec 11 02:20:36 crc kubenswrapper[4824]: I1211 02:20:36.402354 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f896c8c65-c7jtf" event={"ID":"35d304da-8076-40eb-8314-f4ff5405b6f4","Type":"ContainerDied","Data":"17695eb12243c44967c773d72597a8dd8e1c14f36ed21c57a368474369ccf1fd"} Dec 11 02:20:36 crc kubenswrapper[4824]: I1211 02:20:36.402405 4824 scope.go:117] "RemoveContainer" containerID="226339d7735db5963c656beebd3d1b62220a6c88f98dac9e9fa12270179191c7" Dec 11 02:20:36 crc kubenswrapper[4824]: I1211 02:20:36.402506 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f896c8c65-c7jtf" Dec 11 02:20:36 crc kubenswrapper[4824]: I1211 02:20:36.402641 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Dec 11 02:20:36 crc kubenswrapper[4824]: I1211 02:20:36.411781 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-5j6cw" Dec 11 02:20:36 crc kubenswrapper[4824]: I1211 02:20:36.472829 4824 scope.go:117] "RemoveContainer" containerID="e6b504b5d603835a10b87caa4f8c6f591f8d57fb3917df1c9d68b01dc53242cd" Dec 11 02:20:36 crc kubenswrapper[4824]: I1211 02:20:36.472996 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Dec 11 02:20:36 crc kubenswrapper[4824]: I1211 02:20:36.501415 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qxzb9\" (UniqueName: \"kubernetes.io/projected/35d304da-8076-40eb-8314-f4ff5405b6f4-kube-api-access-qxzb9\") pod \"35d304da-8076-40eb-8314-f4ff5405b6f4\" (UID: \"35d304da-8076-40eb-8314-f4ff5405b6f4\") " Dec 11 02:20:36 crc kubenswrapper[4824]: I1211 02:20:36.501723 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/35d304da-8076-40eb-8314-f4ff5405b6f4-dns-svc\") pod \"35d304da-8076-40eb-8314-f4ff5405b6f4\" (UID: \"35d304da-8076-40eb-8314-f4ff5405b6f4\") " Dec 11 02:20:36 crc kubenswrapper[4824]: I1211 02:20:36.501852 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/35d304da-8076-40eb-8314-f4ff5405b6f4-ovsdbserver-sb\") pod \"35d304da-8076-40eb-8314-f4ff5405b6f4\" (UID: \"35d304da-8076-40eb-8314-f4ff5405b6f4\") " Dec 11 02:20:36 crc kubenswrapper[4824]: I1211 02:20:36.501927 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35d304da-8076-40eb-8314-f4ff5405b6f4-config\") pod \"35d304da-8076-40eb-8314-f4ff5405b6f4\" (UID: \"35d304da-8076-40eb-8314-f4ff5405b6f4\") " Dec 11 02:20:36 crc kubenswrapper[4824]: I1211 02:20:36.507598 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35d304da-8076-40eb-8314-f4ff5405b6f4-kube-api-access-qxzb9" (OuterVolumeSpecName: "kube-api-access-qxzb9") pod "35d304da-8076-40eb-8314-f4ff5405b6f4" (UID: "35d304da-8076-40eb-8314-f4ff5405b6f4"). InnerVolumeSpecName "kube-api-access-qxzb9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:20:36 crc kubenswrapper[4824]: I1211 02:20:36.543505 4824 scope.go:117] "RemoveContainer" containerID="226339d7735db5963c656beebd3d1b62220a6c88f98dac9e9fa12270179191c7" Dec 11 02:20:36 crc kubenswrapper[4824]: E1211 02:20:36.545474 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"226339d7735db5963c656beebd3d1b62220a6c88f98dac9e9fa12270179191c7\": container with ID starting with 226339d7735db5963c656beebd3d1b62220a6c88f98dac9e9fa12270179191c7 not found: ID does not exist" containerID="226339d7735db5963c656beebd3d1b62220a6c88f98dac9e9fa12270179191c7" Dec 11 02:20:36 crc kubenswrapper[4824]: I1211 02:20:36.545519 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"226339d7735db5963c656beebd3d1b62220a6c88f98dac9e9fa12270179191c7"} err="failed to get container status \"226339d7735db5963c656beebd3d1b62220a6c88f98dac9e9fa12270179191c7\": rpc error: code = NotFound desc = could not find container \"226339d7735db5963c656beebd3d1b62220a6c88f98dac9e9fa12270179191c7\": container with ID starting with 226339d7735db5963c656beebd3d1b62220a6c88f98dac9e9fa12270179191c7 not found: ID does not exist" Dec 11 02:20:36 crc kubenswrapper[4824]: I1211 02:20:36.545546 4824 scope.go:117] "RemoveContainer" containerID="e6b504b5d603835a10b87caa4f8c6f591f8d57fb3917df1c9d68b01dc53242cd" Dec 11 02:20:36 crc kubenswrapper[4824]: E1211 02:20:36.545818 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e6b504b5d603835a10b87caa4f8c6f591f8d57fb3917df1c9d68b01dc53242cd\": container with ID starting with e6b504b5d603835a10b87caa4f8c6f591f8d57fb3917df1c9d68b01dc53242cd not found: ID does not exist" containerID="e6b504b5d603835a10b87caa4f8c6f591f8d57fb3917df1c9d68b01dc53242cd" Dec 11 02:20:36 crc kubenswrapper[4824]: I1211 02:20:36.545851 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e6b504b5d603835a10b87caa4f8c6f591f8d57fb3917df1c9d68b01dc53242cd"} err="failed to get container status \"e6b504b5d603835a10b87caa4f8c6f591f8d57fb3917df1c9d68b01dc53242cd\": rpc error: code = NotFound desc = could not find container \"e6b504b5d603835a10b87caa4f8c6f591f8d57fb3917df1c9d68b01dc53242cd\": container with ID starting with e6b504b5d603835a10b87caa4f8c6f591f8d57fb3917df1c9d68b01dc53242cd not found: ID does not exist" Dec 11 02:20:36 crc kubenswrapper[4824]: I1211 02:20:36.550216 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35d304da-8076-40eb-8314-f4ff5405b6f4-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "35d304da-8076-40eb-8314-f4ff5405b6f4" (UID: "35d304da-8076-40eb-8314-f4ff5405b6f4"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:20:36 crc kubenswrapper[4824]: I1211 02:20:36.571951 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35d304da-8076-40eb-8314-f4ff5405b6f4-config" (OuterVolumeSpecName: "config") pod "35d304da-8076-40eb-8314-f4ff5405b6f4" (UID: "35d304da-8076-40eb-8314-f4ff5405b6f4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:20:36 crc kubenswrapper[4824]: I1211 02:20:36.595180 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35d304da-8076-40eb-8314-f4ff5405b6f4-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "35d304da-8076-40eb-8314-f4ff5405b6f4" (UID: "35d304da-8076-40eb-8314-f4ff5405b6f4"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:20:36 crc kubenswrapper[4824]: I1211 02:20:36.603662 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/35d304da-8076-40eb-8314-f4ff5405b6f4-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 11 02:20:36 crc kubenswrapper[4824]: I1211 02:20:36.603682 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35d304da-8076-40eb-8314-f4ff5405b6f4-config\") on node \"crc\" DevicePath \"\"" Dec 11 02:20:36 crc kubenswrapper[4824]: I1211 02:20:36.603691 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qxzb9\" (UniqueName: \"kubernetes.io/projected/35d304da-8076-40eb-8314-f4ff5405b6f4-kube-api-access-qxzb9\") on node \"crc\" DevicePath \"\"" Dec 11 02:20:36 crc kubenswrapper[4824]: I1211 02:20:36.603701 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/35d304da-8076-40eb-8314-f4ff5405b6f4-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 11 02:20:36 crc kubenswrapper[4824]: I1211 02:20:36.698316 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Dec 11 02:20:36 crc kubenswrapper[4824]: E1211 02:20:36.698686 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35d304da-8076-40eb-8314-f4ff5405b6f4" containerName="init" Dec 11 02:20:36 crc kubenswrapper[4824]: I1211 02:20:36.698703 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="35d304da-8076-40eb-8314-f4ff5405b6f4" containerName="init" Dec 11 02:20:36 crc kubenswrapper[4824]: E1211 02:20:36.698717 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35d304da-8076-40eb-8314-f4ff5405b6f4" containerName="dnsmasq-dns" Dec 11 02:20:36 crc kubenswrapper[4824]: I1211 02:20:36.698723 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="35d304da-8076-40eb-8314-f4ff5405b6f4" containerName="dnsmasq-dns" Dec 11 02:20:36 crc kubenswrapper[4824]: I1211 02:20:36.698880 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="35d304da-8076-40eb-8314-f4ff5405b6f4" containerName="dnsmasq-dns" Dec 11 02:20:36 crc kubenswrapper[4824]: I1211 02:20:36.712238 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 11 02:20:36 crc kubenswrapper[4824]: I1211 02:20:36.712319 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 11 02:20:36 crc kubenswrapper[4824]: I1211 02:20:36.714616 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Dec 11 02:20:36 crc kubenswrapper[4824]: I1211 02:20:36.716705 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-zvqxv" Dec 11 02:20:36 crc kubenswrapper[4824]: I1211 02:20:36.716830 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Dec 11 02:20:36 crc kubenswrapper[4824]: I1211 02:20:36.717012 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Dec 11 02:20:36 crc kubenswrapper[4824]: I1211 02:20:36.736157 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-c7jtf"] Dec 11 02:20:36 crc kubenswrapper[4824]: I1211 02:20:36.745621 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-c7jtf"] Dec 11 02:20:36 crc kubenswrapper[4824]: I1211 02:20:36.805720 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/964729d8-30e0-4d0b-ae8e-6f9bfc2536d0-scripts\") pod \"ovn-northd-0\" (UID: \"964729d8-30e0-4d0b-ae8e-6f9bfc2536d0\") " pod="openstack/ovn-northd-0" Dec 11 02:20:36 crc kubenswrapper[4824]: I1211 02:20:36.805785 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/964729d8-30e0-4d0b-ae8e-6f9bfc2536d0-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"964729d8-30e0-4d0b-ae8e-6f9bfc2536d0\") " pod="openstack/ovn-northd-0" Dec 11 02:20:36 crc kubenswrapper[4824]: I1211 02:20:36.805806 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/964729d8-30e0-4d0b-ae8e-6f9bfc2536d0-config\") pod \"ovn-northd-0\" (UID: \"964729d8-30e0-4d0b-ae8e-6f9bfc2536d0\") " pod="openstack/ovn-northd-0" Dec 11 02:20:36 crc kubenswrapper[4824]: I1211 02:20:36.805840 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/964729d8-30e0-4d0b-ae8e-6f9bfc2536d0-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"964729d8-30e0-4d0b-ae8e-6f9bfc2536d0\") " pod="openstack/ovn-northd-0" Dec 11 02:20:36 crc kubenswrapper[4824]: I1211 02:20:36.805864 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/964729d8-30e0-4d0b-ae8e-6f9bfc2536d0-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"964729d8-30e0-4d0b-ae8e-6f9bfc2536d0\") " pod="openstack/ovn-northd-0" Dec 11 02:20:36 crc kubenswrapper[4824]: I1211 02:20:36.805892 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/964729d8-30e0-4d0b-ae8e-6f9bfc2536d0-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"964729d8-30e0-4d0b-ae8e-6f9bfc2536d0\") " pod="openstack/ovn-northd-0" Dec 11 02:20:36 crc kubenswrapper[4824]: I1211 02:20:36.805922 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h67hh\" (UniqueName: \"kubernetes.io/projected/964729d8-30e0-4d0b-ae8e-6f9bfc2536d0-kube-api-access-h67hh\") pod \"ovn-northd-0\" (UID: \"964729d8-30e0-4d0b-ae8e-6f9bfc2536d0\") " pod="openstack/ovn-northd-0" Dec 11 02:20:36 crc kubenswrapper[4824]: I1211 02:20:36.907192 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/964729d8-30e0-4d0b-ae8e-6f9bfc2536d0-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"964729d8-30e0-4d0b-ae8e-6f9bfc2536d0\") " pod="openstack/ovn-northd-0" Dec 11 02:20:36 crc kubenswrapper[4824]: I1211 02:20:36.907233 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/964729d8-30e0-4d0b-ae8e-6f9bfc2536d0-config\") pod \"ovn-northd-0\" (UID: \"964729d8-30e0-4d0b-ae8e-6f9bfc2536d0\") " pod="openstack/ovn-northd-0" Dec 11 02:20:36 crc kubenswrapper[4824]: I1211 02:20:36.907274 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/964729d8-30e0-4d0b-ae8e-6f9bfc2536d0-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"964729d8-30e0-4d0b-ae8e-6f9bfc2536d0\") " pod="openstack/ovn-northd-0" Dec 11 02:20:36 crc kubenswrapper[4824]: I1211 02:20:36.907297 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/964729d8-30e0-4d0b-ae8e-6f9bfc2536d0-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"964729d8-30e0-4d0b-ae8e-6f9bfc2536d0\") " pod="openstack/ovn-northd-0" Dec 11 02:20:36 crc kubenswrapper[4824]: I1211 02:20:36.907325 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/964729d8-30e0-4d0b-ae8e-6f9bfc2536d0-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"964729d8-30e0-4d0b-ae8e-6f9bfc2536d0\") " pod="openstack/ovn-northd-0" Dec 11 02:20:36 crc kubenswrapper[4824]: I1211 02:20:36.907356 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h67hh\" (UniqueName: \"kubernetes.io/projected/964729d8-30e0-4d0b-ae8e-6f9bfc2536d0-kube-api-access-h67hh\") pod \"ovn-northd-0\" (UID: \"964729d8-30e0-4d0b-ae8e-6f9bfc2536d0\") " pod="openstack/ovn-northd-0" Dec 11 02:20:36 crc kubenswrapper[4824]: I1211 02:20:36.907389 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/964729d8-30e0-4d0b-ae8e-6f9bfc2536d0-scripts\") pod \"ovn-northd-0\" (UID: \"964729d8-30e0-4d0b-ae8e-6f9bfc2536d0\") " pod="openstack/ovn-northd-0" Dec 11 02:20:36 crc kubenswrapper[4824]: I1211 02:20:36.908163 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/964729d8-30e0-4d0b-ae8e-6f9bfc2536d0-scripts\") pod \"ovn-northd-0\" (UID: \"964729d8-30e0-4d0b-ae8e-6f9bfc2536d0\") " pod="openstack/ovn-northd-0" Dec 11 02:20:36 crc kubenswrapper[4824]: I1211 02:20:36.909484 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/964729d8-30e0-4d0b-ae8e-6f9bfc2536d0-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"964729d8-30e0-4d0b-ae8e-6f9bfc2536d0\") " pod="openstack/ovn-northd-0" Dec 11 02:20:36 crc kubenswrapper[4824]: I1211 02:20:36.909562 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/964729d8-30e0-4d0b-ae8e-6f9bfc2536d0-config\") pod \"ovn-northd-0\" (UID: \"964729d8-30e0-4d0b-ae8e-6f9bfc2536d0\") " pod="openstack/ovn-northd-0" Dec 11 02:20:36 crc kubenswrapper[4824]: I1211 02:20:36.915361 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/964729d8-30e0-4d0b-ae8e-6f9bfc2536d0-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"964729d8-30e0-4d0b-ae8e-6f9bfc2536d0\") " pod="openstack/ovn-northd-0" Dec 11 02:20:36 crc kubenswrapper[4824]: I1211 02:20:36.915387 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/964729d8-30e0-4d0b-ae8e-6f9bfc2536d0-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"964729d8-30e0-4d0b-ae8e-6f9bfc2536d0\") " pod="openstack/ovn-northd-0" Dec 11 02:20:36 crc kubenswrapper[4824]: I1211 02:20:36.916350 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/964729d8-30e0-4d0b-ae8e-6f9bfc2536d0-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"964729d8-30e0-4d0b-ae8e-6f9bfc2536d0\") " pod="openstack/ovn-northd-0" Dec 11 02:20:36 crc kubenswrapper[4824]: I1211 02:20:36.926651 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h67hh\" (UniqueName: \"kubernetes.io/projected/964729d8-30e0-4d0b-ae8e-6f9bfc2536d0-kube-api-access-h67hh\") pod \"ovn-northd-0\" (UID: \"964729d8-30e0-4d0b-ae8e-6f9bfc2536d0\") " pod="openstack/ovn-northd-0" Dec 11 02:20:36 crc kubenswrapper[4824]: I1211 02:20:36.956544 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-5j6cw"] Dec 11 02:20:36 crc kubenswrapper[4824]: W1211 02:20:36.963707 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod342737ab_7fa3_4654_9493_e826dacbaa6f.slice/crio-f08525d1ef2fb0038ef806add621b4536a921ff2b14d914262cc47a0dc69773d WatchSource:0}: Error finding container f08525d1ef2fb0038ef806add621b4536a921ff2b14d914262cc47a0dc69773d: Status 404 returned error can't find the container with id f08525d1ef2fb0038ef806add621b4536a921ff2b14d914262cc47a0dc69773d Dec 11 02:20:37 crc kubenswrapper[4824]: I1211 02:20:37.037260 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 11 02:20:37 crc kubenswrapper[4824]: I1211 02:20:37.093903 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Dec 11 02:20:37 crc kubenswrapper[4824]: I1211 02:20:37.101839 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 11 02:20:37 crc kubenswrapper[4824]: I1211 02:20:37.104206 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Dec 11 02:20:37 crc kubenswrapper[4824]: I1211 02:20:37.104217 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Dec 11 02:20:37 crc kubenswrapper[4824]: I1211 02:20:37.104298 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-d2r5q" Dec 11 02:20:37 crc kubenswrapper[4824]: I1211 02:20:37.104334 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Dec 11 02:20:37 crc kubenswrapper[4824]: I1211 02:20:37.126062 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 11 02:20:37 crc kubenswrapper[4824]: I1211 02:20:37.211418 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/7913c5de-48cb-4e9a-8b9b-ee976476bab2-lock\") pod \"swift-storage-0\" (UID: \"7913c5de-48cb-4e9a-8b9b-ee976476bab2\") " pod="openstack/swift-storage-0" Dec 11 02:20:37 crc kubenswrapper[4824]: I1211 02:20:37.211464 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/7913c5de-48cb-4e9a-8b9b-ee976476bab2-cache\") pod \"swift-storage-0\" (UID: \"7913c5de-48cb-4e9a-8b9b-ee976476bab2\") " pod="openstack/swift-storage-0" Dec 11 02:20:37 crc kubenswrapper[4824]: I1211 02:20:37.211547 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mgp77\" (UniqueName: \"kubernetes.io/projected/7913c5de-48cb-4e9a-8b9b-ee976476bab2-kube-api-access-mgp77\") pod \"swift-storage-0\" (UID: \"7913c5de-48cb-4e9a-8b9b-ee976476bab2\") " pod="openstack/swift-storage-0" Dec 11 02:20:37 crc kubenswrapper[4824]: I1211 02:20:37.211589 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7913c5de-48cb-4e9a-8b9b-ee976476bab2-etc-swift\") pod \"swift-storage-0\" (UID: \"7913c5de-48cb-4e9a-8b9b-ee976476bab2\") " pod="openstack/swift-storage-0" Dec 11 02:20:37 crc kubenswrapper[4824]: I1211 02:20:37.211688 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-0\" (UID: \"7913c5de-48cb-4e9a-8b9b-ee976476bab2\") " pod="openstack/swift-storage-0" Dec 11 02:20:37 crc kubenswrapper[4824]: I1211 02:20:37.252577 4824 patch_prober.go:28] interesting pod/machine-config-daemon-gx6xt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 02:20:37 crc kubenswrapper[4824]: I1211 02:20:37.252863 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 02:20:37 crc kubenswrapper[4824]: I1211 02:20:37.252909 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" Dec 11 02:20:37 crc kubenswrapper[4824]: I1211 02:20:37.253577 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d1afa0a444a8a2d00d6d09bdd1c078096ee2bfd1a3af1878373a7a899d9f6039"} pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 11 02:20:37 crc kubenswrapper[4824]: I1211 02:20:37.253634 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" containerName="machine-config-daemon" containerID="cri-o://d1afa0a444a8a2d00d6d09bdd1c078096ee2bfd1a3af1878373a7a899d9f6039" gracePeriod=600 Dec 11 02:20:37 crc kubenswrapper[4824]: I1211 02:20:37.312742 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/7913c5de-48cb-4e9a-8b9b-ee976476bab2-lock\") pod \"swift-storage-0\" (UID: \"7913c5de-48cb-4e9a-8b9b-ee976476bab2\") " pod="openstack/swift-storage-0" Dec 11 02:20:37 crc kubenswrapper[4824]: I1211 02:20:37.312789 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/7913c5de-48cb-4e9a-8b9b-ee976476bab2-cache\") pod \"swift-storage-0\" (UID: \"7913c5de-48cb-4e9a-8b9b-ee976476bab2\") " pod="openstack/swift-storage-0" Dec 11 02:20:37 crc kubenswrapper[4824]: I1211 02:20:37.312832 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mgp77\" (UniqueName: \"kubernetes.io/projected/7913c5de-48cb-4e9a-8b9b-ee976476bab2-kube-api-access-mgp77\") pod \"swift-storage-0\" (UID: \"7913c5de-48cb-4e9a-8b9b-ee976476bab2\") " pod="openstack/swift-storage-0" Dec 11 02:20:37 crc kubenswrapper[4824]: I1211 02:20:37.312872 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7913c5de-48cb-4e9a-8b9b-ee976476bab2-etc-swift\") pod \"swift-storage-0\" (UID: \"7913c5de-48cb-4e9a-8b9b-ee976476bab2\") " pod="openstack/swift-storage-0" Dec 11 02:20:37 crc kubenswrapper[4824]: I1211 02:20:37.312898 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-0\" (UID: \"7913c5de-48cb-4e9a-8b9b-ee976476bab2\") " pod="openstack/swift-storage-0" Dec 11 02:20:37 crc kubenswrapper[4824]: E1211 02:20:37.313069 4824 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 11 02:20:37 crc kubenswrapper[4824]: E1211 02:20:37.313097 4824 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 11 02:20:37 crc kubenswrapper[4824]: E1211 02:20:37.313177 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/7913c5de-48cb-4e9a-8b9b-ee976476bab2-etc-swift podName:7913c5de-48cb-4e9a-8b9b-ee976476bab2 nodeName:}" failed. No retries permitted until 2025-12-11 02:20:37.813157963 +0000 UTC m=+1179.502195362 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/7913c5de-48cb-4e9a-8b9b-ee976476bab2-etc-swift") pod "swift-storage-0" (UID: "7913c5de-48cb-4e9a-8b9b-ee976476bab2") : configmap "swift-ring-files" not found Dec 11 02:20:37 crc kubenswrapper[4824]: I1211 02:20:37.313409 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/7913c5de-48cb-4e9a-8b9b-ee976476bab2-cache\") pod \"swift-storage-0\" (UID: \"7913c5de-48cb-4e9a-8b9b-ee976476bab2\") " pod="openstack/swift-storage-0" Dec 11 02:20:37 crc kubenswrapper[4824]: I1211 02:20:37.313422 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/7913c5de-48cb-4e9a-8b9b-ee976476bab2-lock\") pod \"swift-storage-0\" (UID: \"7913c5de-48cb-4e9a-8b9b-ee976476bab2\") " pod="openstack/swift-storage-0" Dec 11 02:20:37 crc kubenswrapper[4824]: I1211 02:20:37.313458 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-0\" (UID: \"7913c5de-48cb-4e9a-8b9b-ee976476bab2\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/swift-storage-0" Dec 11 02:20:37 crc kubenswrapper[4824]: I1211 02:20:37.333322 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-0\" (UID: \"7913c5de-48cb-4e9a-8b9b-ee976476bab2\") " pod="openstack/swift-storage-0" Dec 11 02:20:37 crc kubenswrapper[4824]: I1211 02:20:37.337940 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mgp77\" (UniqueName: \"kubernetes.io/projected/7913c5de-48cb-4e9a-8b9b-ee976476bab2-kube-api-access-mgp77\") pod \"swift-storage-0\" (UID: \"7913c5de-48cb-4e9a-8b9b-ee976476bab2\") " pod="openstack/swift-storage-0" Dec 11 02:20:37 crc kubenswrapper[4824]: I1211 02:20:37.413324 4824 generic.go:334] "Generic (PLEG): container finished" podID="342737ab-7fa3-4654-9493-e826dacbaa6f" containerID="6e401da432977c68f6a09e8a58b770b58dccae00320a2be0b3ea4389be82ecb4" exitCode=0 Dec 11 02:20:37 crc kubenswrapper[4824]: I1211 02:20:37.413429 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-5j6cw" event={"ID":"342737ab-7fa3-4654-9493-e826dacbaa6f","Type":"ContainerDied","Data":"6e401da432977c68f6a09e8a58b770b58dccae00320a2be0b3ea4389be82ecb4"} Dec 11 02:20:37 crc kubenswrapper[4824]: I1211 02:20:37.413471 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-5j6cw" event={"ID":"342737ab-7fa3-4654-9493-e826dacbaa6f","Type":"ContainerStarted","Data":"f08525d1ef2fb0038ef806add621b4536a921ff2b14d914262cc47a0dc69773d"} Dec 11 02:20:37 crc kubenswrapper[4824]: I1211 02:20:37.421653 4824 generic.go:334] "Generic (PLEG): container finished" podID="44b156e4-64a4-4d45-aa5e-9b10a862faed" containerID="d1afa0a444a8a2d00d6d09bdd1c078096ee2bfd1a3af1878373a7a899d9f6039" exitCode=0 Dec 11 02:20:37 crc kubenswrapper[4824]: I1211 02:20:37.422203 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" event={"ID":"44b156e4-64a4-4d45-aa5e-9b10a862faed","Type":"ContainerDied","Data":"d1afa0a444a8a2d00d6d09bdd1c078096ee2bfd1a3af1878373a7a899d9f6039"} Dec 11 02:20:37 crc kubenswrapper[4824]: I1211 02:20:37.422275 4824 scope.go:117] "RemoveContainer" containerID="65dbaed5184dd7e582c2d25d2c2754654e16b28021f7e2281401ebdcef00a13e" Dec 11 02:20:37 crc kubenswrapper[4824]: I1211 02:20:37.531950 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 11 02:20:37 crc kubenswrapper[4824]: I1211 02:20:37.820910 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7913c5de-48cb-4e9a-8b9b-ee976476bab2-etc-swift\") pod \"swift-storage-0\" (UID: \"7913c5de-48cb-4e9a-8b9b-ee976476bab2\") " pod="openstack/swift-storage-0" Dec 11 02:20:37 crc kubenswrapper[4824]: E1211 02:20:37.821470 4824 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 11 02:20:37 crc kubenswrapper[4824]: E1211 02:20:37.821515 4824 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 11 02:20:37 crc kubenswrapper[4824]: E1211 02:20:37.821598 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/7913c5de-48cb-4e9a-8b9b-ee976476bab2-etc-swift podName:7913c5de-48cb-4e9a-8b9b-ee976476bab2 nodeName:}" failed. No retries permitted until 2025-12-11 02:20:38.821572066 +0000 UTC m=+1180.510609485 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/7913c5de-48cb-4e9a-8b9b-ee976476bab2-etc-swift") pod "swift-storage-0" (UID: "7913c5de-48cb-4e9a-8b9b-ee976476bab2") : configmap "swift-ring-files" not found Dec 11 02:20:38 crc kubenswrapper[4824]: I1211 02:20:38.451217 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-5j6cw" event={"ID":"342737ab-7fa3-4654-9493-e826dacbaa6f","Type":"ContainerStarted","Data":"60884c2e4f625ce1fc400158671d70c7dab41b7db11c1f8de1113467ae4206e9"} Dec 11 02:20:38 crc kubenswrapper[4824]: I1211 02:20:38.455573 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-5j6cw" Dec 11 02:20:38 crc kubenswrapper[4824]: I1211 02:20:38.457531 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" event={"ID":"44b156e4-64a4-4d45-aa5e-9b10a862faed","Type":"ContainerStarted","Data":"d40ff17e029de5775ac2507f904eb2259da25d15d3aa3e3a0e4c811c4717820d"} Dec 11 02:20:38 crc kubenswrapper[4824]: I1211 02:20:38.462379 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"964729d8-30e0-4d0b-ae8e-6f9bfc2536d0","Type":"ContainerStarted","Data":"a65c1f618e9c99690c0b4821993fd39a0f85a01c55007834ceded4d1d3058045"} Dec 11 02:20:38 crc kubenswrapper[4824]: I1211 02:20:38.477067 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-698758b865-5j6cw" podStartSLOduration=3.477044452 podStartE2EDuration="3.477044452s" podCreationTimestamp="2025-12-11 02:20:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:20:38.474092831 +0000 UTC m=+1180.163130220" watchObservedRunningTime="2025-12-11 02:20:38.477044452 +0000 UTC m=+1180.166081851" Dec 11 02:20:38 crc kubenswrapper[4824]: I1211 02:20:38.641763 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35d304da-8076-40eb-8314-f4ff5405b6f4" path="/var/lib/kubelet/pods/35d304da-8076-40eb-8314-f4ff5405b6f4/volumes" Dec 11 02:20:38 crc kubenswrapper[4824]: I1211 02:20:38.836610 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7913c5de-48cb-4e9a-8b9b-ee976476bab2-etc-swift\") pod \"swift-storage-0\" (UID: \"7913c5de-48cb-4e9a-8b9b-ee976476bab2\") " pod="openstack/swift-storage-0" Dec 11 02:20:38 crc kubenswrapper[4824]: E1211 02:20:38.836910 4824 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 11 02:20:38 crc kubenswrapper[4824]: E1211 02:20:38.836931 4824 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 11 02:20:38 crc kubenswrapper[4824]: E1211 02:20:38.836979 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/7913c5de-48cb-4e9a-8b9b-ee976476bab2-etc-swift podName:7913c5de-48cb-4e9a-8b9b-ee976476bab2 nodeName:}" failed. No retries permitted until 2025-12-11 02:20:40.836963094 +0000 UTC m=+1182.526000483 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/7913c5de-48cb-4e9a-8b9b-ee976476bab2-etc-swift") pod "swift-storage-0" (UID: "7913c5de-48cb-4e9a-8b9b-ee976476bab2") : configmap "swift-ring-files" not found Dec 11 02:20:39 crc kubenswrapper[4824]: I1211 02:20:39.472821 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"964729d8-30e0-4d0b-ae8e-6f9bfc2536d0","Type":"ContainerStarted","Data":"b3a125b64eadbf6b9fa48ab146f50124f44bb49a927fb08b1f63ddfb2c05a30a"} Dec 11 02:20:40 crc kubenswrapper[4824]: I1211 02:20:40.485952 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"964729d8-30e0-4d0b-ae8e-6f9bfc2536d0","Type":"ContainerStarted","Data":"c3a7668253ba40632f10440b10806ea6efccf9cede51732acf45deaeb0ff17c3"} Dec 11 02:20:40 crc kubenswrapper[4824]: I1211 02:20:40.521891 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=3.016380325 podStartE2EDuration="4.521860072s" podCreationTimestamp="2025-12-11 02:20:36 +0000 UTC" firstStartedPulling="2025-12-11 02:20:37.540356704 +0000 UTC m=+1179.229394083" lastFinishedPulling="2025-12-11 02:20:39.045836451 +0000 UTC m=+1180.734873830" observedRunningTime="2025-12-11 02:20:40.514874811 +0000 UTC m=+1182.203912230" watchObservedRunningTime="2025-12-11 02:20:40.521860072 +0000 UTC m=+1182.210897481" Dec 11 02:20:40 crc kubenswrapper[4824]: I1211 02:20:40.870816 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7913c5de-48cb-4e9a-8b9b-ee976476bab2-etc-swift\") pod \"swift-storage-0\" (UID: \"7913c5de-48cb-4e9a-8b9b-ee976476bab2\") " pod="openstack/swift-storage-0" Dec 11 02:20:40 crc kubenswrapper[4824]: E1211 02:20:40.871064 4824 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 11 02:20:40 crc kubenswrapper[4824]: E1211 02:20:40.871459 4824 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 11 02:20:40 crc kubenswrapper[4824]: E1211 02:20:40.871574 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/7913c5de-48cb-4e9a-8b9b-ee976476bab2-etc-swift podName:7913c5de-48cb-4e9a-8b9b-ee976476bab2 nodeName:}" failed. No retries permitted until 2025-12-11 02:20:44.871538422 +0000 UTC m=+1186.560575851 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/7913c5de-48cb-4e9a-8b9b-ee976476bab2-etc-swift") pod "swift-storage-0" (UID: "7913c5de-48cb-4e9a-8b9b-ee976476bab2") : configmap "swift-ring-files" not found Dec 11 02:20:41 crc kubenswrapper[4824]: I1211 02:20:41.069544 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-zbgp5"] Dec 11 02:20:41 crc kubenswrapper[4824]: I1211 02:20:41.070509 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-zbgp5" Dec 11 02:20:41 crc kubenswrapper[4824]: I1211 02:20:41.072758 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Dec 11 02:20:41 crc kubenswrapper[4824]: I1211 02:20:41.073104 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 11 02:20:41 crc kubenswrapper[4824]: I1211 02:20:41.074752 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5719adff-b313-426e-afd1-69986bdc81bc-etc-swift\") pod \"swift-ring-rebalance-zbgp5\" (UID: \"5719adff-b313-426e-afd1-69986bdc81bc\") " pod="openstack/swift-ring-rebalance-zbgp5" Dec 11 02:20:41 crc kubenswrapper[4824]: I1211 02:20:41.074838 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5719adff-b313-426e-afd1-69986bdc81bc-scripts\") pod \"swift-ring-rebalance-zbgp5\" (UID: \"5719adff-b313-426e-afd1-69986bdc81bc\") " pod="openstack/swift-ring-rebalance-zbgp5" Dec 11 02:20:41 crc kubenswrapper[4824]: I1211 02:20:41.074923 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5719adff-b313-426e-afd1-69986bdc81bc-ring-data-devices\") pod \"swift-ring-rebalance-zbgp5\" (UID: \"5719adff-b313-426e-afd1-69986bdc81bc\") " pod="openstack/swift-ring-rebalance-zbgp5" Dec 11 02:20:41 crc kubenswrapper[4824]: I1211 02:20:41.074959 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5719adff-b313-426e-afd1-69986bdc81bc-swiftconf\") pod \"swift-ring-rebalance-zbgp5\" (UID: \"5719adff-b313-426e-afd1-69986bdc81bc\") " pod="openstack/swift-ring-rebalance-zbgp5" Dec 11 02:20:41 crc kubenswrapper[4824]: I1211 02:20:41.075003 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5719adff-b313-426e-afd1-69986bdc81bc-combined-ca-bundle\") pod \"swift-ring-rebalance-zbgp5\" (UID: \"5719adff-b313-426e-afd1-69986bdc81bc\") " pod="openstack/swift-ring-rebalance-zbgp5" Dec 11 02:20:41 crc kubenswrapper[4824]: I1211 02:20:41.075038 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z55vb\" (UniqueName: \"kubernetes.io/projected/5719adff-b313-426e-afd1-69986bdc81bc-kube-api-access-z55vb\") pod \"swift-ring-rebalance-zbgp5\" (UID: \"5719adff-b313-426e-afd1-69986bdc81bc\") " pod="openstack/swift-ring-rebalance-zbgp5" Dec 11 02:20:41 crc kubenswrapper[4824]: I1211 02:20:41.075099 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5719adff-b313-426e-afd1-69986bdc81bc-dispersionconf\") pod \"swift-ring-rebalance-zbgp5\" (UID: \"5719adff-b313-426e-afd1-69986bdc81bc\") " pod="openstack/swift-ring-rebalance-zbgp5" Dec 11 02:20:41 crc kubenswrapper[4824]: I1211 02:20:41.075800 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Dec 11 02:20:41 crc kubenswrapper[4824]: I1211 02:20:41.095914 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-zbgp5"] Dec 11 02:20:41 crc kubenswrapper[4824]: I1211 02:20:41.176660 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5719adff-b313-426e-afd1-69986bdc81bc-scripts\") pod \"swift-ring-rebalance-zbgp5\" (UID: \"5719adff-b313-426e-afd1-69986bdc81bc\") " pod="openstack/swift-ring-rebalance-zbgp5" Dec 11 02:20:41 crc kubenswrapper[4824]: I1211 02:20:41.176739 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5719adff-b313-426e-afd1-69986bdc81bc-ring-data-devices\") pod \"swift-ring-rebalance-zbgp5\" (UID: \"5719adff-b313-426e-afd1-69986bdc81bc\") " pod="openstack/swift-ring-rebalance-zbgp5" Dec 11 02:20:41 crc kubenswrapper[4824]: I1211 02:20:41.176788 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5719adff-b313-426e-afd1-69986bdc81bc-swiftconf\") pod \"swift-ring-rebalance-zbgp5\" (UID: \"5719adff-b313-426e-afd1-69986bdc81bc\") " pod="openstack/swift-ring-rebalance-zbgp5" Dec 11 02:20:41 crc kubenswrapper[4824]: I1211 02:20:41.176811 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5719adff-b313-426e-afd1-69986bdc81bc-combined-ca-bundle\") pod \"swift-ring-rebalance-zbgp5\" (UID: \"5719adff-b313-426e-afd1-69986bdc81bc\") " pod="openstack/swift-ring-rebalance-zbgp5" Dec 11 02:20:41 crc kubenswrapper[4824]: I1211 02:20:41.176835 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z55vb\" (UniqueName: \"kubernetes.io/projected/5719adff-b313-426e-afd1-69986bdc81bc-kube-api-access-z55vb\") pod \"swift-ring-rebalance-zbgp5\" (UID: \"5719adff-b313-426e-afd1-69986bdc81bc\") " pod="openstack/swift-ring-rebalance-zbgp5" Dec 11 02:20:41 crc kubenswrapper[4824]: I1211 02:20:41.176871 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5719adff-b313-426e-afd1-69986bdc81bc-dispersionconf\") pod \"swift-ring-rebalance-zbgp5\" (UID: \"5719adff-b313-426e-afd1-69986bdc81bc\") " pod="openstack/swift-ring-rebalance-zbgp5" Dec 11 02:20:41 crc kubenswrapper[4824]: I1211 02:20:41.176994 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5719adff-b313-426e-afd1-69986bdc81bc-etc-swift\") pod \"swift-ring-rebalance-zbgp5\" (UID: \"5719adff-b313-426e-afd1-69986bdc81bc\") " pod="openstack/swift-ring-rebalance-zbgp5" Dec 11 02:20:41 crc kubenswrapper[4824]: I1211 02:20:41.177732 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5719adff-b313-426e-afd1-69986bdc81bc-etc-swift\") pod \"swift-ring-rebalance-zbgp5\" (UID: \"5719adff-b313-426e-afd1-69986bdc81bc\") " pod="openstack/swift-ring-rebalance-zbgp5" Dec 11 02:20:41 crc kubenswrapper[4824]: I1211 02:20:41.178655 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5719adff-b313-426e-afd1-69986bdc81bc-scripts\") pod \"swift-ring-rebalance-zbgp5\" (UID: \"5719adff-b313-426e-afd1-69986bdc81bc\") " pod="openstack/swift-ring-rebalance-zbgp5" Dec 11 02:20:41 crc kubenswrapper[4824]: I1211 02:20:41.179302 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5719adff-b313-426e-afd1-69986bdc81bc-ring-data-devices\") pod \"swift-ring-rebalance-zbgp5\" (UID: \"5719adff-b313-426e-afd1-69986bdc81bc\") " pod="openstack/swift-ring-rebalance-zbgp5" Dec 11 02:20:41 crc kubenswrapper[4824]: I1211 02:20:41.183694 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5719adff-b313-426e-afd1-69986bdc81bc-dispersionconf\") pod \"swift-ring-rebalance-zbgp5\" (UID: \"5719adff-b313-426e-afd1-69986bdc81bc\") " pod="openstack/swift-ring-rebalance-zbgp5" Dec 11 02:20:41 crc kubenswrapper[4824]: I1211 02:20:41.184447 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5719adff-b313-426e-afd1-69986bdc81bc-swiftconf\") pod \"swift-ring-rebalance-zbgp5\" (UID: \"5719adff-b313-426e-afd1-69986bdc81bc\") " pod="openstack/swift-ring-rebalance-zbgp5" Dec 11 02:20:41 crc kubenswrapper[4824]: I1211 02:20:41.186891 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5719adff-b313-426e-afd1-69986bdc81bc-combined-ca-bundle\") pod \"swift-ring-rebalance-zbgp5\" (UID: \"5719adff-b313-426e-afd1-69986bdc81bc\") " pod="openstack/swift-ring-rebalance-zbgp5" Dec 11 02:20:41 crc kubenswrapper[4824]: I1211 02:20:41.210562 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z55vb\" (UniqueName: \"kubernetes.io/projected/5719adff-b313-426e-afd1-69986bdc81bc-kube-api-access-z55vb\") pod \"swift-ring-rebalance-zbgp5\" (UID: \"5719adff-b313-426e-afd1-69986bdc81bc\") " pod="openstack/swift-ring-rebalance-zbgp5" Dec 11 02:20:41 crc kubenswrapper[4824]: I1211 02:20:41.400956 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-zbgp5" Dec 11 02:20:41 crc kubenswrapper[4824]: I1211 02:20:41.496419 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Dec 11 02:20:41 crc kubenswrapper[4824]: I1211 02:20:41.530242 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-86db49b7ff-2cvmc" Dec 11 02:20:41 crc kubenswrapper[4824]: I1211 02:20:41.890416 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-zbgp5"] Dec 11 02:20:42 crc kubenswrapper[4824]: I1211 02:20:42.269074 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Dec 11 02:20:42 crc kubenswrapper[4824]: I1211 02:20:42.269529 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Dec 11 02:20:42 crc kubenswrapper[4824]: I1211 02:20:42.372282 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Dec 11 02:20:42 crc kubenswrapper[4824]: I1211 02:20:42.505747 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-zbgp5" event={"ID":"5719adff-b313-426e-afd1-69986bdc81bc","Type":"ContainerStarted","Data":"7afc8648378b67da8f8ede79e10c66a50be938dcd693e835dc589cdc4f561a27"} Dec 11 02:20:42 crc kubenswrapper[4824]: I1211 02:20:42.605351 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Dec 11 02:20:43 crc kubenswrapper[4824]: I1211 02:20:43.554584 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-7znqz"] Dec 11 02:20:43 crc kubenswrapper[4824]: I1211 02:20:43.555638 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-7znqz" Dec 11 02:20:43 crc kubenswrapper[4824]: I1211 02:20:43.561492 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-d963-account-create-update-k6d7n"] Dec 11 02:20:43 crc kubenswrapper[4824]: I1211 02:20:43.562653 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-d963-account-create-update-k6d7n" Dec 11 02:20:43 crc kubenswrapper[4824]: I1211 02:20:43.566450 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Dec 11 02:20:43 crc kubenswrapper[4824]: I1211 02:20:43.569566 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-d963-account-create-update-k6d7n"] Dec 11 02:20:43 crc kubenswrapper[4824]: I1211 02:20:43.577087 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Dec 11 02:20:43 crc kubenswrapper[4824]: I1211 02:20:43.578652 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-7znqz"] Dec 11 02:20:43 crc kubenswrapper[4824]: I1211 02:20:43.579515 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Dec 11 02:20:43 crc kubenswrapper[4824]: I1211 02:20:43.682508 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Dec 11 02:20:43 crc kubenswrapper[4824]: I1211 02:20:43.718392 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d7b35324-f201-4847-9e8e-062029d43cee-operator-scripts\") pod \"keystone-d963-account-create-update-k6d7n\" (UID: \"d7b35324-f201-4847-9e8e-062029d43cee\") " pod="openstack/keystone-d963-account-create-update-k6d7n" Dec 11 02:20:43 crc kubenswrapper[4824]: I1211 02:20:43.718454 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/77484236-1254-407f-950e-2a4861f10ead-operator-scripts\") pod \"keystone-db-create-7znqz\" (UID: \"77484236-1254-407f-950e-2a4861f10ead\") " pod="openstack/keystone-db-create-7znqz" Dec 11 02:20:43 crc kubenswrapper[4824]: I1211 02:20:43.718569 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vzhht\" (UniqueName: \"kubernetes.io/projected/77484236-1254-407f-950e-2a4861f10ead-kube-api-access-vzhht\") pod \"keystone-db-create-7znqz\" (UID: \"77484236-1254-407f-950e-2a4861f10ead\") " pod="openstack/keystone-db-create-7znqz" Dec 11 02:20:43 crc kubenswrapper[4824]: I1211 02:20:43.718614 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxtz2\" (UniqueName: \"kubernetes.io/projected/d7b35324-f201-4847-9e8e-062029d43cee-kube-api-access-vxtz2\") pod \"keystone-d963-account-create-update-k6d7n\" (UID: \"d7b35324-f201-4847-9e8e-062029d43cee\") " pod="openstack/keystone-d963-account-create-update-k6d7n" Dec 11 02:20:43 crc kubenswrapper[4824]: I1211 02:20:43.761634 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-cjh7n"] Dec 11 02:20:43 crc kubenswrapper[4824]: I1211 02:20:43.762858 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-cjh7n" Dec 11 02:20:43 crc kubenswrapper[4824]: I1211 02:20:43.771085 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-cjh7n"] Dec 11 02:20:43 crc kubenswrapper[4824]: I1211 02:20:43.820209 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vzhht\" (UniqueName: \"kubernetes.io/projected/77484236-1254-407f-950e-2a4861f10ead-kube-api-access-vzhht\") pod \"keystone-db-create-7znqz\" (UID: \"77484236-1254-407f-950e-2a4861f10ead\") " pod="openstack/keystone-db-create-7znqz" Dec 11 02:20:43 crc kubenswrapper[4824]: I1211 02:20:43.820275 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxtz2\" (UniqueName: \"kubernetes.io/projected/d7b35324-f201-4847-9e8e-062029d43cee-kube-api-access-vxtz2\") pod \"keystone-d963-account-create-update-k6d7n\" (UID: \"d7b35324-f201-4847-9e8e-062029d43cee\") " pod="openstack/keystone-d963-account-create-update-k6d7n" Dec 11 02:20:43 crc kubenswrapper[4824]: I1211 02:20:43.820365 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d7b35324-f201-4847-9e8e-062029d43cee-operator-scripts\") pod \"keystone-d963-account-create-update-k6d7n\" (UID: \"d7b35324-f201-4847-9e8e-062029d43cee\") " pod="openstack/keystone-d963-account-create-update-k6d7n" Dec 11 02:20:43 crc kubenswrapper[4824]: I1211 02:20:43.820402 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/77484236-1254-407f-950e-2a4861f10ead-operator-scripts\") pod \"keystone-db-create-7znqz\" (UID: \"77484236-1254-407f-950e-2a4861f10ead\") " pod="openstack/keystone-db-create-7znqz" Dec 11 02:20:43 crc kubenswrapper[4824]: I1211 02:20:43.821196 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/77484236-1254-407f-950e-2a4861f10ead-operator-scripts\") pod \"keystone-db-create-7znqz\" (UID: \"77484236-1254-407f-950e-2a4861f10ead\") " pod="openstack/keystone-db-create-7znqz" Dec 11 02:20:43 crc kubenswrapper[4824]: I1211 02:20:43.821957 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d7b35324-f201-4847-9e8e-062029d43cee-operator-scripts\") pod \"keystone-d963-account-create-update-k6d7n\" (UID: \"d7b35324-f201-4847-9e8e-062029d43cee\") " pod="openstack/keystone-d963-account-create-update-k6d7n" Dec 11 02:20:43 crc kubenswrapper[4824]: I1211 02:20:43.840077 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxtz2\" (UniqueName: \"kubernetes.io/projected/d7b35324-f201-4847-9e8e-062029d43cee-kube-api-access-vxtz2\") pod \"keystone-d963-account-create-update-k6d7n\" (UID: \"d7b35324-f201-4847-9e8e-062029d43cee\") " pod="openstack/keystone-d963-account-create-update-k6d7n" Dec 11 02:20:43 crc kubenswrapper[4824]: I1211 02:20:43.842528 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vzhht\" (UniqueName: \"kubernetes.io/projected/77484236-1254-407f-950e-2a4861f10ead-kube-api-access-vzhht\") pod \"keystone-db-create-7znqz\" (UID: \"77484236-1254-407f-950e-2a4861f10ead\") " pod="openstack/keystone-db-create-7znqz" Dec 11 02:20:43 crc kubenswrapper[4824]: I1211 02:20:43.872073 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-4d06-account-create-update-59g6f"] Dec 11 02:20:43 crc kubenswrapper[4824]: I1211 02:20:43.873187 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-4d06-account-create-update-59g6f" Dec 11 02:20:43 crc kubenswrapper[4824]: I1211 02:20:43.874970 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Dec 11 02:20:43 crc kubenswrapper[4824]: I1211 02:20:43.881918 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-4d06-account-create-update-59g6f"] Dec 11 02:20:43 crc kubenswrapper[4824]: I1211 02:20:43.919916 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-7znqz" Dec 11 02:20:43 crc kubenswrapper[4824]: I1211 02:20:43.921187 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d62e4855-c5c9-4d8b-ab6d-7695bbbe3f6b-operator-scripts\") pod \"placement-db-create-cjh7n\" (UID: \"d62e4855-c5c9-4d8b-ab6d-7695bbbe3f6b\") " pod="openstack/placement-db-create-cjh7n" Dec 11 02:20:43 crc kubenswrapper[4824]: I1211 02:20:43.921266 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8zq6t\" (UniqueName: \"kubernetes.io/projected/d62e4855-c5c9-4d8b-ab6d-7695bbbe3f6b-kube-api-access-8zq6t\") pod \"placement-db-create-cjh7n\" (UID: \"d62e4855-c5c9-4d8b-ab6d-7695bbbe3f6b\") " pod="openstack/placement-db-create-cjh7n" Dec 11 02:20:43 crc kubenswrapper[4824]: I1211 02:20:43.926855 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-d963-account-create-update-k6d7n" Dec 11 02:20:44 crc kubenswrapper[4824]: I1211 02:20:44.024839 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d62e4855-c5c9-4d8b-ab6d-7695bbbe3f6b-operator-scripts\") pod \"placement-db-create-cjh7n\" (UID: \"d62e4855-c5c9-4d8b-ab6d-7695bbbe3f6b\") " pod="openstack/placement-db-create-cjh7n" Dec 11 02:20:44 crc kubenswrapper[4824]: I1211 02:20:44.025293 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8zq6t\" (UniqueName: \"kubernetes.io/projected/d62e4855-c5c9-4d8b-ab6d-7695bbbe3f6b-kube-api-access-8zq6t\") pod \"placement-db-create-cjh7n\" (UID: \"d62e4855-c5c9-4d8b-ab6d-7695bbbe3f6b\") " pod="openstack/placement-db-create-cjh7n" Dec 11 02:20:44 crc kubenswrapper[4824]: I1211 02:20:44.025385 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nqtf7\" (UniqueName: \"kubernetes.io/projected/cf6d5db8-d6c0-4e6e-b8d1-22b94c561060-kube-api-access-nqtf7\") pod \"placement-4d06-account-create-update-59g6f\" (UID: \"cf6d5db8-d6c0-4e6e-b8d1-22b94c561060\") " pod="openstack/placement-4d06-account-create-update-59g6f" Dec 11 02:20:44 crc kubenswrapper[4824]: I1211 02:20:44.025407 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cf6d5db8-d6c0-4e6e-b8d1-22b94c561060-operator-scripts\") pod \"placement-4d06-account-create-update-59g6f\" (UID: \"cf6d5db8-d6c0-4e6e-b8d1-22b94c561060\") " pod="openstack/placement-4d06-account-create-update-59g6f" Dec 11 02:20:44 crc kubenswrapper[4824]: I1211 02:20:44.026425 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d62e4855-c5c9-4d8b-ab6d-7695bbbe3f6b-operator-scripts\") pod \"placement-db-create-cjh7n\" (UID: \"d62e4855-c5c9-4d8b-ab6d-7695bbbe3f6b\") " pod="openstack/placement-db-create-cjh7n" Dec 11 02:20:44 crc kubenswrapper[4824]: I1211 02:20:44.041321 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-zwsrx"] Dec 11 02:20:44 crc kubenswrapper[4824]: I1211 02:20:44.042986 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-zwsrx" Dec 11 02:20:44 crc kubenswrapper[4824]: I1211 02:20:44.044473 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8zq6t\" (UniqueName: \"kubernetes.io/projected/d62e4855-c5c9-4d8b-ab6d-7695bbbe3f6b-kube-api-access-8zq6t\") pod \"placement-db-create-cjh7n\" (UID: \"d62e4855-c5c9-4d8b-ab6d-7695bbbe3f6b\") " pod="openstack/placement-db-create-cjh7n" Dec 11 02:20:44 crc kubenswrapper[4824]: I1211 02:20:44.048436 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-zwsrx"] Dec 11 02:20:44 crc kubenswrapper[4824]: I1211 02:20:44.081975 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-cjh7n" Dec 11 02:20:44 crc kubenswrapper[4824]: I1211 02:20:44.127665 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2549107d-73ea-4e06-94a5-ff7a1f6b30b6-operator-scripts\") pod \"glance-db-create-zwsrx\" (UID: \"2549107d-73ea-4e06-94a5-ff7a1f6b30b6\") " pod="openstack/glance-db-create-zwsrx" Dec 11 02:20:44 crc kubenswrapper[4824]: I1211 02:20:44.127722 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nqtf7\" (UniqueName: \"kubernetes.io/projected/cf6d5db8-d6c0-4e6e-b8d1-22b94c561060-kube-api-access-nqtf7\") pod \"placement-4d06-account-create-update-59g6f\" (UID: \"cf6d5db8-d6c0-4e6e-b8d1-22b94c561060\") " pod="openstack/placement-4d06-account-create-update-59g6f" Dec 11 02:20:44 crc kubenswrapper[4824]: I1211 02:20:44.127758 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cf6d5db8-d6c0-4e6e-b8d1-22b94c561060-operator-scripts\") pod \"placement-4d06-account-create-update-59g6f\" (UID: \"cf6d5db8-d6c0-4e6e-b8d1-22b94c561060\") " pod="openstack/placement-4d06-account-create-update-59g6f" Dec 11 02:20:44 crc kubenswrapper[4824]: I1211 02:20:44.127791 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kp58q\" (UniqueName: \"kubernetes.io/projected/2549107d-73ea-4e06-94a5-ff7a1f6b30b6-kube-api-access-kp58q\") pod \"glance-db-create-zwsrx\" (UID: \"2549107d-73ea-4e06-94a5-ff7a1f6b30b6\") " pod="openstack/glance-db-create-zwsrx" Dec 11 02:20:44 crc kubenswrapper[4824]: I1211 02:20:44.128594 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cf6d5db8-d6c0-4e6e-b8d1-22b94c561060-operator-scripts\") pod \"placement-4d06-account-create-update-59g6f\" (UID: \"cf6d5db8-d6c0-4e6e-b8d1-22b94c561060\") " pod="openstack/placement-4d06-account-create-update-59g6f" Dec 11 02:20:44 crc kubenswrapper[4824]: I1211 02:20:44.154958 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nqtf7\" (UniqueName: \"kubernetes.io/projected/cf6d5db8-d6c0-4e6e-b8d1-22b94c561060-kube-api-access-nqtf7\") pod \"placement-4d06-account-create-update-59g6f\" (UID: \"cf6d5db8-d6c0-4e6e-b8d1-22b94c561060\") " pod="openstack/placement-4d06-account-create-update-59g6f" Dec 11 02:20:44 crc kubenswrapper[4824]: I1211 02:20:44.174867 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-5da1-account-create-update-g5bs8"] Dec 11 02:20:44 crc kubenswrapper[4824]: I1211 02:20:44.176403 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-5da1-account-create-update-g5bs8" Dec 11 02:20:44 crc kubenswrapper[4824]: I1211 02:20:44.178570 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Dec 11 02:20:44 crc kubenswrapper[4824]: I1211 02:20:44.185195 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-5da1-account-create-update-g5bs8"] Dec 11 02:20:44 crc kubenswrapper[4824]: I1211 02:20:44.213885 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-4d06-account-create-update-59g6f" Dec 11 02:20:44 crc kubenswrapper[4824]: I1211 02:20:44.229453 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2549107d-73ea-4e06-94a5-ff7a1f6b30b6-operator-scripts\") pod \"glance-db-create-zwsrx\" (UID: \"2549107d-73ea-4e06-94a5-ff7a1f6b30b6\") " pod="openstack/glance-db-create-zwsrx" Dec 11 02:20:44 crc kubenswrapper[4824]: I1211 02:20:44.229544 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kp58q\" (UniqueName: \"kubernetes.io/projected/2549107d-73ea-4e06-94a5-ff7a1f6b30b6-kube-api-access-kp58q\") pod \"glance-db-create-zwsrx\" (UID: \"2549107d-73ea-4e06-94a5-ff7a1f6b30b6\") " pod="openstack/glance-db-create-zwsrx" Dec 11 02:20:44 crc kubenswrapper[4824]: I1211 02:20:44.230358 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2549107d-73ea-4e06-94a5-ff7a1f6b30b6-operator-scripts\") pod \"glance-db-create-zwsrx\" (UID: \"2549107d-73ea-4e06-94a5-ff7a1f6b30b6\") " pod="openstack/glance-db-create-zwsrx" Dec 11 02:20:44 crc kubenswrapper[4824]: I1211 02:20:44.246746 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kp58q\" (UniqueName: \"kubernetes.io/projected/2549107d-73ea-4e06-94a5-ff7a1f6b30b6-kube-api-access-kp58q\") pod \"glance-db-create-zwsrx\" (UID: \"2549107d-73ea-4e06-94a5-ff7a1f6b30b6\") " pod="openstack/glance-db-create-zwsrx" Dec 11 02:20:44 crc kubenswrapper[4824]: I1211 02:20:44.331491 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5dhg\" (UniqueName: \"kubernetes.io/projected/7c259c18-68cf-4a86-9870-4f481ba4b12a-kube-api-access-n5dhg\") pod \"glance-5da1-account-create-update-g5bs8\" (UID: \"7c259c18-68cf-4a86-9870-4f481ba4b12a\") " pod="openstack/glance-5da1-account-create-update-g5bs8" Dec 11 02:20:44 crc kubenswrapper[4824]: I1211 02:20:44.331619 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7c259c18-68cf-4a86-9870-4f481ba4b12a-operator-scripts\") pod \"glance-5da1-account-create-update-g5bs8\" (UID: \"7c259c18-68cf-4a86-9870-4f481ba4b12a\") " pod="openstack/glance-5da1-account-create-update-g5bs8" Dec 11 02:20:44 crc kubenswrapper[4824]: I1211 02:20:44.406151 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-zwsrx" Dec 11 02:20:44 crc kubenswrapper[4824]: I1211 02:20:44.434050 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n5dhg\" (UniqueName: \"kubernetes.io/projected/7c259c18-68cf-4a86-9870-4f481ba4b12a-kube-api-access-n5dhg\") pod \"glance-5da1-account-create-update-g5bs8\" (UID: \"7c259c18-68cf-4a86-9870-4f481ba4b12a\") " pod="openstack/glance-5da1-account-create-update-g5bs8" Dec 11 02:20:44 crc kubenswrapper[4824]: I1211 02:20:44.434156 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7c259c18-68cf-4a86-9870-4f481ba4b12a-operator-scripts\") pod \"glance-5da1-account-create-update-g5bs8\" (UID: \"7c259c18-68cf-4a86-9870-4f481ba4b12a\") " pod="openstack/glance-5da1-account-create-update-g5bs8" Dec 11 02:20:44 crc kubenswrapper[4824]: I1211 02:20:44.434903 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7c259c18-68cf-4a86-9870-4f481ba4b12a-operator-scripts\") pod \"glance-5da1-account-create-update-g5bs8\" (UID: \"7c259c18-68cf-4a86-9870-4f481ba4b12a\") " pod="openstack/glance-5da1-account-create-update-g5bs8" Dec 11 02:20:44 crc kubenswrapper[4824]: I1211 02:20:44.463516 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n5dhg\" (UniqueName: \"kubernetes.io/projected/7c259c18-68cf-4a86-9870-4f481ba4b12a-kube-api-access-n5dhg\") pod \"glance-5da1-account-create-update-g5bs8\" (UID: \"7c259c18-68cf-4a86-9870-4f481ba4b12a\") " pod="openstack/glance-5da1-account-create-update-g5bs8" Dec 11 02:20:44 crc kubenswrapper[4824]: I1211 02:20:44.492837 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-5da1-account-create-update-g5bs8" Dec 11 02:20:44 crc kubenswrapper[4824]: I1211 02:20:44.648657 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Dec 11 02:20:44 crc kubenswrapper[4824]: I1211 02:20:44.949684 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7913c5de-48cb-4e9a-8b9b-ee976476bab2-etc-swift\") pod \"swift-storage-0\" (UID: \"7913c5de-48cb-4e9a-8b9b-ee976476bab2\") " pod="openstack/swift-storage-0" Dec 11 02:20:44 crc kubenswrapper[4824]: E1211 02:20:44.949903 4824 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 11 02:20:44 crc kubenswrapper[4824]: E1211 02:20:44.949936 4824 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 11 02:20:44 crc kubenswrapper[4824]: E1211 02:20:44.950020 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/7913c5de-48cb-4e9a-8b9b-ee976476bab2-etc-swift podName:7913c5de-48cb-4e9a-8b9b-ee976476bab2 nodeName:}" failed. No retries permitted until 2025-12-11 02:20:52.949982575 +0000 UTC m=+1194.639019964 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/7913c5de-48cb-4e9a-8b9b-ee976476bab2-etc-swift") pod "swift-storage-0" (UID: "7913c5de-48cb-4e9a-8b9b-ee976476bab2") : configmap "swift-ring-files" not found Dec 11 02:20:46 crc kubenswrapper[4824]: I1211 02:20:46.420785 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-698758b865-5j6cw" Dec 11 02:20:46 crc kubenswrapper[4824]: I1211 02:20:46.475353 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-2cvmc"] Dec 11 02:20:46 crc kubenswrapper[4824]: I1211 02:20:46.475795 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-86db49b7ff-2cvmc" podUID="fdb2591b-b56c-48f0-82a4-c5752e9107b0" containerName="dnsmasq-dns" containerID="cri-o://7e10ab73d64d3d111fbfbd8c0b836b6565fe58bf39f1c603e715c1d9ca2abab1" gracePeriod=10 Dec 11 02:20:46 crc kubenswrapper[4824]: I1211 02:20:46.529282 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-86db49b7ff-2cvmc" podUID="fdb2591b-b56c-48f0-82a4-c5752e9107b0" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.109:5353: connect: connection refused" Dec 11 02:20:46 crc kubenswrapper[4824]: I1211 02:20:46.540225 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-zbgp5" event={"ID":"5719adff-b313-426e-afd1-69986bdc81bc","Type":"ContainerStarted","Data":"a15261146f01f11324c688f9e1a3aabe5b76507fa3748a92cbe03c2cc4d4f211"} Dec 11 02:20:46 crc kubenswrapper[4824]: I1211 02:20:46.559557 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-zbgp5" podStartSLOduration=1.209370081 podStartE2EDuration="5.559538887s" podCreationTimestamp="2025-12-11 02:20:41 +0000 UTC" firstStartedPulling="2025-12-11 02:20:41.904721127 +0000 UTC m=+1183.593758506" lastFinishedPulling="2025-12-11 02:20:46.254889923 +0000 UTC m=+1187.943927312" observedRunningTime="2025-12-11 02:20:46.556764621 +0000 UTC m=+1188.245802000" watchObservedRunningTime="2025-12-11 02:20:46.559538887 +0000 UTC m=+1188.248576256" Dec 11 02:20:46 crc kubenswrapper[4824]: I1211 02:20:46.708562 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-4d06-account-create-update-59g6f"] Dec 11 02:20:46 crc kubenswrapper[4824]: I1211 02:20:46.738348 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-zwsrx"] Dec 11 02:20:46 crc kubenswrapper[4824]: W1211 02:20:46.746265 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2549107d_73ea_4e06_94a5_ff7a1f6b30b6.slice/crio-88da64e8df0332f2fccd308293891a89c1a171139cd682b1e805929274afdbe6 WatchSource:0}: Error finding container 88da64e8df0332f2fccd308293891a89c1a171139cd682b1e805929274afdbe6: Status 404 returned error can't find the container with id 88da64e8df0332f2fccd308293891a89c1a171139cd682b1e805929274afdbe6 Dec 11 02:20:46 crc kubenswrapper[4824]: I1211 02:20:46.991899 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-2cvmc" Dec 11 02:20:47 crc kubenswrapper[4824]: I1211 02:20:47.086235 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-cjh7n"] Dec 11 02:20:47 crc kubenswrapper[4824]: I1211 02:20:47.095410 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-7znqz"] Dec 11 02:20:47 crc kubenswrapper[4824]: I1211 02:20:47.099007 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fdb2591b-b56c-48f0-82a4-c5752e9107b0-dns-svc\") pod \"fdb2591b-b56c-48f0-82a4-c5752e9107b0\" (UID: \"fdb2591b-b56c-48f0-82a4-c5752e9107b0\") " Dec 11 02:20:47 crc kubenswrapper[4824]: I1211 02:20:47.099152 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fdb2591b-b56c-48f0-82a4-c5752e9107b0-config\") pod \"fdb2591b-b56c-48f0-82a4-c5752e9107b0\" (UID: \"fdb2591b-b56c-48f0-82a4-c5752e9107b0\") " Dec 11 02:20:47 crc kubenswrapper[4824]: I1211 02:20:47.099256 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fdb2591b-b56c-48f0-82a4-c5752e9107b0-ovsdbserver-sb\") pod \"fdb2591b-b56c-48f0-82a4-c5752e9107b0\" (UID: \"fdb2591b-b56c-48f0-82a4-c5752e9107b0\") " Dec 11 02:20:47 crc kubenswrapper[4824]: I1211 02:20:47.099294 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fdb2591b-b56c-48f0-82a4-c5752e9107b0-ovsdbserver-nb\") pod \"fdb2591b-b56c-48f0-82a4-c5752e9107b0\" (UID: \"fdb2591b-b56c-48f0-82a4-c5752e9107b0\") " Dec 11 02:20:47 crc kubenswrapper[4824]: I1211 02:20:47.099321 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l8jjl\" (UniqueName: \"kubernetes.io/projected/fdb2591b-b56c-48f0-82a4-c5752e9107b0-kube-api-access-l8jjl\") pod \"fdb2591b-b56c-48f0-82a4-c5752e9107b0\" (UID: \"fdb2591b-b56c-48f0-82a4-c5752e9107b0\") " Dec 11 02:20:47 crc kubenswrapper[4824]: I1211 02:20:47.106447 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-d963-account-create-update-k6d7n"] Dec 11 02:20:47 crc kubenswrapper[4824]: I1211 02:20:47.106682 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fdb2591b-b56c-48f0-82a4-c5752e9107b0-kube-api-access-l8jjl" (OuterVolumeSpecName: "kube-api-access-l8jjl") pod "fdb2591b-b56c-48f0-82a4-c5752e9107b0" (UID: "fdb2591b-b56c-48f0-82a4-c5752e9107b0"). InnerVolumeSpecName "kube-api-access-l8jjl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:20:47 crc kubenswrapper[4824]: I1211 02:20:47.114022 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-5da1-account-create-update-g5bs8"] Dec 11 02:20:47 crc kubenswrapper[4824]: I1211 02:20:47.143053 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fdb2591b-b56c-48f0-82a4-c5752e9107b0-config" (OuterVolumeSpecName: "config") pod "fdb2591b-b56c-48f0-82a4-c5752e9107b0" (UID: "fdb2591b-b56c-48f0-82a4-c5752e9107b0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:20:47 crc kubenswrapper[4824]: I1211 02:20:47.155850 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fdb2591b-b56c-48f0-82a4-c5752e9107b0-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "fdb2591b-b56c-48f0-82a4-c5752e9107b0" (UID: "fdb2591b-b56c-48f0-82a4-c5752e9107b0"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:20:47 crc kubenswrapper[4824]: I1211 02:20:47.157755 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fdb2591b-b56c-48f0-82a4-c5752e9107b0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "fdb2591b-b56c-48f0-82a4-c5752e9107b0" (UID: "fdb2591b-b56c-48f0-82a4-c5752e9107b0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:20:47 crc kubenswrapper[4824]: I1211 02:20:47.177033 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fdb2591b-b56c-48f0-82a4-c5752e9107b0-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "fdb2591b-b56c-48f0-82a4-c5752e9107b0" (UID: "fdb2591b-b56c-48f0-82a4-c5752e9107b0"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:20:47 crc kubenswrapper[4824]: I1211 02:20:47.201899 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fdb2591b-b56c-48f0-82a4-c5752e9107b0-config\") on node \"crc\" DevicePath \"\"" Dec 11 02:20:47 crc kubenswrapper[4824]: I1211 02:20:47.201956 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fdb2591b-b56c-48f0-82a4-c5752e9107b0-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 11 02:20:47 crc kubenswrapper[4824]: I1211 02:20:47.201967 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fdb2591b-b56c-48f0-82a4-c5752e9107b0-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 11 02:20:47 crc kubenswrapper[4824]: I1211 02:20:47.201976 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l8jjl\" (UniqueName: \"kubernetes.io/projected/fdb2591b-b56c-48f0-82a4-c5752e9107b0-kube-api-access-l8jjl\") on node \"crc\" DevicePath \"\"" Dec 11 02:20:47 crc kubenswrapper[4824]: I1211 02:20:47.201986 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fdb2591b-b56c-48f0-82a4-c5752e9107b0-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 11 02:20:47 crc kubenswrapper[4824]: I1211 02:20:47.555406 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-7znqz" event={"ID":"77484236-1254-407f-950e-2a4861f10ead","Type":"ContainerStarted","Data":"e8f49b6cf914032bf3ef3528ec6072ebe413fa5371ea0fffa28848847f52611c"} Dec 11 02:20:47 crc kubenswrapper[4824]: I1211 02:20:47.555714 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-7znqz" event={"ID":"77484236-1254-407f-950e-2a4861f10ead","Type":"ContainerStarted","Data":"1fef7fe46aea33effab1ac2d9426526c068c91e8e13e3d225bd3be7168a65afd"} Dec 11 02:20:47 crc kubenswrapper[4824]: I1211 02:20:47.557192 4824 generic.go:334] "Generic (PLEG): container finished" podID="2549107d-73ea-4e06-94a5-ff7a1f6b30b6" containerID="0943d6e1ee6b3e01f1a85b58a6a6a5d5fda8468e382b9355445cb109fc0a87a1" exitCode=0 Dec 11 02:20:47 crc kubenswrapper[4824]: I1211 02:20:47.557265 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-zwsrx" event={"ID":"2549107d-73ea-4e06-94a5-ff7a1f6b30b6","Type":"ContainerDied","Data":"0943d6e1ee6b3e01f1a85b58a6a6a5d5fda8468e382b9355445cb109fc0a87a1"} Dec 11 02:20:47 crc kubenswrapper[4824]: I1211 02:20:47.557292 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-zwsrx" event={"ID":"2549107d-73ea-4e06-94a5-ff7a1f6b30b6","Type":"ContainerStarted","Data":"88da64e8df0332f2fccd308293891a89c1a171139cd682b1e805929274afdbe6"} Dec 11 02:20:47 crc kubenswrapper[4824]: I1211 02:20:47.559584 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-d963-account-create-update-k6d7n" event={"ID":"d7b35324-f201-4847-9e8e-062029d43cee","Type":"ContainerStarted","Data":"5fc6bc4a209be71b47ee41b3aff839a4ccef0e1483672ccc600c113e0e00f29e"} Dec 11 02:20:47 crc kubenswrapper[4824]: I1211 02:20:47.559640 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-d963-account-create-update-k6d7n" event={"ID":"d7b35324-f201-4847-9e8e-062029d43cee","Type":"ContainerStarted","Data":"f6ae44cf16296e7a2f1b434f373e18637341fa36567ad2bf4e79f8bb057805e4"} Dec 11 02:20:47 crc kubenswrapper[4824]: I1211 02:20:47.564928 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-cjh7n" event={"ID":"d62e4855-c5c9-4d8b-ab6d-7695bbbe3f6b","Type":"ContainerStarted","Data":"d2188a3d6e6bd3e7da5caae01449214156d57fdd87701c6fdbe7d78f10cd6e90"} Dec 11 02:20:47 crc kubenswrapper[4824]: I1211 02:20:47.564969 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-cjh7n" event={"ID":"d62e4855-c5c9-4d8b-ab6d-7695bbbe3f6b","Type":"ContainerStarted","Data":"612af26a8e96129bb5cad9e53da0ac1dbdea79c1320f4754f1635a289a12561c"} Dec 11 02:20:47 crc kubenswrapper[4824]: I1211 02:20:47.569425 4824 generic.go:334] "Generic (PLEG): container finished" podID="cf6d5db8-d6c0-4e6e-b8d1-22b94c561060" containerID="6ad4a4876b1daca5369d39a7eece8a9decffb4d62224968b7a5f400151acfd56" exitCode=0 Dec 11 02:20:47 crc kubenswrapper[4824]: I1211 02:20:47.569478 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-4d06-account-create-update-59g6f" event={"ID":"cf6d5db8-d6c0-4e6e-b8d1-22b94c561060","Type":"ContainerDied","Data":"6ad4a4876b1daca5369d39a7eece8a9decffb4d62224968b7a5f400151acfd56"} Dec 11 02:20:47 crc kubenswrapper[4824]: I1211 02:20:47.569494 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-4d06-account-create-update-59g6f" event={"ID":"cf6d5db8-d6c0-4e6e-b8d1-22b94c561060","Type":"ContainerStarted","Data":"a468f2f01557f08c822b7aca8c3e1062a6c9c40720ce50520ff73643c9d46440"} Dec 11 02:20:47 crc kubenswrapper[4824]: I1211 02:20:47.571038 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-create-7znqz" podStartSLOduration=4.571018337 podStartE2EDuration="4.571018337s" podCreationTimestamp="2025-12-11 02:20:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:20:47.567878821 +0000 UTC m=+1189.256916200" watchObservedRunningTime="2025-12-11 02:20:47.571018337 +0000 UTC m=+1189.260055716" Dec 11 02:20:47 crc kubenswrapper[4824]: I1211 02:20:47.575310 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-5da1-account-create-update-g5bs8" event={"ID":"7c259c18-68cf-4a86-9870-4f481ba4b12a","Type":"ContainerStarted","Data":"56177de03d14f23a4b5e3f313a8924b97023cd805443017f164b97c499007198"} Dec 11 02:20:47 crc kubenswrapper[4824]: I1211 02:20:47.575351 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-5da1-account-create-update-g5bs8" event={"ID":"7c259c18-68cf-4a86-9870-4f481ba4b12a","Type":"ContainerStarted","Data":"5ceb29abde83466e6439348c9c7ff03490f6a91811262de2fd25320129d2eccd"} Dec 11 02:20:47 crc kubenswrapper[4824]: I1211 02:20:47.577591 4824 generic.go:334] "Generic (PLEG): container finished" podID="fdb2591b-b56c-48f0-82a4-c5752e9107b0" containerID="7e10ab73d64d3d111fbfbd8c0b836b6565fe58bf39f1c603e715c1d9ca2abab1" exitCode=0 Dec 11 02:20:47 crc kubenswrapper[4824]: I1211 02:20:47.578475 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-2cvmc" Dec 11 02:20:47 crc kubenswrapper[4824]: I1211 02:20:47.580198 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-2cvmc" event={"ID":"fdb2591b-b56c-48f0-82a4-c5752e9107b0","Type":"ContainerDied","Data":"7e10ab73d64d3d111fbfbd8c0b836b6565fe58bf39f1c603e715c1d9ca2abab1"} Dec 11 02:20:47 crc kubenswrapper[4824]: I1211 02:20:47.580247 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-2cvmc" event={"ID":"fdb2591b-b56c-48f0-82a4-c5752e9107b0","Type":"ContainerDied","Data":"e0b8f12355030c117508163acfd8d5ca4c6c33999e3771b455c8e77865647a1d"} Dec 11 02:20:47 crc kubenswrapper[4824]: I1211 02:20:47.580270 4824 scope.go:117] "RemoveContainer" containerID="7e10ab73d64d3d111fbfbd8c0b836b6565fe58bf39f1c603e715c1d9ca2abab1" Dec 11 02:20:47 crc kubenswrapper[4824]: I1211 02:20:47.610743 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-create-cjh7n" podStartSLOduration=4.610727207 podStartE2EDuration="4.610727207s" podCreationTimestamp="2025-12-11 02:20:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:20:47.608769953 +0000 UTC m=+1189.297807332" watchObservedRunningTime="2025-12-11 02:20:47.610727207 +0000 UTC m=+1189.299764586" Dec 11 02:20:47 crc kubenswrapper[4824]: I1211 02:20:47.617981 4824 scope.go:117] "RemoveContainer" containerID="d023802bf723a4402fc9ac5bb3d51be9aae723d79401b98aa6f5c63fb8ff67f1" Dec 11 02:20:47 crc kubenswrapper[4824]: I1211 02:20:47.630813 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-d963-account-create-update-k6d7n" podStartSLOduration=4.630799367 podStartE2EDuration="4.630799367s" podCreationTimestamp="2025-12-11 02:20:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:20:47.625277826 +0000 UTC m=+1189.314315205" watchObservedRunningTime="2025-12-11 02:20:47.630799367 +0000 UTC m=+1189.319836746" Dec 11 02:20:47 crc kubenswrapper[4824]: I1211 02:20:47.652251 4824 scope.go:117] "RemoveContainer" containerID="7e10ab73d64d3d111fbfbd8c0b836b6565fe58bf39f1c603e715c1d9ca2abab1" Dec 11 02:20:47 crc kubenswrapper[4824]: E1211 02:20:47.654311 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7e10ab73d64d3d111fbfbd8c0b836b6565fe58bf39f1c603e715c1d9ca2abab1\": container with ID starting with 7e10ab73d64d3d111fbfbd8c0b836b6565fe58bf39f1c603e715c1d9ca2abab1 not found: ID does not exist" containerID="7e10ab73d64d3d111fbfbd8c0b836b6565fe58bf39f1c603e715c1d9ca2abab1" Dec 11 02:20:47 crc kubenswrapper[4824]: I1211 02:20:47.654345 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e10ab73d64d3d111fbfbd8c0b836b6565fe58bf39f1c603e715c1d9ca2abab1"} err="failed to get container status \"7e10ab73d64d3d111fbfbd8c0b836b6565fe58bf39f1c603e715c1d9ca2abab1\": rpc error: code = NotFound desc = could not find container \"7e10ab73d64d3d111fbfbd8c0b836b6565fe58bf39f1c603e715c1d9ca2abab1\": container with ID starting with 7e10ab73d64d3d111fbfbd8c0b836b6565fe58bf39f1c603e715c1d9ca2abab1 not found: ID does not exist" Dec 11 02:20:47 crc kubenswrapper[4824]: I1211 02:20:47.654367 4824 scope.go:117] "RemoveContainer" containerID="d023802bf723a4402fc9ac5bb3d51be9aae723d79401b98aa6f5c63fb8ff67f1" Dec 11 02:20:47 crc kubenswrapper[4824]: E1211 02:20:47.655181 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d023802bf723a4402fc9ac5bb3d51be9aae723d79401b98aa6f5c63fb8ff67f1\": container with ID starting with d023802bf723a4402fc9ac5bb3d51be9aae723d79401b98aa6f5c63fb8ff67f1 not found: ID does not exist" containerID="d023802bf723a4402fc9ac5bb3d51be9aae723d79401b98aa6f5c63fb8ff67f1" Dec 11 02:20:47 crc kubenswrapper[4824]: I1211 02:20:47.655205 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d023802bf723a4402fc9ac5bb3d51be9aae723d79401b98aa6f5c63fb8ff67f1"} err="failed to get container status \"d023802bf723a4402fc9ac5bb3d51be9aae723d79401b98aa6f5c63fb8ff67f1\": rpc error: code = NotFound desc = could not find container \"d023802bf723a4402fc9ac5bb3d51be9aae723d79401b98aa6f5c63fb8ff67f1\": container with ID starting with d023802bf723a4402fc9ac5bb3d51be9aae723d79401b98aa6f5c63fb8ff67f1 not found: ID does not exist" Dec 11 02:20:47 crc kubenswrapper[4824]: I1211 02:20:47.656654 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-2cvmc"] Dec 11 02:20:47 crc kubenswrapper[4824]: I1211 02:20:47.668176 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-2cvmc"] Dec 11 02:20:47 crc kubenswrapper[4824]: I1211 02:20:47.687224 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-5da1-account-create-update-g5bs8" podStartSLOduration=3.687204974 podStartE2EDuration="3.687204974s" podCreationTimestamp="2025-12-11 02:20:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:20:47.684570132 +0000 UTC m=+1189.373607521" watchObservedRunningTime="2025-12-11 02:20:47.687204974 +0000 UTC m=+1189.376242363" Dec 11 02:20:48 crc kubenswrapper[4824]: I1211 02:20:48.590401 4824 generic.go:334] "Generic (PLEG): container finished" podID="77484236-1254-407f-950e-2a4861f10ead" containerID="e8f49b6cf914032bf3ef3528ec6072ebe413fa5371ea0fffa28848847f52611c" exitCode=0 Dec 11 02:20:48 crc kubenswrapper[4824]: I1211 02:20:48.590470 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-7znqz" event={"ID":"77484236-1254-407f-950e-2a4861f10ead","Type":"ContainerDied","Data":"e8f49b6cf914032bf3ef3528ec6072ebe413fa5371ea0fffa28848847f52611c"} Dec 11 02:20:48 crc kubenswrapper[4824]: I1211 02:20:48.592702 4824 generic.go:334] "Generic (PLEG): container finished" podID="d7b35324-f201-4847-9e8e-062029d43cee" containerID="5fc6bc4a209be71b47ee41b3aff839a4ccef0e1483672ccc600c113e0e00f29e" exitCode=0 Dec 11 02:20:48 crc kubenswrapper[4824]: I1211 02:20:48.592830 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-d963-account-create-update-k6d7n" event={"ID":"d7b35324-f201-4847-9e8e-062029d43cee","Type":"ContainerDied","Data":"5fc6bc4a209be71b47ee41b3aff839a4ccef0e1483672ccc600c113e0e00f29e"} Dec 11 02:20:48 crc kubenswrapper[4824]: I1211 02:20:48.595506 4824 generic.go:334] "Generic (PLEG): container finished" podID="d62e4855-c5c9-4d8b-ab6d-7695bbbe3f6b" containerID="d2188a3d6e6bd3e7da5caae01449214156d57fdd87701c6fdbe7d78f10cd6e90" exitCode=0 Dec 11 02:20:48 crc kubenswrapper[4824]: I1211 02:20:48.595571 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-cjh7n" event={"ID":"d62e4855-c5c9-4d8b-ab6d-7695bbbe3f6b","Type":"ContainerDied","Data":"d2188a3d6e6bd3e7da5caae01449214156d57fdd87701c6fdbe7d78f10cd6e90"} Dec 11 02:20:48 crc kubenswrapper[4824]: I1211 02:20:48.598702 4824 generic.go:334] "Generic (PLEG): container finished" podID="7c259c18-68cf-4a86-9870-4f481ba4b12a" containerID="56177de03d14f23a4b5e3f313a8924b97023cd805443017f164b97c499007198" exitCode=0 Dec 11 02:20:48 crc kubenswrapper[4824]: I1211 02:20:48.598768 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-5da1-account-create-update-g5bs8" event={"ID":"7c259c18-68cf-4a86-9870-4f481ba4b12a","Type":"ContainerDied","Data":"56177de03d14f23a4b5e3f313a8924b97023cd805443017f164b97c499007198"} Dec 11 02:20:48 crc kubenswrapper[4824]: I1211 02:20:48.658892 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fdb2591b-b56c-48f0-82a4-c5752e9107b0" path="/var/lib/kubelet/pods/fdb2591b-b56c-48f0-82a4-c5752e9107b0/volumes" Dec 11 02:20:49 crc kubenswrapper[4824]: I1211 02:20:49.081581 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-zwsrx" Dec 11 02:20:49 crc kubenswrapper[4824]: I1211 02:20:49.092009 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-4d06-account-create-update-59g6f" Dec 11 02:20:49 crc kubenswrapper[4824]: I1211 02:20:49.244556 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nqtf7\" (UniqueName: \"kubernetes.io/projected/cf6d5db8-d6c0-4e6e-b8d1-22b94c561060-kube-api-access-nqtf7\") pod \"cf6d5db8-d6c0-4e6e-b8d1-22b94c561060\" (UID: \"cf6d5db8-d6c0-4e6e-b8d1-22b94c561060\") " Dec 11 02:20:49 crc kubenswrapper[4824]: I1211 02:20:49.244635 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2549107d-73ea-4e06-94a5-ff7a1f6b30b6-operator-scripts\") pod \"2549107d-73ea-4e06-94a5-ff7a1f6b30b6\" (UID: \"2549107d-73ea-4e06-94a5-ff7a1f6b30b6\") " Dec 11 02:20:49 crc kubenswrapper[4824]: I1211 02:20:49.244656 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kp58q\" (UniqueName: \"kubernetes.io/projected/2549107d-73ea-4e06-94a5-ff7a1f6b30b6-kube-api-access-kp58q\") pod \"2549107d-73ea-4e06-94a5-ff7a1f6b30b6\" (UID: \"2549107d-73ea-4e06-94a5-ff7a1f6b30b6\") " Dec 11 02:20:49 crc kubenswrapper[4824]: I1211 02:20:49.244732 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cf6d5db8-d6c0-4e6e-b8d1-22b94c561060-operator-scripts\") pod \"cf6d5db8-d6c0-4e6e-b8d1-22b94c561060\" (UID: \"cf6d5db8-d6c0-4e6e-b8d1-22b94c561060\") " Dec 11 02:20:49 crc kubenswrapper[4824]: I1211 02:20:49.245280 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2549107d-73ea-4e06-94a5-ff7a1f6b30b6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2549107d-73ea-4e06-94a5-ff7a1f6b30b6" (UID: "2549107d-73ea-4e06-94a5-ff7a1f6b30b6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:20:49 crc kubenswrapper[4824]: I1211 02:20:49.245470 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cf6d5db8-d6c0-4e6e-b8d1-22b94c561060-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "cf6d5db8-d6c0-4e6e-b8d1-22b94c561060" (UID: "cf6d5db8-d6c0-4e6e-b8d1-22b94c561060"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:20:49 crc kubenswrapper[4824]: I1211 02:20:49.256649 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf6d5db8-d6c0-4e6e-b8d1-22b94c561060-kube-api-access-nqtf7" (OuterVolumeSpecName: "kube-api-access-nqtf7") pod "cf6d5db8-d6c0-4e6e-b8d1-22b94c561060" (UID: "cf6d5db8-d6c0-4e6e-b8d1-22b94c561060"). InnerVolumeSpecName "kube-api-access-nqtf7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:20:49 crc kubenswrapper[4824]: I1211 02:20:49.257892 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2549107d-73ea-4e06-94a5-ff7a1f6b30b6-kube-api-access-kp58q" (OuterVolumeSpecName: "kube-api-access-kp58q") pod "2549107d-73ea-4e06-94a5-ff7a1f6b30b6" (UID: "2549107d-73ea-4e06-94a5-ff7a1f6b30b6"). InnerVolumeSpecName "kube-api-access-kp58q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:20:49 crc kubenswrapper[4824]: I1211 02:20:49.346915 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cf6d5db8-d6c0-4e6e-b8d1-22b94c561060-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 02:20:49 crc kubenswrapper[4824]: I1211 02:20:49.346963 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nqtf7\" (UniqueName: \"kubernetes.io/projected/cf6d5db8-d6c0-4e6e-b8d1-22b94c561060-kube-api-access-nqtf7\") on node \"crc\" DevicePath \"\"" Dec 11 02:20:49 crc kubenswrapper[4824]: I1211 02:20:49.346978 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2549107d-73ea-4e06-94a5-ff7a1f6b30b6-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 02:20:49 crc kubenswrapper[4824]: I1211 02:20:49.346989 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kp58q\" (UniqueName: \"kubernetes.io/projected/2549107d-73ea-4e06-94a5-ff7a1f6b30b6-kube-api-access-kp58q\") on node \"crc\" DevicePath \"\"" Dec 11 02:20:49 crc kubenswrapper[4824]: I1211 02:20:49.612238 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-zwsrx" event={"ID":"2549107d-73ea-4e06-94a5-ff7a1f6b30b6","Type":"ContainerDied","Data":"88da64e8df0332f2fccd308293891a89c1a171139cd682b1e805929274afdbe6"} Dec 11 02:20:49 crc kubenswrapper[4824]: I1211 02:20:49.612732 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="88da64e8df0332f2fccd308293891a89c1a171139cd682b1e805929274afdbe6" Dec 11 02:20:49 crc kubenswrapper[4824]: I1211 02:20:49.612468 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-zwsrx" Dec 11 02:20:49 crc kubenswrapper[4824]: I1211 02:20:49.614401 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-4d06-account-create-update-59g6f" Dec 11 02:20:49 crc kubenswrapper[4824]: I1211 02:20:49.620208 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-4d06-account-create-update-59g6f" event={"ID":"cf6d5db8-d6c0-4e6e-b8d1-22b94c561060","Type":"ContainerDied","Data":"a468f2f01557f08c822b7aca8c3e1062a6c9c40720ce50520ff73643c9d46440"} Dec 11 02:20:49 crc kubenswrapper[4824]: I1211 02:20:49.620295 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a468f2f01557f08c822b7aca8c3e1062a6c9c40720ce50520ff73643c9d46440" Dec 11 02:20:50 crc kubenswrapper[4824]: I1211 02:20:50.117478 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-d963-account-create-update-k6d7n" Dec 11 02:20:50 crc kubenswrapper[4824]: I1211 02:20:50.263028 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d7b35324-f201-4847-9e8e-062029d43cee-operator-scripts\") pod \"d7b35324-f201-4847-9e8e-062029d43cee\" (UID: \"d7b35324-f201-4847-9e8e-062029d43cee\") " Dec 11 02:20:50 crc kubenswrapper[4824]: I1211 02:20:50.263142 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vxtz2\" (UniqueName: \"kubernetes.io/projected/d7b35324-f201-4847-9e8e-062029d43cee-kube-api-access-vxtz2\") pod \"d7b35324-f201-4847-9e8e-062029d43cee\" (UID: \"d7b35324-f201-4847-9e8e-062029d43cee\") " Dec 11 02:20:50 crc kubenswrapper[4824]: I1211 02:20:50.264068 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7b35324-f201-4847-9e8e-062029d43cee-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d7b35324-f201-4847-9e8e-062029d43cee" (UID: "d7b35324-f201-4847-9e8e-062029d43cee"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:20:50 crc kubenswrapper[4824]: I1211 02:20:50.269900 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7b35324-f201-4847-9e8e-062029d43cee-kube-api-access-vxtz2" (OuterVolumeSpecName: "kube-api-access-vxtz2") pod "d7b35324-f201-4847-9e8e-062029d43cee" (UID: "d7b35324-f201-4847-9e8e-062029d43cee"). InnerVolumeSpecName "kube-api-access-vxtz2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:20:50 crc kubenswrapper[4824]: I1211 02:20:50.277955 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-7znqz" Dec 11 02:20:50 crc kubenswrapper[4824]: I1211 02:20:50.282522 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-cjh7n" Dec 11 02:20:50 crc kubenswrapper[4824]: I1211 02:20:50.287421 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-5da1-account-create-update-g5bs8" Dec 11 02:20:50 crc kubenswrapper[4824]: I1211 02:20:50.364030 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n5dhg\" (UniqueName: \"kubernetes.io/projected/7c259c18-68cf-4a86-9870-4f481ba4b12a-kube-api-access-n5dhg\") pod \"7c259c18-68cf-4a86-9870-4f481ba4b12a\" (UID: \"7c259c18-68cf-4a86-9870-4f481ba4b12a\") " Dec 11 02:20:50 crc kubenswrapper[4824]: I1211 02:20:50.364079 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vzhht\" (UniqueName: \"kubernetes.io/projected/77484236-1254-407f-950e-2a4861f10ead-kube-api-access-vzhht\") pod \"77484236-1254-407f-950e-2a4861f10ead\" (UID: \"77484236-1254-407f-950e-2a4861f10ead\") " Dec 11 02:20:50 crc kubenswrapper[4824]: I1211 02:20:50.364192 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8zq6t\" (UniqueName: \"kubernetes.io/projected/d62e4855-c5c9-4d8b-ab6d-7695bbbe3f6b-kube-api-access-8zq6t\") pod \"d62e4855-c5c9-4d8b-ab6d-7695bbbe3f6b\" (UID: \"d62e4855-c5c9-4d8b-ab6d-7695bbbe3f6b\") " Dec 11 02:20:50 crc kubenswrapper[4824]: I1211 02:20:50.364220 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d62e4855-c5c9-4d8b-ab6d-7695bbbe3f6b-operator-scripts\") pod \"d62e4855-c5c9-4d8b-ab6d-7695bbbe3f6b\" (UID: \"d62e4855-c5c9-4d8b-ab6d-7695bbbe3f6b\") " Dec 11 02:20:50 crc kubenswrapper[4824]: I1211 02:20:50.364312 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/77484236-1254-407f-950e-2a4861f10ead-operator-scripts\") pod \"77484236-1254-407f-950e-2a4861f10ead\" (UID: \"77484236-1254-407f-950e-2a4861f10ead\") " Dec 11 02:20:50 crc kubenswrapper[4824]: I1211 02:20:50.364451 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7c259c18-68cf-4a86-9870-4f481ba4b12a-operator-scripts\") pod \"7c259c18-68cf-4a86-9870-4f481ba4b12a\" (UID: \"7c259c18-68cf-4a86-9870-4f481ba4b12a\") " Dec 11 02:20:50 crc kubenswrapper[4824]: I1211 02:20:50.364755 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d62e4855-c5c9-4d8b-ab6d-7695bbbe3f6b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d62e4855-c5c9-4d8b-ab6d-7695bbbe3f6b" (UID: "d62e4855-c5c9-4d8b-ab6d-7695bbbe3f6b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:20:50 crc kubenswrapper[4824]: I1211 02:20:50.364836 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d7b35324-f201-4847-9e8e-062029d43cee-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 02:20:50 crc kubenswrapper[4824]: I1211 02:20:50.364851 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d62e4855-c5c9-4d8b-ab6d-7695bbbe3f6b-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 02:20:50 crc kubenswrapper[4824]: I1211 02:20:50.364864 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vxtz2\" (UniqueName: \"kubernetes.io/projected/d7b35324-f201-4847-9e8e-062029d43cee-kube-api-access-vxtz2\") on node \"crc\" DevicePath \"\"" Dec 11 02:20:50 crc kubenswrapper[4824]: I1211 02:20:50.365145 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/77484236-1254-407f-950e-2a4861f10ead-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "77484236-1254-407f-950e-2a4861f10ead" (UID: "77484236-1254-407f-950e-2a4861f10ead"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:20:50 crc kubenswrapper[4824]: I1211 02:20:50.365177 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c259c18-68cf-4a86-9870-4f481ba4b12a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7c259c18-68cf-4a86-9870-4f481ba4b12a" (UID: "7c259c18-68cf-4a86-9870-4f481ba4b12a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:20:50 crc kubenswrapper[4824]: I1211 02:20:50.367893 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c259c18-68cf-4a86-9870-4f481ba4b12a-kube-api-access-n5dhg" (OuterVolumeSpecName: "kube-api-access-n5dhg") pod "7c259c18-68cf-4a86-9870-4f481ba4b12a" (UID: "7c259c18-68cf-4a86-9870-4f481ba4b12a"). InnerVolumeSpecName "kube-api-access-n5dhg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:20:50 crc kubenswrapper[4824]: I1211 02:20:50.369365 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d62e4855-c5c9-4d8b-ab6d-7695bbbe3f6b-kube-api-access-8zq6t" (OuterVolumeSpecName: "kube-api-access-8zq6t") pod "d62e4855-c5c9-4d8b-ab6d-7695bbbe3f6b" (UID: "d62e4855-c5c9-4d8b-ab6d-7695bbbe3f6b"). InnerVolumeSpecName "kube-api-access-8zq6t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:20:50 crc kubenswrapper[4824]: I1211 02:20:50.369607 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77484236-1254-407f-950e-2a4861f10ead-kube-api-access-vzhht" (OuterVolumeSpecName: "kube-api-access-vzhht") pod "77484236-1254-407f-950e-2a4861f10ead" (UID: "77484236-1254-407f-950e-2a4861f10ead"). InnerVolumeSpecName "kube-api-access-vzhht". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:20:50 crc kubenswrapper[4824]: I1211 02:20:50.467154 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7c259c18-68cf-4a86-9870-4f481ba4b12a-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 02:20:50 crc kubenswrapper[4824]: I1211 02:20:50.467211 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vzhht\" (UniqueName: \"kubernetes.io/projected/77484236-1254-407f-950e-2a4861f10ead-kube-api-access-vzhht\") on node \"crc\" DevicePath \"\"" Dec 11 02:20:50 crc kubenswrapper[4824]: I1211 02:20:50.467237 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n5dhg\" (UniqueName: \"kubernetes.io/projected/7c259c18-68cf-4a86-9870-4f481ba4b12a-kube-api-access-n5dhg\") on node \"crc\" DevicePath \"\"" Dec 11 02:20:50 crc kubenswrapper[4824]: I1211 02:20:50.467257 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8zq6t\" (UniqueName: \"kubernetes.io/projected/d62e4855-c5c9-4d8b-ab6d-7695bbbe3f6b-kube-api-access-8zq6t\") on node \"crc\" DevicePath \"\"" Dec 11 02:20:50 crc kubenswrapper[4824]: I1211 02:20:50.467275 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/77484236-1254-407f-950e-2a4861f10ead-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 02:20:50 crc kubenswrapper[4824]: I1211 02:20:50.629075 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-cjh7n" event={"ID":"d62e4855-c5c9-4d8b-ab6d-7695bbbe3f6b","Type":"ContainerDied","Data":"612af26a8e96129bb5cad9e53da0ac1dbdea79c1320f4754f1635a289a12561c"} Dec 11 02:20:50 crc kubenswrapper[4824]: I1211 02:20:50.629549 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="612af26a8e96129bb5cad9e53da0ac1dbdea79c1320f4754f1635a289a12561c" Dec 11 02:20:50 crc kubenswrapper[4824]: I1211 02:20:50.629085 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-cjh7n" Dec 11 02:20:50 crc kubenswrapper[4824]: I1211 02:20:50.631584 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-5da1-account-create-update-g5bs8" Dec 11 02:20:50 crc kubenswrapper[4824]: I1211 02:20:50.641652 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-7znqz" Dec 11 02:20:50 crc kubenswrapper[4824]: I1211 02:20:50.647887 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-d963-account-create-update-k6d7n" Dec 11 02:20:50 crc kubenswrapper[4824]: I1211 02:20:50.652175 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-5da1-account-create-update-g5bs8" event={"ID":"7c259c18-68cf-4a86-9870-4f481ba4b12a","Type":"ContainerDied","Data":"5ceb29abde83466e6439348c9c7ff03490f6a91811262de2fd25320129d2eccd"} Dec 11 02:20:50 crc kubenswrapper[4824]: I1211 02:20:50.652230 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5ceb29abde83466e6439348c9c7ff03490f6a91811262de2fd25320129d2eccd" Dec 11 02:20:50 crc kubenswrapper[4824]: I1211 02:20:50.652252 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-7znqz" event={"ID":"77484236-1254-407f-950e-2a4861f10ead","Type":"ContainerDied","Data":"1fef7fe46aea33effab1ac2d9426526c068c91e8e13e3d225bd3be7168a65afd"} Dec 11 02:20:50 crc kubenswrapper[4824]: I1211 02:20:50.652269 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1fef7fe46aea33effab1ac2d9426526c068c91e8e13e3d225bd3be7168a65afd" Dec 11 02:20:50 crc kubenswrapper[4824]: I1211 02:20:50.652287 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-d963-account-create-update-k6d7n" event={"ID":"d7b35324-f201-4847-9e8e-062029d43cee","Type":"ContainerDied","Data":"f6ae44cf16296e7a2f1b434f373e18637341fa36567ad2bf4e79f8bb057805e4"} Dec 11 02:20:50 crc kubenswrapper[4824]: I1211 02:20:50.652305 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f6ae44cf16296e7a2f1b434f373e18637341fa36567ad2bf4e79f8bb057805e4" Dec 11 02:20:52 crc kubenswrapper[4824]: I1211 02:20:52.128819 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Dec 11 02:20:53 crc kubenswrapper[4824]: I1211 02:20:53.022662 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7913c5de-48cb-4e9a-8b9b-ee976476bab2-etc-swift\") pod \"swift-storage-0\" (UID: \"7913c5de-48cb-4e9a-8b9b-ee976476bab2\") " pod="openstack/swift-storage-0" Dec 11 02:20:53 crc kubenswrapper[4824]: E1211 02:20:53.022935 4824 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 11 02:20:53 crc kubenswrapper[4824]: E1211 02:20:53.022976 4824 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 11 02:20:53 crc kubenswrapper[4824]: E1211 02:20:53.023073 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/7913c5de-48cb-4e9a-8b9b-ee976476bab2-etc-swift podName:7913c5de-48cb-4e9a-8b9b-ee976476bab2 nodeName:}" failed. No retries permitted until 2025-12-11 02:21:09.023044701 +0000 UTC m=+1210.712082110 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/7913c5de-48cb-4e9a-8b9b-ee976476bab2-etc-swift") pod "swift-storage-0" (UID: "7913c5de-48cb-4e9a-8b9b-ee976476bab2") : configmap "swift-ring-files" not found Dec 11 02:20:53 crc kubenswrapper[4824]: I1211 02:20:53.679735 4824 generic.go:334] "Generic (PLEG): container finished" podID="5719adff-b313-426e-afd1-69986bdc81bc" containerID="a15261146f01f11324c688f9e1a3aabe5b76507fa3748a92cbe03c2cc4d4f211" exitCode=0 Dec 11 02:20:53 crc kubenswrapper[4824]: I1211 02:20:53.679793 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-zbgp5" event={"ID":"5719adff-b313-426e-afd1-69986bdc81bc","Type":"ContainerDied","Data":"a15261146f01f11324c688f9e1a3aabe5b76507fa3748a92cbe03c2cc4d4f211"} Dec 11 02:20:54 crc kubenswrapper[4824]: I1211 02:20:54.311880 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-4ctff"] Dec 11 02:20:54 crc kubenswrapper[4824]: E1211 02:20:54.312338 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7b35324-f201-4847-9e8e-062029d43cee" containerName="mariadb-account-create-update" Dec 11 02:20:54 crc kubenswrapper[4824]: I1211 02:20:54.312365 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7b35324-f201-4847-9e8e-062029d43cee" containerName="mariadb-account-create-update" Dec 11 02:20:54 crc kubenswrapper[4824]: E1211 02:20:54.312377 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fdb2591b-b56c-48f0-82a4-c5752e9107b0" containerName="init" Dec 11 02:20:54 crc kubenswrapper[4824]: I1211 02:20:54.312385 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="fdb2591b-b56c-48f0-82a4-c5752e9107b0" containerName="init" Dec 11 02:20:54 crc kubenswrapper[4824]: E1211 02:20:54.312401 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c259c18-68cf-4a86-9870-4f481ba4b12a" containerName="mariadb-account-create-update" Dec 11 02:20:54 crc kubenswrapper[4824]: I1211 02:20:54.312410 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c259c18-68cf-4a86-9870-4f481ba4b12a" containerName="mariadb-account-create-update" Dec 11 02:20:54 crc kubenswrapper[4824]: E1211 02:20:54.312426 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fdb2591b-b56c-48f0-82a4-c5752e9107b0" containerName="dnsmasq-dns" Dec 11 02:20:54 crc kubenswrapper[4824]: I1211 02:20:54.312435 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="fdb2591b-b56c-48f0-82a4-c5752e9107b0" containerName="dnsmasq-dns" Dec 11 02:20:54 crc kubenswrapper[4824]: E1211 02:20:54.312448 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77484236-1254-407f-950e-2a4861f10ead" containerName="mariadb-database-create" Dec 11 02:20:54 crc kubenswrapper[4824]: I1211 02:20:54.312457 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="77484236-1254-407f-950e-2a4861f10ead" containerName="mariadb-database-create" Dec 11 02:20:54 crc kubenswrapper[4824]: E1211 02:20:54.312474 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d62e4855-c5c9-4d8b-ab6d-7695bbbe3f6b" containerName="mariadb-database-create" Dec 11 02:20:54 crc kubenswrapper[4824]: I1211 02:20:54.312483 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d62e4855-c5c9-4d8b-ab6d-7695bbbe3f6b" containerName="mariadb-database-create" Dec 11 02:20:54 crc kubenswrapper[4824]: E1211 02:20:54.312493 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2549107d-73ea-4e06-94a5-ff7a1f6b30b6" containerName="mariadb-database-create" Dec 11 02:20:54 crc kubenswrapper[4824]: I1211 02:20:54.312501 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="2549107d-73ea-4e06-94a5-ff7a1f6b30b6" containerName="mariadb-database-create" Dec 11 02:20:54 crc kubenswrapper[4824]: E1211 02:20:54.312521 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf6d5db8-d6c0-4e6e-b8d1-22b94c561060" containerName="mariadb-account-create-update" Dec 11 02:20:54 crc kubenswrapper[4824]: I1211 02:20:54.312528 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf6d5db8-d6c0-4e6e-b8d1-22b94c561060" containerName="mariadb-account-create-update" Dec 11 02:20:54 crc kubenswrapper[4824]: I1211 02:20:54.312750 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf6d5db8-d6c0-4e6e-b8d1-22b94c561060" containerName="mariadb-account-create-update" Dec 11 02:20:54 crc kubenswrapper[4824]: I1211 02:20:54.312763 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="77484236-1254-407f-950e-2a4861f10ead" containerName="mariadb-database-create" Dec 11 02:20:54 crc kubenswrapper[4824]: I1211 02:20:54.312776 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="d62e4855-c5c9-4d8b-ab6d-7695bbbe3f6b" containerName="mariadb-database-create" Dec 11 02:20:54 crc kubenswrapper[4824]: I1211 02:20:54.312796 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c259c18-68cf-4a86-9870-4f481ba4b12a" containerName="mariadb-account-create-update" Dec 11 02:20:54 crc kubenswrapper[4824]: I1211 02:20:54.312808 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="2549107d-73ea-4e06-94a5-ff7a1f6b30b6" containerName="mariadb-database-create" Dec 11 02:20:54 crc kubenswrapper[4824]: I1211 02:20:54.312821 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7b35324-f201-4847-9e8e-062029d43cee" containerName="mariadb-account-create-update" Dec 11 02:20:54 crc kubenswrapper[4824]: I1211 02:20:54.312834 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="fdb2591b-b56c-48f0-82a4-c5752e9107b0" containerName="dnsmasq-dns" Dec 11 02:20:54 crc kubenswrapper[4824]: I1211 02:20:54.313431 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-4ctff" Dec 11 02:20:54 crc kubenswrapper[4824]: I1211 02:20:54.317483 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Dec 11 02:20:54 crc kubenswrapper[4824]: I1211 02:20:54.317768 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-5plrn" Dec 11 02:20:54 crc kubenswrapper[4824]: I1211 02:20:54.330631 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-4ctff"] Dec 11 02:20:54 crc kubenswrapper[4824]: I1211 02:20:54.447776 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3b4a47f6-97f5-4395-b82a-166b934174b5-db-sync-config-data\") pod \"glance-db-sync-4ctff\" (UID: \"3b4a47f6-97f5-4395-b82a-166b934174b5\") " pod="openstack/glance-db-sync-4ctff" Dec 11 02:20:54 crc kubenswrapper[4824]: I1211 02:20:54.447885 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b4a47f6-97f5-4395-b82a-166b934174b5-config-data\") pod \"glance-db-sync-4ctff\" (UID: \"3b4a47f6-97f5-4395-b82a-166b934174b5\") " pod="openstack/glance-db-sync-4ctff" Dec 11 02:20:54 crc kubenswrapper[4824]: I1211 02:20:54.448063 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cxxvl\" (UniqueName: \"kubernetes.io/projected/3b4a47f6-97f5-4395-b82a-166b934174b5-kube-api-access-cxxvl\") pod \"glance-db-sync-4ctff\" (UID: \"3b4a47f6-97f5-4395-b82a-166b934174b5\") " pod="openstack/glance-db-sync-4ctff" Dec 11 02:20:54 crc kubenswrapper[4824]: I1211 02:20:54.448260 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b4a47f6-97f5-4395-b82a-166b934174b5-combined-ca-bundle\") pod \"glance-db-sync-4ctff\" (UID: \"3b4a47f6-97f5-4395-b82a-166b934174b5\") " pod="openstack/glance-db-sync-4ctff" Dec 11 02:20:54 crc kubenswrapper[4824]: I1211 02:20:54.551036 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3b4a47f6-97f5-4395-b82a-166b934174b5-db-sync-config-data\") pod \"glance-db-sync-4ctff\" (UID: \"3b4a47f6-97f5-4395-b82a-166b934174b5\") " pod="openstack/glance-db-sync-4ctff" Dec 11 02:20:54 crc kubenswrapper[4824]: I1211 02:20:54.551142 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b4a47f6-97f5-4395-b82a-166b934174b5-config-data\") pod \"glance-db-sync-4ctff\" (UID: \"3b4a47f6-97f5-4395-b82a-166b934174b5\") " pod="openstack/glance-db-sync-4ctff" Dec 11 02:20:54 crc kubenswrapper[4824]: I1211 02:20:54.551235 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cxxvl\" (UniqueName: \"kubernetes.io/projected/3b4a47f6-97f5-4395-b82a-166b934174b5-kube-api-access-cxxvl\") pod \"glance-db-sync-4ctff\" (UID: \"3b4a47f6-97f5-4395-b82a-166b934174b5\") " pod="openstack/glance-db-sync-4ctff" Dec 11 02:20:54 crc kubenswrapper[4824]: I1211 02:20:54.551341 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b4a47f6-97f5-4395-b82a-166b934174b5-combined-ca-bundle\") pod \"glance-db-sync-4ctff\" (UID: \"3b4a47f6-97f5-4395-b82a-166b934174b5\") " pod="openstack/glance-db-sync-4ctff" Dec 11 02:20:54 crc kubenswrapper[4824]: I1211 02:20:54.558960 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b4a47f6-97f5-4395-b82a-166b934174b5-config-data\") pod \"glance-db-sync-4ctff\" (UID: \"3b4a47f6-97f5-4395-b82a-166b934174b5\") " pod="openstack/glance-db-sync-4ctff" Dec 11 02:20:54 crc kubenswrapper[4824]: I1211 02:20:54.559467 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b4a47f6-97f5-4395-b82a-166b934174b5-combined-ca-bundle\") pod \"glance-db-sync-4ctff\" (UID: \"3b4a47f6-97f5-4395-b82a-166b934174b5\") " pod="openstack/glance-db-sync-4ctff" Dec 11 02:20:54 crc kubenswrapper[4824]: I1211 02:20:54.559841 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3b4a47f6-97f5-4395-b82a-166b934174b5-db-sync-config-data\") pod \"glance-db-sync-4ctff\" (UID: \"3b4a47f6-97f5-4395-b82a-166b934174b5\") " pod="openstack/glance-db-sync-4ctff" Dec 11 02:20:54 crc kubenswrapper[4824]: I1211 02:20:54.585728 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cxxvl\" (UniqueName: \"kubernetes.io/projected/3b4a47f6-97f5-4395-b82a-166b934174b5-kube-api-access-cxxvl\") pod \"glance-db-sync-4ctff\" (UID: \"3b4a47f6-97f5-4395-b82a-166b934174b5\") " pod="openstack/glance-db-sync-4ctff" Dec 11 02:20:54 crc kubenswrapper[4824]: I1211 02:20:54.639894 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-4ctff" Dec 11 02:20:55 crc kubenswrapper[4824]: I1211 02:20:55.095358 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-zbgp5" Dec 11 02:20:55 crc kubenswrapper[4824]: I1211 02:20:55.263023 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5719adff-b313-426e-afd1-69986bdc81bc-swiftconf\") pod \"5719adff-b313-426e-afd1-69986bdc81bc\" (UID: \"5719adff-b313-426e-afd1-69986bdc81bc\") " Dec 11 02:20:55 crc kubenswrapper[4824]: I1211 02:20:55.263193 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5719adff-b313-426e-afd1-69986bdc81bc-scripts\") pod \"5719adff-b313-426e-afd1-69986bdc81bc\" (UID: \"5719adff-b313-426e-afd1-69986bdc81bc\") " Dec 11 02:20:55 crc kubenswrapper[4824]: I1211 02:20:55.263271 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z55vb\" (UniqueName: \"kubernetes.io/projected/5719adff-b313-426e-afd1-69986bdc81bc-kube-api-access-z55vb\") pod \"5719adff-b313-426e-afd1-69986bdc81bc\" (UID: \"5719adff-b313-426e-afd1-69986bdc81bc\") " Dec 11 02:20:55 crc kubenswrapper[4824]: I1211 02:20:55.263301 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5719adff-b313-426e-afd1-69986bdc81bc-etc-swift\") pod \"5719adff-b313-426e-afd1-69986bdc81bc\" (UID: \"5719adff-b313-426e-afd1-69986bdc81bc\") " Dec 11 02:20:55 crc kubenswrapper[4824]: I1211 02:20:55.263325 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5719adff-b313-426e-afd1-69986bdc81bc-ring-data-devices\") pod \"5719adff-b313-426e-afd1-69986bdc81bc\" (UID: \"5719adff-b313-426e-afd1-69986bdc81bc\") " Dec 11 02:20:55 crc kubenswrapper[4824]: I1211 02:20:55.263385 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5719adff-b313-426e-afd1-69986bdc81bc-dispersionconf\") pod \"5719adff-b313-426e-afd1-69986bdc81bc\" (UID: \"5719adff-b313-426e-afd1-69986bdc81bc\") " Dec 11 02:20:55 crc kubenswrapper[4824]: I1211 02:20:55.263415 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5719adff-b313-426e-afd1-69986bdc81bc-combined-ca-bundle\") pod \"5719adff-b313-426e-afd1-69986bdc81bc\" (UID: \"5719adff-b313-426e-afd1-69986bdc81bc\") " Dec 11 02:20:55 crc kubenswrapper[4824]: I1211 02:20:55.265222 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5719adff-b313-426e-afd1-69986bdc81bc-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "5719adff-b313-426e-afd1-69986bdc81bc" (UID: "5719adff-b313-426e-afd1-69986bdc81bc"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:20:55 crc kubenswrapper[4824]: I1211 02:20:55.265378 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5719adff-b313-426e-afd1-69986bdc81bc-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "5719adff-b313-426e-afd1-69986bdc81bc" (UID: "5719adff-b313-426e-afd1-69986bdc81bc"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:20:55 crc kubenswrapper[4824]: I1211 02:20:55.269650 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5719adff-b313-426e-afd1-69986bdc81bc-kube-api-access-z55vb" (OuterVolumeSpecName: "kube-api-access-z55vb") pod "5719adff-b313-426e-afd1-69986bdc81bc" (UID: "5719adff-b313-426e-afd1-69986bdc81bc"). InnerVolumeSpecName "kube-api-access-z55vb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:20:55 crc kubenswrapper[4824]: I1211 02:20:55.270783 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5719adff-b313-426e-afd1-69986bdc81bc-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "5719adff-b313-426e-afd1-69986bdc81bc" (UID: "5719adff-b313-426e-afd1-69986bdc81bc"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:20:55 crc kubenswrapper[4824]: I1211 02:20:55.287136 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5719adff-b313-426e-afd1-69986bdc81bc-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "5719adff-b313-426e-afd1-69986bdc81bc" (UID: "5719adff-b313-426e-afd1-69986bdc81bc"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:20:55 crc kubenswrapper[4824]: I1211 02:20:55.296970 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5719adff-b313-426e-afd1-69986bdc81bc-scripts" (OuterVolumeSpecName: "scripts") pod "5719adff-b313-426e-afd1-69986bdc81bc" (UID: "5719adff-b313-426e-afd1-69986bdc81bc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:20:55 crc kubenswrapper[4824]: I1211 02:20:55.305155 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5719adff-b313-426e-afd1-69986bdc81bc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5719adff-b313-426e-afd1-69986bdc81bc" (UID: "5719adff-b313-426e-afd1-69986bdc81bc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:20:55 crc kubenswrapper[4824]: I1211 02:20:55.320973 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-4ctff"] Dec 11 02:20:55 crc kubenswrapper[4824]: I1211 02:20:55.366693 4824 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5719adff-b313-426e-afd1-69986bdc81bc-swiftconf\") on node \"crc\" DevicePath \"\"" Dec 11 02:20:55 crc kubenswrapper[4824]: I1211 02:20:55.366726 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5719adff-b313-426e-afd1-69986bdc81bc-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 02:20:55 crc kubenswrapper[4824]: I1211 02:20:55.366740 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z55vb\" (UniqueName: \"kubernetes.io/projected/5719adff-b313-426e-afd1-69986bdc81bc-kube-api-access-z55vb\") on node \"crc\" DevicePath \"\"" Dec 11 02:20:55 crc kubenswrapper[4824]: I1211 02:20:55.366755 4824 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5719adff-b313-426e-afd1-69986bdc81bc-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 11 02:20:55 crc kubenswrapper[4824]: I1211 02:20:55.366766 4824 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5719adff-b313-426e-afd1-69986bdc81bc-ring-data-devices\") on node \"crc\" DevicePath \"\"" Dec 11 02:20:55 crc kubenswrapper[4824]: I1211 02:20:55.366778 4824 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5719adff-b313-426e-afd1-69986bdc81bc-dispersionconf\") on node \"crc\" DevicePath \"\"" Dec 11 02:20:55 crc kubenswrapper[4824]: I1211 02:20:55.366790 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5719adff-b313-426e-afd1-69986bdc81bc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 02:20:55 crc kubenswrapper[4824]: I1211 02:20:55.696033 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-zbgp5" event={"ID":"5719adff-b313-426e-afd1-69986bdc81bc","Type":"ContainerDied","Data":"7afc8648378b67da8f8ede79e10c66a50be938dcd693e835dc589cdc4f561a27"} Dec 11 02:20:55 crc kubenswrapper[4824]: I1211 02:20:55.696090 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7afc8648378b67da8f8ede79e10c66a50be938dcd693e835dc589cdc4f561a27" Dec 11 02:20:55 crc kubenswrapper[4824]: I1211 02:20:55.696179 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-zbgp5" Dec 11 02:20:55 crc kubenswrapper[4824]: I1211 02:20:55.697632 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-4ctff" event={"ID":"3b4a47f6-97f5-4395-b82a-166b934174b5","Type":"ContainerStarted","Data":"38a31d452ee585811a8351977fa04871cf57dabd2fa50c5688836ba6f3cfdf95"} Dec 11 02:20:59 crc kubenswrapper[4824]: I1211 02:20:59.629533 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-779x4" podUID="148c69f8-3121-4a55-b3a6-be44b816f643" containerName="ovn-controller" probeResult="failure" output=< Dec 11 02:20:59 crc kubenswrapper[4824]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 11 02:20:59 crc kubenswrapper[4824]: > Dec 11 02:21:02 crc kubenswrapper[4824]: I1211 02:21:02.753168 4824 generic.go:334] "Generic (PLEG): container finished" podID="1fd840b6-4ace-45b8-a8d2-38a041110095" containerID="690854c03401211031fd0831cdfab50c3cc32ac8bb1ccb2f7ab9e7897c3a6efe" exitCode=0 Dec 11 02:21:02 crc kubenswrapper[4824]: I1211 02:21:02.753314 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1fd840b6-4ace-45b8-a8d2-38a041110095","Type":"ContainerDied","Data":"690854c03401211031fd0831cdfab50c3cc32ac8bb1ccb2f7ab9e7897c3a6efe"} Dec 11 02:21:03 crc kubenswrapper[4824]: I1211 02:21:03.773259 4824 generic.go:334] "Generic (PLEG): container finished" podID="ff37fa2f-dc06-430a-a4fe-af2b379fe563" containerID="1d02f24dc933f44c12ca63c4374960b6bf5f14ea8005dd772dbd07f54f6c8f51" exitCode=0 Dec 11 02:21:03 crc kubenswrapper[4824]: I1211 02:21:03.773341 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ff37fa2f-dc06-430a-a4fe-af2b379fe563","Type":"ContainerDied","Data":"1d02f24dc933f44c12ca63c4374960b6bf5f14ea8005dd772dbd07f54f6c8f51"} Dec 11 02:21:04 crc kubenswrapper[4824]: I1211 02:21:04.628735 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-779x4" podUID="148c69f8-3121-4a55-b3a6-be44b816f643" containerName="ovn-controller" probeResult="failure" output=< Dec 11 02:21:04 crc kubenswrapper[4824]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 11 02:21:04 crc kubenswrapper[4824]: > Dec 11 02:21:04 crc kubenswrapper[4824]: I1211 02:21:04.676166 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-xzkzd" Dec 11 02:21:04 crc kubenswrapper[4824]: I1211 02:21:04.690901 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-xzkzd" Dec 11 02:21:04 crc kubenswrapper[4824]: I1211 02:21:04.949893 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-779x4-config-v9wcm"] Dec 11 02:21:04 crc kubenswrapper[4824]: E1211 02:21:04.950376 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5719adff-b313-426e-afd1-69986bdc81bc" containerName="swift-ring-rebalance" Dec 11 02:21:04 crc kubenswrapper[4824]: I1211 02:21:04.950428 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="5719adff-b313-426e-afd1-69986bdc81bc" containerName="swift-ring-rebalance" Dec 11 02:21:04 crc kubenswrapper[4824]: I1211 02:21:04.950676 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="5719adff-b313-426e-afd1-69986bdc81bc" containerName="swift-ring-rebalance" Dec 11 02:21:04 crc kubenswrapper[4824]: I1211 02:21:04.951391 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-779x4-config-v9wcm" Dec 11 02:21:04 crc kubenswrapper[4824]: I1211 02:21:04.954255 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 11 02:21:04 crc kubenswrapper[4824]: I1211 02:21:04.957357 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-779x4-config-v9wcm"] Dec 11 02:21:05 crc kubenswrapper[4824]: I1211 02:21:05.043995 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/249d9815-d46f-4cf1-a8b9-f8831450b708-var-log-ovn\") pod \"ovn-controller-779x4-config-v9wcm\" (UID: \"249d9815-d46f-4cf1-a8b9-f8831450b708\") " pod="openstack/ovn-controller-779x4-config-v9wcm" Dec 11 02:21:05 crc kubenswrapper[4824]: I1211 02:21:05.044083 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/249d9815-d46f-4cf1-a8b9-f8831450b708-additional-scripts\") pod \"ovn-controller-779x4-config-v9wcm\" (UID: \"249d9815-d46f-4cf1-a8b9-f8831450b708\") " pod="openstack/ovn-controller-779x4-config-v9wcm" Dec 11 02:21:05 crc kubenswrapper[4824]: I1211 02:21:05.044142 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/249d9815-d46f-4cf1-a8b9-f8831450b708-var-run-ovn\") pod \"ovn-controller-779x4-config-v9wcm\" (UID: \"249d9815-d46f-4cf1-a8b9-f8831450b708\") " pod="openstack/ovn-controller-779x4-config-v9wcm" Dec 11 02:21:05 crc kubenswrapper[4824]: I1211 02:21:05.044184 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/249d9815-d46f-4cf1-a8b9-f8831450b708-scripts\") pod \"ovn-controller-779x4-config-v9wcm\" (UID: \"249d9815-d46f-4cf1-a8b9-f8831450b708\") " pod="openstack/ovn-controller-779x4-config-v9wcm" Dec 11 02:21:05 crc kubenswrapper[4824]: I1211 02:21:05.044243 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/249d9815-d46f-4cf1-a8b9-f8831450b708-var-run\") pod \"ovn-controller-779x4-config-v9wcm\" (UID: \"249d9815-d46f-4cf1-a8b9-f8831450b708\") " pod="openstack/ovn-controller-779x4-config-v9wcm" Dec 11 02:21:05 crc kubenswrapper[4824]: I1211 02:21:05.044383 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q7c4v\" (UniqueName: \"kubernetes.io/projected/249d9815-d46f-4cf1-a8b9-f8831450b708-kube-api-access-q7c4v\") pod \"ovn-controller-779x4-config-v9wcm\" (UID: \"249d9815-d46f-4cf1-a8b9-f8831450b708\") " pod="openstack/ovn-controller-779x4-config-v9wcm" Dec 11 02:21:05 crc kubenswrapper[4824]: I1211 02:21:05.146089 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/249d9815-d46f-4cf1-a8b9-f8831450b708-scripts\") pod \"ovn-controller-779x4-config-v9wcm\" (UID: \"249d9815-d46f-4cf1-a8b9-f8831450b708\") " pod="openstack/ovn-controller-779x4-config-v9wcm" Dec 11 02:21:05 crc kubenswrapper[4824]: I1211 02:21:05.146188 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/249d9815-d46f-4cf1-a8b9-f8831450b708-var-run\") pod \"ovn-controller-779x4-config-v9wcm\" (UID: \"249d9815-d46f-4cf1-a8b9-f8831450b708\") " pod="openstack/ovn-controller-779x4-config-v9wcm" Dec 11 02:21:05 crc kubenswrapper[4824]: I1211 02:21:05.146237 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q7c4v\" (UniqueName: \"kubernetes.io/projected/249d9815-d46f-4cf1-a8b9-f8831450b708-kube-api-access-q7c4v\") pod \"ovn-controller-779x4-config-v9wcm\" (UID: \"249d9815-d46f-4cf1-a8b9-f8831450b708\") " pod="openstack/ovn-controller-779x4-config-v9wcm" Dec 11 02:21:05 crc kubenswrapper[4824]: I1211 02:21:05.146259 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/249d9815-d46f-4cf1-a8b9-f8831450b708-var-log-ovn\") pod \"ovn-controller-779x4-config-v9wcm\" (UID: \"249d9815-d46f-4cf1-a8b9-f8831450b708\") " pod="openstack/ovn-controller-779x4-config-v9wcm" Dec 11 02:21:05 crc kubenswrapper[4824]: I1211 02:21:05.146304 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/249d9815-d46f-4cf1-a8b9-f8831450b708-additional-scripts\") pod \"ovn-controller-779x4-config-v9wcm\" (UID: \"249d9815-d46f-4cf1-a8b9-f8831450b708\") " pod="openstack/ovn-controller-779x4-config-v9wcm" Dec 11 02:21:05 crc kubenswrapper[4824]: I1211 02:21:05.146337 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/249d9815-d46f-4cf1-a8b9-f8831450b708-var-run-ovn\") pod \"ovn-controller-779x4-config-v9wcm\" (UID: \"249d9815-d46f-4cf1-a8b9-f8831450b708\") " pod="openstack/ovn-controller-779x4-config-v9wcm" Dec 11 02:21:05 crc kubenswrapper[4824]: I1211 02:21:05.146607 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/249d9815-d46f-4cf1-a8b9-f8831450b708-var-run-ovn\") pod \"ovn-controller-779x4-config-v9wcm\" (UID: \"249d9815-d46f-4cf1-a8b9-f8831450b708\") " pod="openstack/ovn-controller-779x4-config-v9wcm" Dec 11 02:21:05 crc kubenswrapper[4824]: I1211 02:21:05.146617 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/249d9815-d46f-4cf1-a8b9-f8831450b708-var-run\") pod \"ovn-controller-779x4-config-v9wcm\" (UID: \"249d9815-d46f-4cf1-a8b9-f8831450b708\") " pod="openstack/ovn-controller-779x4-config-v9wcm" Dec 11 02:21:05 crc kubenswrapper[4824]: I1211 02:21:05.146668 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/249d9815-d46f-4cf1-a8b9-f8831450b708-var-log-ovn\") pod \"ovn-controller-779x4-config-v9wcm\" (UID: \"249d9815-d46f-4cf1-a8b9-f8831450b708\") " pod="openstack/ovn-controller-779x4-config-v9wcm" Dec 11 02:21:05 crc kubenswrapper[4824]: I1211 02:21:05.148439 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/249d9815-d46f-4cf1-a8b9-f8831450b708-additional-scripts\") pod \"ovn-controller-779x4-config-v9wcm\" (UID: \"249d9815-d46f-4cf1-a8b9-f8831450b708\") " pod="openstack/ovn-controller-779x4-config-v9wcm" Dec 11 02:21:05 crc kubenswrapper[4824]: I1211 02:21:05.156611 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/249d9815-d46f-4cf1-a8b9-f8831450b708-scripts\") pod \"ovn-controller-779x4-config-v9wcm\" (UID: \"249d9815-d46f-4cf1-a8b9-f8831450b708\") " pod="openstack/ovn-controller-779x4-config-v9wcm" Dec 11 02:21:05 crc kubenswrapper[4824]: I1211 02:21:05.167723 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q7c4v\" (UniqueName: \"kubernetes.io/projected/249d9815-d46f-4cf1-a8b9-f8831450b708-kube-api-access-q7c4v\") pod \"ovn-controller-779x4-config-v9wcm\" (UID: \"249d9815-d46f-4cf1-a8b9-f8831450b708\") " pod="openstack/ovn-controller-779x4-config-v9wcm" Dec 11 02:21:05 crc kubenswrapper[4824]: I1211 02:21:05.269020 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-779x4-config-v9wcm" Dec 11 02:21:08 crc kubenswrapper[4824]: I1211 02:21:08.204586 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-779x4-config-v9wcm"] Dec 11 02:21:08 crc kubenswrapper[4824]: W1211 02:21:08.211630 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod249d9815_d46f_4cf1_a8b9_f8831450b708.slice/crio-a5437e1c9d224935429b40914741013a94ffe2958da104ffe1cb99e879a921e8 WatchSource:0}: Error finding container a5437e1c9d224935429b40914741013a94ffe2958da104ffe1cb99e879a921e8: Status 404 returned error can't find the container with id a5437e1c9d224935429b40914741013a94ffe2958da104ffe1cb99e879a921e8 Dec 11 02:21:08 crc kubenswrapper[4824]: I1211 02:21:08.818644 4824 generic.go:334] "Generic (PLEG): container finished" podID="249d9815-d46f-4cf1-a8b9-f8831450b708" containerID="cbe3be8fd77c4538b4c67e74d833cf6343d64462c38ce95f2eb81652eb1887a2" exitCode=0 Dec 11 02:21:08 crc kubenswrapper[4824]: I1211 02:21:08.818705 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-779x4-config-v9wcm" event={"ID":"249d9815-d46f-4cf1-a8b9-f8831450b708","Type":"ContainerDied","Data":"cbe3be8fd77c4538b4c67e74d833cf6343d64462c38ce95f2eb81652eb1887a2"} Dec 11 02:21:08 crc kubenswrapper[4824]: I1211 02:21:08.818952 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-779x4-config-v9wcm" event={"ID":"249d9815-d46f-4cf1-a8b9-f8831450b708","Type":"ContainerStarted","Data":"a5437e1c9d224935429b40914741013a94ffe2958da104ffe1cb99e879a921e8"} Dec 11 02:21:08 crc kubenswrapper[4824]: I1211 02:21:08.821026 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ff37fa2f-dc06-430a-a4fe-af2b379fe563","Type":"ContainerStarted","Data":"69c1077bbc1e742ec896e90a037d4489755f60af0094fc7227c893dc106fa13b"} Dec 11 02:21:08 crc kubenswrapper[4824]: I1211 02:21:08.821256 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 11 02:21:08 crc kubenswrapper[4824]: I1211 02:21:08.822767 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-4ctff" event={"ID":"3b4a47f6-97f5-4395-b82a-166b934174b5","Type":"ContainerStarted","Data":"13a2e99f97a6fd8fd4b48427476f27d577b17501906194f19debf85ca7f2b4c4"} Dec 11 02:21:08 crc kubenswrapper[4824]: I1211 02:21:08.824576 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1fd840b6-4ace-45b8-a8d2-38a041110095","Type":"ContainerStarted","Data":"8dee72102c0022a859a56e54e0f34c0048fded2348fc7ecb0d3de490ad11c4cd"} Dec 11 02:21:08 crc kubenswrapper[4824]: I1211 02:21:08.824765 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 11 02:21:08 crc kubenswrapper[4824]: I1211 02:21:08.862934 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-4ctff" podStartSLOduration=2.338807295 podStartE2EDuration="14.862915892s" podCreationTimestamp="2025-12-11 02:20:54 +0000 UTC" firstStartedPulling="2025-12-11 02:20:55.337955447 +0000 UTC m=+1197.026992866" lastFinishedPulling="2025-12-11 02:21:07.862064084 +0000 UTC m=+1209.551101463" observedRunningTime="2025-12-11 02:21:08.85482456 +0000 UTC m=+1210.543861979" watchObservedRunningTime="2025-12-11 02:21:08.862915892 +0000 UTC m=+1210.551953271" Dec 11 02:21:08 crc kubenswrapper[4824]: I1211 02:21:08.892857 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=62.161117735 podStartE2EDuration="1m9.892833573s" podCreationTimestamp="2025-12-11 02:19:59 +0000 UTC" firstStartedPulling="2025-12-11 02:20:19.477740677 +0000 UTC m=+1161.166778056" lastFinishedPulling="2025-12-11 02:20:27.209456475 +0000 UTC m=+1168.898493894" observedRunningTime="2025-12-11 02:21:08.887078405 +0000 UTC m=+1210.576115814" watchObservedRunningTime="2025-12-11 02:21:08.892833573 +0000 UTC m=+1210.581870992" Dec 11 02:21:09 crc kubenswrapper[4824]: I1211 02:21:09.116002 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7913c5de-48cb-4e9a-8b9b-ee976476bab2-etc-swift\") pod \"swift-storage-0\" (UID: \"7913c5de-48cb-4e9a-8b9b-ee976476bab2\") " pod="openstack/swift-storage-0" Dec 11 02:21:09 crc kubenswrapper[4824]: I1211 02:21:09.126798 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7913c5de-48cb-4e9a-8b9b-ee976476bab2-etc-swift\") pod \"swift-storage-0\" (UID: \"7913c5de-48cb-4e9a-8b9b-ee976476bab2\") " pod="openstack/swift-storage-0" Dec 11 02:21:09 crc kubenswrapper[4824]: I1211 02:21:09.288096 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 11 02:21:09 crc kubenswrapper[4824]: I1211 02:21:09.605556 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-779x4" Dec 11 02:21:09 crc kubenswrapper[4824]: I1211 02:21:09.625375 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=62.304574506 podStartE2EDuration="1m10.625359862s" podCreationTimestamp="2025-12-11 02:19:59 +0000 UTC" firstStartedPulling="2025-12-11 02:20:19.477799618 +0000 UTC m=+1161.166836997" lastFinishedPulling="2025-12-11 02:20:27.798584964 +0000 UTC m=+1169.487622353" observedRunningTime="2025-12-11 02:21:08.927503474 +0000 UTC m=+1210.616540893" watchObservedRunningTime="2025-12-11 02:21:09.625359862 +0000 UTC m=+1211.314397231" Dec 11 02:21:09 crc kubenswrapper[4824]: I1211 02:21:09.812993 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 11 02:21:09 crc kubenswrapper[4824]: W1211 02:21:09.817561 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7913c5de_48cb_4e9a_8b9b_ee976476bab2.slice/crio-36aa0773644915663404b62b9e9e5a73c6ba6df96169ac9155a1e4e3432c5f6c WatchSource:0}: Error finding container 36aa0773644915663404b62b9e9e5a73c6ba6df96169ac9155a1e4e3432c5f6c: Status 404 returned error can't find the container with id 36aa0773644915663404b62b9e9e5a73c6ba6df96169ac9155a1e4e3432c5f6c Dec 11 02:21:09 crc kubenswrapper[4824]: I1211 02:21:09.833283 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"7913c5de-48cb-4e9a-8b9b-ee976476bab2","Type":"ContainerStarted","Data":"36aa0773644915663404b62b9e9e5a73c6ba6df96169ac9155a1e4e3432c5f6c"} Dec 11 02:21:10 crc kubenswrapper[4824]: I1211 02:21:10.191180 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-779x4-config-v9wcm" Dec 11 02:21:10 crc kubenswrapper[4824]: I1211 02:21:10.254747 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/249d9815-d46f-4cf1-a8b9-f8831450b708-var-run-ovn\") pod \"249d9815-d46f-4cf1-a8b9-f8831450b708\" (UID: \"249d9815-d46f-4cf1-a8b9-f8831450b708\") " Dec 11 02:21:10 crc kubenswrapper[4824]: I1211 02:21:10.254833 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/249d9815-d46f-4cf1-a8b9-f8831450b708-scripts\") pod \"249d9815-d46f-4cf1-a8b9-f8831450b708\" (UID: \"249d9815-d46f-4cf1-a8b9-f8831450b708\") " Dec 11 02:21:10 crc kubenswrapper[4824]: I1211 02:21:10.254853 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/249d9815-d46f-4cf1-a8b9-f8831450b708-var-log-ovn\") pod \"249d9815-d46f-4cf1-a8b9-f8831450b708\" (UID: \"249d9815-d46f-4cf1-a8b9-f8831450b708\") " Dec 11 02:21:10 crc kubenswrapper[4824]: I1211 02:21:10.254842 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/249d9815-d46f-4cf1-a8b9-f8831450b708-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "249d9815-d46f-4cf1-a8b9-f8831450b708" (UID: "249d9815-d46f-4cf1-a8b9-f8831450b708"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 02:21:10 crc kubenswrapper[4824]: I1211 02:21:10.254892 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/249d9815-d46f-4cf1-a8b9-f8831450b708-additional-scripts\") pod \"249d9815-d46f-4cf1-a8b9-f8831450b708\" (UID: \"249d9815-d46f-4cf1-a8b9-f8831450b708\") " Dec 11 02:21:10 crc kubenswrapper[4824]: I1211 02:21:10.254912 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/249d9815-d46f-4cf1-a8b9-f8831450b708-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "249d9815-d46f-4cf1-a8b9-f8831450b708" (UID: "249d9815-d46f-4cf1-a8b9-f8831450b708"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 02:21:10 crc kubenswrapper[4824]: I1211 02:21:10.254930 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/249d9815-d46f-4cf1-a8b9-f8831450b708-var-run\") pod \"249d9815-d46f-4cf1-a8b9-f8831450b708\" (UID: \"249d9815-d46f-4cf1-a8b9-f8831450b708\") " Dec 11 02:21:10 crc kubenswrapper[4824]: I1211 02:21:10.254949 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q7c4v\" (UniqueName: \"kubernetes.io/projected/249d9815-d46f-4cf1-a8b9-f8831450b708-kube-api-access-q7c4v\") pod \"249d9815-d46f-4cf1-a8b9-f8831450b708\" (UID: \"249d9815-d46f-4cf1-a8b9-f8831450b708\") " Dec 11 02:21:10 crc kubenswrapper[4824]: I1211 02:21:10.254998 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/249d9815-d46f-4cf1-a8b9-f8831450b708-var-run" (OuterVolumeSpecName: "var-run") pod "249d9815-d46f-4cf1-a8b9-f8831450b708" (UID: "249d9815-d46f-4cf1-a8b9-f8831450b708"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 02:21:10 crc kubenswrapper[4824]: I1211 02:21:10.255290 4824 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/249d9815-d46f-4cf1-a8b9-f8831450b708-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 11 02:21:10 crc kubenswrapper[4824]: I1211 02:21:10.255302 4824 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/249d9815-d46f-4cf1-a8b9-f8831450b708-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 11 02:21:10 crc kubenswrapper[4824]: I1211 02:21:10.255310 4824 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/249d9815-d46f-4cf1-a8b9-f8831450b708-var-run\") on node \"crc\" DevicePath \"\"" Dec 11 02:21:10 crc kubenswrapper[4824]: I1211 02:21:10.255731 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/249d9815-d46f-4cf1-a8b9-f8831450b708-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "249d9815-d46f-4cf1-a8b9-f8831450b708" (UID: "249d9815-d46f-4cf1-a8b9-f8831450b708"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:21:10 crc kubenswrapper[4824]: I1211 02:21:10.256001 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/249d9815-d46f-4cf1-a8b9-f8831450b708-scripts" (OuterVolumeSpecName: "scripts") pod "249d9815-d46f-4cf1-a8b9-f8831450b708" (UID: "249d9815-d46f-4cf1-a8b9-f8831450b708"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:21:10 crc kubenswrapper[4824]: I1211 02:21:10.277008 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/249d9815-d46f-4cf1-a8b9-f8831450b708-kube-api-access-q7c4v" (OuterVolumeSpecName: "kube-api-access-q7c4v") pod "249d9815-d46f-4cf1-a8b9-f8831450b708" (UID: "249d9815-d46f-4cf1-a8b9-f8831450b708"). InnerVolumeSpecName "kube-api-access-q7c4v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:21:10 crc kubenswrapper[4824]: I1211 02:21:10.357037 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/249d9815-d46f-4cf1-a8b9-f8831450b708-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 02:21:10 crc kubenswrapper[4824]: I1211 02:21:10.357086 4824 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/249d9815-d46f-4cf1-a8b9-f8831450b708-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 02:21:10 crc kubenswrapper[4824]: I1211 02:21:10.357098 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q7c4v\" (UniqueName: \"kubernetes.io/projected/249d9815-d46f-4cf1-a8b9-f8831450b708-kube-api-access-q7c4v\") on node \"crc\" DevicePath \"\"" Dec 11 02:21:10 crc kubenswrapper[4824]: I1211 02:21:10.842556 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-779x4-config-v9wcm" event={"ID":"249d9815-d46f-4cf1-a8b9-f8831450b708","Type":"ContainerDied","Data":"a5437e1c9d224935429b40914741013a94ffe2958da104ffe1cb99e879a921e8"} Dec 11 02:21:10 crc kubenswrapper[4824]: I1211 02:21:10.842608 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a5437e1c9d224935429b40914741013a94ffe2958da104ffe1cb99e879a921e8" Dec 11 02:21:10 crc kubenswrapper[4824]: I1211 02:21:10.842672 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-779x4-config-v9wcm" Dec 11 02:21:11 crc kubenswrapper[4824]: I1211 02:21:11.311095 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-779x4-config-v9wcm"] Dec 11 02:21:11 crc kubenswrapper[4824]: I1211 02:21:11.325582 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-779x4-config-v9wcm"] Dec 11 02:21:11 crc kubenswrapper[4824]: I1211 02:21:11.437748 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-779x4-config-b56vn"] Dec 11 02:21:11 crc kubenswrapper[4824]: E1211 02:21:11.438152 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="249d9815-d46f-4cf1-a8b9-f8831450b708" containerName="ovn-config" Dec 11 02:21:11 crc kubenswrapper[4824]: I1211 02:21:11.438166 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="249d9815-d46f-4cf1-a8b9-f8831450b708" containerName="ovn-config" Dec 11 02:21:11 crc kubenswrapper[4824]: I1211 02:21:11.438343 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="249d9815-d46f-4cf1-a8b9-f8831450b708" containerName="ovn-config" Dec 11 02:21:11 crc kubenswrapper[4824]: I1211 02:21:11.438954 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-779x4-config-b56vn" Dec 11 02:21:11 crc kubenswrapper[4824]: I1211 02:21:11.440731 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 11 02:21:11 crc kubenswrapper[4824]: I1211 02:21:11.455468 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-779x4-config-b56vn"] Dec 11 02:21:11 crc kubenswrapper[4824]: I1211 02:21:11.477669 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/ebda38df-902c-4ea4-a8cf-89701d29ab5e-additional-scripts\") pod \"ovn-controller-779x4-config-b56vn\" (UID: \"ebda38df-902c-4ea4-a8cf-89701d29ab5e\") " pod="openstack/ovn-controller-779x4-config-b56vn" Dec 11 02:21:11 crc kubenswrapper[4824]: I1211 02:21:11.477762 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/ebda38df-902c-4ea4-a8cf-89701d29ab5e-var-run-ovn\") pod \"ovn-controller-779x4-config-b56vn\" (UID: \"ebda38df-902c-4ea4-a8cf-89701d29ab5e\") " pod="openstack/ovn-controller-779x4-config-b56vn" Dec 11 02:21:11 crc kubenswrapper[4824]: I1211 02:21:11.477809 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ebda38df-902c-4ea4-a8cf-89701d29ab5e-var-run\") pod \"ovn-controller-779x4-config-b56vn\" (UID: \"ebda38df-902c-4ea4-a8cf-89701d29ab5e\") " pod="openstack/ovn-controller-779x4-config-b56vn" Dec 11 02:21:11 crc kubenswrapper[4824]: I1211 02:21:11.477834 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ebda38df-902c-4ea4-a8cf-89701d29ab5e-scripts\") pod \"ovn-controller-779x4-config-b56vn\" (UID: \"ebda38df-902c-4ea4-a8cf-89701d29ab5e\") " pod="openstack/ovn-controller-779x4-config-b56vn" Dec 11 02:21:11 crc kubenswrapper[4824]: I1211 02:21:11.477867 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-494p9\" (UniqueName: \"kubernetes.io/projected/ebda38df-902c-4ea4-a8cf-89701d29ab5e-kube-api-access-494p9\") pod \"ovn-controller-779x4-config-b56vn\" (UID: \"ebda38df-902c-4ea4-a8cf-89701d29ab5e\") " pod="openstack/ovn-controller-779x4-config-b56vn" Dec 11 02:21:11 crc kubenswrapper[4824]: I1211 02:21:11.478001 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/ebda38df-902c-4ea4-a8cf-89701d29ab5e-var-log-ovn\") pod \"ovn-controller-779x4-config-b56vn\" (UID: \"ebda38df-902c-4ea4-a8cf-89701d29ab5e\") " pod="openstack/ovn-controller-779x4-config-b56vn" Dec 11 02:21:11 crc kubenswrapper[4824]: I1211 02:21:11.579457 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/ebda38df-902c-4ea4-a8cf-89701d29ab5e-var-run-ovn\") pod \"ovn-controller-779x4-config-b56vn\" (UID: \"ebda38df-902c-4ea4-a8cf-89701d29ab5e\") " pod="openstack/ovn-controller-779x4-config-b56vn" Dec 11 02:21:11 crc kubenswrapper[4824]: I1211 02:21:11.579524 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ebda38df-902c-4ea4-a8cf-89701d29ab5e-var-run\") pod \"ovn-controller-779x4-config-b56vn\" (UID: \"ebda38df-902c-4ea4-a8cf-89701d29ab5e\") " pod="openstack/ovn-controller-779x4-config-b56vn" Dec 11 02:21:11 crc kubenswrapper[4824]: I1211 02:21:11.579575 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ebda38df-902c-4ea4-a8cf-89701d29ab5e-scripts\") pod \"ovn-controller-779x4-config-b56vn\" (UID: \"ebda38df-902c-4ea4-a8cf-89701d29ab5e\") " pod="openstack/ovn-controller-779x4-config-b56vn" Dec 11 02:21:11 crc kubenswrapper[4824]: I1211 02:21:11.579823 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/ebda38df-902c-4ea4-a8cf-89701d29ab5e-var-run-ovn\") pod \"ovn-controller-779x4-config-b56vn\" (UID: \"ebda38df-902c-4ea4-a8cf-89701d29ab5e\") " pod="openstack/ovn-controller-779x4-config-b56vn" Dec 11 02:21:11 crc kubenswrapper[4824]: I1211 02:21:11.579842 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ebda38df-902c-4ea4-a8cf-89701d29ab5e-var-run\") pod \"ovn-controller-779x4-config-b56vn\" (UID: \"ebda38df-902c-4ea4-a8cf-89701d29ab5e\") " pod="openstack/ovn-controller-779x4-config-b56vn" Dec 11 02:21:11 crc kubenswrapper[4824]: I1211 02:21:11.579884 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-494p9\" (UniqueName: \"kubernetes.io/projected/ebda38df-902c-4ea4-a8cf-89701d29ab5e-kube-api-access-494p9\") pod \"ovn-controller-779x4-config-b56vn\" (UID: \"ebda38df-902c-4ea4-a8cf-89701d29ab5e\") " pod="openstack/ovn-controller-779x4-config-b56vn" Dec 11 02:21:11 crc kubenswrapper[4824]: I1211 02:21:11.579978 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/ebda38df-902c-4ea4-a8cf-89701d29ab5e-var-log-ovn\") pod \"ovn-controller-779x4-config-b56vn\" (UID: \"ebda38df-902c-4ea4-a8cf-89701d29ab5e\") " pod="openstack/ovn-controller-779x4-config-b56vn" Dec 11 02:21:11 crc kubenswrapper[4824]: I1211 02:21:11.580025 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/ebda38df-902c-4ea4-a8cf-89701d29ab5e-additional-scripts\") pod \"ovn-controller-779x4-config-b56vn\" (UID: \"ebda38df-902c-4ea4-a8cf-89701d29ab5e\") " pod="openstack/ovn-controller-779x4-config-b56vn" Dec 11 02:21:11 crc kubenswrapper[4824]: I1211 02:21:11.580095 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/ebda38df-902c-4ea4-a8cf-89701d29ab5e-var-log-ovn\") pod \"ovn-controller-779x4-config-b56vn\" (UID: \"ebda38df-902c-4ea4-a8cf-89701d29ab5e\") " pod="openstack/ovn-controller-779x4-config-b56vn" Dec 11 02:21:11 crc kubenswrapper[4824]: I1211 02:21:11.580958 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/ebda38df-902c-4ea4-a8cf-89701d29ab5e-additional-scripts\") pod \"ovn-controller-779x4-config-b56vn\" (UID: \"ebda38df-902c-4ea4-a8cf-89701d29ab5e\") " pod="openstack/ovn-controller-779x4-config-b56vn" Dec 11 02:21:11 crc kubenswrapper[4824]: I1211 02:21:11.583789 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ebda38df-902c-4ea4-a8cf-89701d29ab5e-scripts\") pod \"ovn-controller-779x4-config-b56vn\" (UID: \"ebda38df-902c-4ea4-a8cf-89701d29ab5e\") " pod="openstack/ovn-controller-779x4-config-b56vn" Dec 11 02:21:11 crc kubenswrapper[4824]: I1211 02:21:11.603266 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-494p9\" (UniqueName: \"kubernetes.io/projected/ebda38df-902c-4ea4-a8cf-89701d29ab5e-kube-api-access-494p9\") pod \"ovn-controller-779x4-config-b56vn\" (UID: \"ebda38df-902c-4ea4-a8cf-89701d29ab5e\") " pod="openstack/ovn-controller-779x4-config-b56vn" Dec 11 02:21:11 crc kubenswrapper[4824]: I1211 02:21:11.767919 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-779x4-config-b56vn" Dec 11 02:21:11 crc kubenswrapper[4824]: I1211 02:21:11.858068 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"7913c5de-48cb-4e9a-8b9b-ee976476bab2","Type":"ContainerStarted","Data":"ad211c4bcad8500815ba9efc5f404582abdcbfe968f09a1eedf2321fac5b5434"} Dec 11 02:21:11 crc kubenswrapper[4824]: I1211 02:21:11.858128 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"7913c5de-48cb-4e9a-8b9b-ee976476bab2","Type":"ContainerStarted","Data":"a95a70fb8df9000388c60c1672dfae108d762f30fa46d9d2c6565378ae89df2e"} Dec 11 02:21:11 crc kubenswrapper[4824]: I1211 02:21:11.858140 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"7913c5de-48cb-4e9a-8b9b-ee976476bab2","Type":"ContainerStarted","Data":"c97edd85281badf1eb6a45e98ee624e3e12de5a5025c0711148367313aab1d7d"} Dec 11 02:21:12 crc kubenswrapper[4824]: I1211 02:21:12.350760 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-779x4-config-b56vn"] Dec 11 02:21:12 crc kubenswrapper[4824]: W1211 02:21:12.356761 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podebda38df_902c_4ea4_a8cf_89701d29ab5e.slice/crio-f679d1e4b9a9cec9b5588a2dabde5fbf0bde8d30ecb9cc5c30edb475e67f862c WatchSource:0}: Error finding container f679d1e4b9a9cec9b5588a2dabde5fbf0bde8d30ecb9cc5c30edb475e67f862c: Status 404 returned error can't find the container with id f679d1e4b9a9cec9b5588a2dabde5fbf0bde8d30ecb9cc5c30edb475e67f862c Dec 11 02:21:12 crc kubenswrapper[4824]: I1211 02:21:12.652968 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="249d9815-d46f-4cf1-a8b9-f8831450b708" path="/var/lib/kubelet/pods/249d9815-d46f-4cf1-a8b9-f8831450b708/volumes" Dec 11 02:21:12 crc kubenswrapper[4824]: I1211 02:21:12.868220 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"7913c5de-48cb-4e9a-8b9b-ee976476bab2","Type":"ContainerStarted","Data":"6baf8b7f9dc6835b78679e08cf1a4ba415ee0489fda51226af43687d31d01d9a"} Dec 11 02:21:12 crc kubenswrapper[4824]: I1211 02:21:12.870545 4824 generic.go:334] "Generic (PLEG): container finished" podID="ebda38df-902c-4ea4-a8cf-89701d29ab5e" containerID="ed28097c0c3c423f890fb5bc1a668d40016ed48cdda8c3ac88e9a5c4ad01814d" exitCode=0 Dec 11 02:21:12 crc kubenswrapper[4824]: I1211 02:21:12.870600 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-779x4-config-b56vn" event={"ID":"ebda38df-902c-4ea4-a8cf-89701d29ab5e","Type":"ContainerDied","Data":"ed28097c0c3c423f890fb5bc1a668d40016ed48cdda8c3ac88e9a5c4ad01814d"} Dec 11 02:21:12 crc kubenswrapper[4824]: I1211 02:21:12.870623 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-779x4-config-b56vn" event={"ID":"ebda38df-902c-4ea4-a8cf-89701d29ab5e","Type":"ContainerStarted","Data":"f679d1e4b9a9cec9b5588a2dabde5fbf0bde8d30ecb9cc5c30edb475e67f862c"} Dec 11 02:21:13 crc kubenswrapper[4824]: I1211 02:21:13.882635 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"7913c5de-48cb-4e9a-8b9b-ee976476bab2","Type":"ContainerStarted","Data":"e4f448c5d0d1e1806a867878a545a571fa433f3b18f8bc3078d6823e85d66c69"} Dec 11 02:21:13 crc kubenswrapper[4824]: I1211 02:21:13.883425 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"7913c5de-48cb-4e9a-8b9b-ee976476bab2","Type":"ContainerStarted","Data":"161f763b94f9d2bfca77fb13ae32afaddf55bd785c32a59e3c5505625dbb5b27"} Dec 11 02:21:14 crc kubenswrapper[4824]: I1211 02:21:14.180057 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-779x4-config-b56vn" Dec 11 02:21:14 crc kubenswrapper[4824]: I1211 02:21:14.327569 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/ebda38df-902c-4ea4-a8cf-89701d29ab5e-additional-scripts\") pod \"ebda38df-902c-4ea4-a8cf-89701d29ab5e\" (UID: \"ebda38df-902c-4ea4-a8cf-89701d29ab5e\") " Dec 11 02:21:14 crc kubenswrapper[4824]: I1211 02:21:14.327738 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/ebda38df-902c-4ea4-a8cf-89701d29ab5e-var-log-ovn\") pod \"ebda38df-902c-4ea4-a8cf-89701d29ab5e\" (UID: \"ebda38df-902c-4ea4-a8cf-89701d29ab5e\") " Dec 11 02:21:14 crc kubenswrapper[4824]: I1211 02:21:14.327782 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/ebda38df-902c-4ea4-a8cf-89701d29ab5e-var-run-ovn\") pod \"ebda38df-902c-4ea4-a8cf-89701d29ab5e\" (UID: \"ebda38df-902c-4ea4-a8cf-89701d29ab5e\") " Dec 11 02:21:14 crc kubenswrapper[4824]: I1211 02:21:14.327833 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ebda38df-902c-4ea4-a8cf-89701d29ab5e-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "ebda38df-902c-4ea4-a8cf-89701d29ab5e" (UID: "ebda38df-902c-4ea4-a8cf-89701d29ab5e"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 02:21:14 crc kubenswrapper[4824]: I1211 02:21:14.327897 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ebda38df-902c-4ea4-a8cf-89701d29ab5e-scripts\") pod \"ebda38df-902c-4ea4-a8cf-89701d29ab5e\" (UID: \"ebda38df-902c-4ea4-a8cf-89701d29ab5e\") " Dec 11 02:21:14 crc kubenswrapper[4824]: I1211 02:21:14.327909 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ebda38df-902c-4ea4-a8cf-89701d29ab5e-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "ebda38df-902c-4ea4-a8cf-89701d29ab5e" (UID: "ebda38df-902c-4ea4-a8cf-89701d29ab5e"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 02:21:14 crc kubenswrapper[4824]: I1211 02:21:14.327933 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-494p9\" (UniqueName: \"kubernetes.io/projected/ebda38df-902c-4ea4-a8cf-89701d29ab5e-kube-api-access-494p9\") pod \"ebda38df-902c-4ea4-a8cf-89701d29ab5e\" (UID: \"ebda38df-902c-4ea4-a8cf-89701d29ab5e\") " Dec 11 02:21:14 crc kubenswrapper[4824]: I1211 02:21:14.327996 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ebda38df-902c-4ea4-a8cf-89701d29ab5e-var-run\") pod \"ebda38df-902c-4ea4-a8cf-89701d29ab5e\" (UID: \"ebda38df-902c-4ea4-a8cf-89701d29ab5e\") " Dec 11 02:21:14 crc kubenswrapper[4824]: I1211 02:21:14.328141 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ebda38df-902c-4ea4-a8cf-89701d29ab5e-var-run" (OuterVolumeSpecName: "var-run") pod "ebda38df-902c-4ea4-a8cf-89701d29ab5e" (UID: "ebda38df-902c-4ea4-a8cf-89701d29ab5e"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 02:21:14 crc kubenswrapper[4824]: I1211 02:21:14.328445 4824 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/ebda38df-902c-4ea4-a8cf-89701d29ab5e-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 11 02:21:14 crc kubenswrapper[4824]: I1211 02:21:14.328472 4824 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/ebda38df-902c-4ea4-a8cf-89701d29ab5e-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 11 02:21:14 crc kubenswrapper[4824]: I1211 02:21:14.328487 4824 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ebda38df-902c-4ea4-a8cf-89701d29ab5e-var-run\") on node \"crc\" DevicePath \"\"" Dec 11 02:21:14 crc kubenswrapper[4824]: I1211 02:21:14.330850 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ebda38df-902c-4ea4-a8cf-89701d29ab5e-scripts" (OuterVolumeSpecName: "scripts") pod "ebda38df-902c-4ea4-a8cf-89701d29ab5e" (UID: "ebda38df-902c-4ea4-a8cf-89701d29ab5e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:21:14 crc kubenswrapper[4824]: I1211 02:21:14.331871 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ebda38df-902c-4ea4-a8cf-89701d29ab5e-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "ebda38df-902c-4ea4-a8cf-89701d29ab5e" (UID: "ebda38df-902c-4ea4-a8cf-89701d29ab5e"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:21:14 crc kubenswrapper[4824]: I1211 02:21:14.336439 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ebda38df-902c-4ea4-a8cf-89701d29ab5e-kube-api-access-494p9" (OuterVolumeSpecName: "kube-api-access-494p9") pod "ebda38df-902c-4ea4-a8cf-89701d29ab5e" (UID: "ebda38df-902c-4ea4-a8cf-89701d29ab5e"). InnerVolumeSpecName "kube-api-access-494p9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:21:14 crc kubenswrapper[4824]: I1211 02:21:14.429362 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ebda38df-902c-4ea4-a8cf-89701d29ab5e-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 02:21:14 crc kubenswrapper[4824]: I1211 02:21:14.429395 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-494p9\" (UniqueName: \"kubernetes.io/projected/ebda38df-902c-4ea4-a8cf-89701d29ab5e-kube-api-access-494p9\") on node \"crc\" DevicePath \"\"" Dec 11 02:21:14 crc kubenswrapper[4824]: I1211 02:21:14.429407 4824 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/ebda38df-902c-4ea4-a8cf-89701d29ab5e-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 02:21:14 crc kubenswrapper[4824]: I1211 02:21:14.893543 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-779x4-config-b56vn" Dec 11 02:21:14 crc kubenswrapper[4824]: I1211 02:21:14.893539 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-779x4-config-b56vn" event={"ID":"ebda38df-902c-4ea4-a8cf-89701d29ab5e","Type":"ContainerDied","Data":"f679d1e4b9a9cec9b5588a2dabde5fbf0bde8d30ecb9cc5c30edb475e67f862c"} Dec 11 02:21:14 crc kubenswrapper[4824]: I1211 02:21:14.894074 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f679d1e4b9a9cec9b5588a2dabde5fbf0bde8d30ecb9cc5c30edb475e67f862c" Dec 11 02:21:14 crc kubenswrapper[4824]: I1211 02:21:14.899392 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"7913c5de-48cb-4e9a-8b9b-ee976476bab2","Type":"ContainerStarted","Data":"db15893396e5aa5fe6ef8d3d848a6f95d05faa88bdcacddd7809a3504a18e45b"} Dec 11 02:21:14 crc kubenswrapper[4824]: I1211 02:21:14.899447 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"7913c5de-48cb-4e9a-8b9b-ee976476bab2","Type":"ContainerStarted","Data":"cb92e67bc4e93e0a27a01325219fd2d6361fb9a1f0ecb352e7972f6119993cea"} Dec 11 02:21:15 crc kubenswrapper[4824]: I1211 02:21:15.255296 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-779x4-config-b56vn"] Dec 11 02:21:15 crc kubenswrapper[4824]: I1211 02:21:15.267240 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-779x4-config-b56vn"] Dec 11 02:21:15 crc kubenswrapper[4824]: I1211 02:21:15.913751 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"7913c5de-48cb-4e9a-8b9b-ee976476bab2","Type":"ContainerStarted","Data":"9e4204dc522f2de9e3c24a93866b60be0e72b5ebff9a15878d928ee69a24a933"} Dec 11 02:21:16 crc kubenswrapper[4824]: I1211 02:21:16.691262 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ebda38df-902c-4ea4-a8cf-89701d29ab5e" path="/var/lib/kubelet/pods/ebda38df-902c-4ea4-a8cf-89701d29ab5e/volumes" Dec 11 02:21:16 crc kubenswrapper[4824]: I1211 02:21:16.926242 4824 generic.go:334] "Generic (PLEG): container finished" podID="3b4a47f6-97f5-4395-b82a-166b934174b5" containerID="13a2e99f97a6fd8fd4b48427476f27d577b17501906194f19debf85ca7f2b4c4" exitCode=0 Dec 11 02:21:16 crc kubenswrapper[4824]: I1211 02:21:16.926329 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-4ctff" event={"ID":"3b4a47f6-97f5-4395-b82a-166b934174b5","Type":"ContainerDied","Data":"13a2e99f97a6fd8fd4b48427476f27d577b17501906194f19debf85ca7f2b4c4"} Dec 11 02:21:16 crc kubenswrapper[4824]: I1211 02:21:16.937398 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"7913c5de-48cb-4e9a-8b9b-ee976476bab2","Type":"ContainerStarted","Data":"dbd16c30a4f97b1cabfc7897fe960566d3aaf6adbbdf3c0804655548f2556740"} Dec 11 02:21:16 crc kubenswrapper[4824]: I1211 02:21:16.937440 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"7913c5de-48cb-4e9a-8b9b-ee976476bab2","Type":"ContainerStarted","Data":"ecfcac30ed4a8b70b031b758ec738ddbb5c0564ff92ac90e31c609fa16bdb691"} Dec 11 02:21:16 crc kubenswrapper[4824]: I1211 02:21:16.937452 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"7913c5de-48cb-4e9a-8b9b-ee976476bab2","Type":"ContainerStarted","Data":"b51e47deda37d80234c3589ac44e991a76ecf405b4efd89b18d60e2e8218369e"} Dec 11 02:21:16 crc kubenswrapper[4824]: I1211 02:21:16.937461 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"7913c5de-48cb-4e9a-8b9b-ee976476bab2","Type":"ContainerStarted","Data":"165ff871e9a18aa54a5ac5e67663b045c37f7e5c92cc73b6c4a440ad02312d30"} Dec 11 02:21:17 crc kubenswrapper[4824]: I1211 02:21:17.966392 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"7913c5de-48cb-4e9a-8b9b-ee976476bab2","Type":"ContainerStarted","Data":"ae67fe9b18f6a6bdc609b5644c0c654d8be5cb3f13743ef7d035721b57503d79"} Dec 11 02:21:18 crc kubenswrapper[4824]: I1211 02:21:18.426635 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-4ctff" Dec 11 02:21:18 crc kubenswrapper[4824]: I1211 02:21:18.604802 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3b4a47f6-97f5-4395-b82a-166b934174b5-db-sync-config-data\") pod \"3b4a47f6-97f5-4395-b82a-166b934174b5\" (UID: \"3b4a47f6-97f5-4395-b82a-166b934174b5\") " Dec 11 02:21:18 crc kubenswrapper[4824]: I1211 02:21:18.604995 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b4a47f6-97f5-4395-b82a-166b934174b5-combined-ca-bundle\") pod \"3b4a47f6-97f5-4395-b82a-166b934174b5\" (UID: \"3b4a47f6-97f5-4395-b82a-166b934174b5\") " Dec 11 02:21:18 crc kubenswrapper[4824]: I1211 02:21:18.605048 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cxxvl\" (UniqueName: \"kubernetes.io/projected/3b4a47f6-97f5-4395-b82a-166b934174b5-kube-api-access-cxxvl\") pod \"3b4a47f6-97f5-4395-b82a-166b934174b5\" (UID: \"3b4a47f6-97f5-4395-b82a-166b934174b5\") " Dec 11 02:21:18 crc kubenswrapper[4824]: I1211 02:21:18.605152 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b4a47f6-97f5-4395-b82a-166b934174b5-config-data\") pod \"3b4a47f6-97f5-4395-b82a-166b934174b5\" (UID: \"3b4a47f6-97f5-4395-b82a-166b934174b5\") " Dec 11 02:21:18 crc kubenswrapper[4824]: I1211 02:21:18.613396 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b4a47f6-97f5-4395-b82a-166b934174b5-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "3b4a47f6-97f5-4395-b82a-166b934174b5" (UID: "3b4a47f6-97f5-4395-b82a-166b934174b5"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:21:18 crc kubenswrapper[4824]: I1211 02:21:18.614182 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b4a47f6-97f5-4395-b82a-166b934174b5-kube-api-access-cxxvl" (OuterVolumeSpecName: "kube-api-access-cxxvl") pod "3b4a47f6-97f5-4395-b82a-166b934174b5" (UID: "3b4a47f6-97f5-4395-b82a-166b934174b5"). InnerVolumeSpecName "kube-api-access-cxxvl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:21:18 crc kubenswrapper[4824]: I1211 02:21:18.644328 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b4a47f6-97f5-4395-b82a-166b934174b5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3b4a47f6-97f5-4395-b82a-166b934174b5" (UID: "3b4a47f6-97f5-4395-b82a-166b934174b5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:21:18 crc kubenswrapper[4824]: I1211 02:21:18.684053 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b4a47f6-97f5-4395-b82a-166b934174b5-config-data" (OuterVolumeSpecName: "config-data") pod "3b4a47f6-97f5-4395-b82a-166b934174b5" (UID: "3b4a47f6-97f5-4395-b82a-166b934174b5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:21:18 crc kubenswrapper[4824]: I1211 02:21:18.707481 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b4a47f6-97f5-4395-b82a-166b934174b5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 02:21:18 crc kubenswrapper[4824]: I1211 02:21:18.707521 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cxxvl\" (UniqueName: \"kubernetes.io/projected/3b4a47f6-97f5-4395-b82a-166b934174b5-kube-api-access-cxxvl\") on node \"crc\" DevicePath \"\"" Dec 11 02:21:18 crc kubenswrapper[4824]: I1211 02:21:18.707535 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b4a47f6-97f5-4395-b82a-166b934174b5-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 02:21:18 crc kubenswrapper[4824]: I1211 02:21:18.707546 4824 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3b4a47f6-97f5-4395-b82a-166b934174b5-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 02:21:18 crc kubenswrapper[4824]: I1211 02:21:18.984228 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-4ctff" event={"ID":"3b4a47f6-97f5-4395-b82a-166b934174b5","Type":"ContainerDied","Data":"38a31d452ee585811a8351977fa04871cf57dabd2fa50c5688836ba6f3cfdf95"} Dec 11 02:21:18 crc kubenswrapper[4824]: I1211 02:21:18.984288 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="38a31d452ee585811a8351977fa04871cf57dabd2fa50c5688836ba6f3cfdf95" Dec 11 02:21:18 crc kubenswrapper[4824]: I1211 02:21:18.984336 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-4ctff" Dec 11 02:21:19 crc kubenswrapper[4824]: I1211 02:21:19.463665 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-6b4w8"] Dec 11 02:21:19 crc kubenswrapper[4824]: E1211 02:21:19.464436 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebda38df-902c-4ea4-a8cf-89701d29ab5e" containerName="ovn-config" Dec 11 02:21:19 crc kubenswrapper[4824]: I1211 02:21:19.464470 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebda38df-902c-4ea4-a8cf-89701d29ab5e" containerName="ovn-config" Dec 11 02:21:19 crc kubenswrapper[4824]: E1211 02:21:19.464493 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b4a47f6-97f5-4395-b82a-166b934174b5" containerName="glance-db-sync" Dec 11 02:21:19 crc kubenswrapper[4824]: I1211 02:21:19.464500 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b4a47f6-97f5-4395-b82a-166b934174b5" containerName="glance-db-sync" Dec 11 02:21:19 crc kubenswrapper[4824]: I1211 02:21:19.464674 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b4a47f6-97f5-4395-b82a-166b934174b5" containerName="glance-db-sync" Dec 11 02:21:19 crc kubenswrapper[4824]: I1211 02:21:19.464705 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebda38df-902c-4ea4-a8cf-89701d29ab5e" containerName="ovn-config" Dec 11 02:21:19 crc kubenswrapper[4824]: I1211 02:21:19.465646 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b946c75cc-6b4w8" Dec 11 02:21:19 crc kubenswrapper[4824]: I1211 02:21:19.522098 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-6b4w8"] Dec 11 02:21:19 crc kubenswrapper[4824]: I1211 02:21:19.621914 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f30f409e-2171-4b62-8e1b-e72ea450ec8d-ovsdbserver-nb\") pod \"dnsmasq-dns-5b946c75cc-6b4w8\" (UID: \"f30f409e-2171-4b62-8e1b-e72ea450ec8d\") " pod="openstack/dnsmasq-dns-5b946c75cc-6b4w8" Dec 11 02:21:19 crc kubenswrapper[4824]: I1211 02:21:19.622245 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f30f409e-2171-4b62-8e1b-e72ea450ec8d-dns-svc\") pod \"dnsmasq-dns-5b946c75cc-6b4w8\" (UID: \"f30f409e-2171-4b62-8e1b-e72ea450ec8d\") " pod="openstack/dnsmasq-dns-5b946c75cc-6b4w8" Dec 11 02:21:19 crc kubenswrapper[4824]: I1211 02:21:19.622370 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f30f409e-2171-4b62-8e1b-e72ea450ec8d-config\") pod \"dnsmasq-dns-5b946c75cc-6b4w8\" (UID: \"f30f409e-2171-4b62-8e1b-e72ea450ec8d\") " pod="openstack/dnsmasq-dns-5b946c75cc-6b4w8" Dec 11 02:21:19 crc kubenswrapper[4824]: I1211 02:21:19.622588 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f30f409e-2171-4b62-8e1b-e72ea450ec8d-ovsdbserver-sb\") pod \"dnsmasq-dns-5b946c75cc-6b4w8\" (UID: \"f30f409e-2171-4b62-8e1b-e72ea450ec8d\") " pod="openstack/dnsmasq-dns-5b946c75cc-6b4w8" Dec 11 02:21:19 crc kubenswrapper[4824]: I1211 02:21:19.622710 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-drwxg\" (UniqueName: \"kubernetes.io/projected/f30f409e-2171-4b62-8e1b-e72ea450ec8d-kube-api-access-drwxg\") pod \"dnsmasq-dns-5b946c75cc-6b4w8\" (UID: \"f30f409e-2171-4b62-8e1b-e72ea450ec8d\") " pod="openstack/dnsmasq-dns-5b946c75cc-6b4w8" Dec 11 02:21:19 crc kubenswrapper[4824]: I1211 02:21:19.724219 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f30f409e-2171-4b62-8e1b-e72ea450ec8d-config\") pod \"dnsmasq-dns-5b946c75cc-6b4w8\" (UID: \"f30f409e-2171-4b62-8e1b-e72ea450ec8d\") " pod="openstack/dnsmasq-dns-5b946c75cc-6b4w8" Dec 11 02:21:19 crc kubenswrapper[4824]: I1211 02:21:19.724352 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f30f409e-2171-4b62-8e1b-e72ea450ec8d-ovsdbserver-sb\") pod \"dnsmasq-dns-5b946c75cc-6b4w8\" (UID: \"f30f409e-2171-4b62-8e1b-e72ea450ec8d\") " pod="openstack/dnsmasq-dns-5b946c75cc-6b4w8" Dec 11 02:21:19 crc kubenswrapper[4824]: I1211 02:21:19.724370 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-drwxg\" (UniqueName: \"kubernetes.io/projected/f30f409e-2171-4b62-8e1b-e72ea450ec8d-kube-api-access-drwxg\") pod \"dnsmasq-dns-5b946c75cc-6b4w8\" (UID: \"f30f409e-2171-4b62-8e1b-e72ea450ec8d\") " pod="openstack/dnsmasq-dns-5b946c75cc-6b4w8" Dec 11 02:21:19 crc kubenswrapper[4824]: I1211 02:21:19.724407 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f30f409e-2171-4b62-8e1b-e72ea450ec8d-ovsdbserver-nb\") pod \"dnsmasq-dns-5b946c75cc-6b4w8\" (UID: \"f30f409e-2171-4b62-8e1b-e72ea450ec8d\") " pod="openstack/dnsmasq-dns-5b946c75cc-6b4w8" Dec 11 02:21:19 crc kubenswrapper[4824]: I1211 02:21:19.724430 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f30f409e-2171-4b62-8e1b-e72ea450ec8d-dns-svc\") pod \"dnsmasq-dns-5b946c75cc-6b4w8\" (UID: \"f30f409e-2171-4b62-8e1b-e72ea450ec8d\") " pod="openstack/dnsmasq-dns-5b946c75cc-6b4w8" Dec 11 02:21:19 crc kubenswrapper[4824]: I1211 02:21:19.725432 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f30f409e-2171-4b62-8e1b-e72ea450ec8d-dns-svc\") pod \"dnsmasq-dns-5b946c75cc-6b4w8\" (UID: \"f30f409e-2171-4b62-8e1b-e72ea450ec8d\") " pod="openstack/dnsmasq-dns-5b946c75cc-6b4w8" Dec 11 02:21:19 crc kubenswrapper[4824]: I1211 02:21:19.725447 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f30f409e-2171-4b62-8e1b-e72ea450ec8d-ovsdbserver-sb\") pod \"dnsmasq-dns-5b946c75cc-6b4w8\" (UID: \"f30f409e-2171-4b62-8e1b-e72ea450ec8d\") " pod="openstack/dnsmasq-dns-5b946c75cc-6b4w8" Dec 11 02:21:19 crc kubenswrapper[4824]: I1211 02:21:19.725458 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f30f409e-2171-4b62-8e1b-e72ea450ec8d-config\") pod \"dnsmasq-dns-5b946c75cc-6b4w8\" (UID: \"f30f409e-2171-4b62-8e1b-e72ea450ec8d\") " pod="openstack/dnsmasq-dns-5b946c75cc-6b4w8" Dec 11 02:21:19 crc kubenswrapper[4824]: I1211 02:21:19.725959 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f30f409e-2171-4b62-8e1b-e72ea450ec8d-ovsdbserver-nb\") pod \"dnsmasq-dns-5b946c75cc-6b4w8\" (UID: \"f30f409e-2171-4b62-8e1b-e72ea450ec8d\") " pod="openstack/dnsmasq-dns-5b946c75cc-6b4w8" Dec 11 02:21:19 crc kubenswrapper[4824]: I1211 02:21:19.741390 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-drwxg\" (UniqueName: \"kubernetes.io/projected/f30f409e-2171-4b62-8e1b-e72ea450ec8d-kube-api-access-drwxg\") pod \"dnsmasq-dns-5b946c75cc-6b4w8\" (UID: \"f30f409e-2171-4b62-8e1b-e72ea450ec8d\") " pod="openstack/dnsmasq-dns-5b946c75cc-6b4w8" Dec 11 02:21:19 crc kubenswrapper[4824]: I1211 02:21:19.820351 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b946c75cc-6b4w8" Dec 11 02:21:20 crc kubenswrapper[4824]: I1211 02:21:20.239074 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-6b4w8"] Dec 11 02:21:20 crc kubenswrapper[4824]: I1211 02:21:20.748295 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 11 02:21:20 crc kubenswrapper[4824]: I1211 02:21:20.823417 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.048849 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"7913c5de-48cb-4e9a-8b9b-ee976476bab2","Type":"ContainerStarted","Data":"cf88ebda790c92b939ed996a950e05c2b86faefb6354dc06b2e2bcb188c8b8a0"} Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.053461 4824 generic.go:334] "Generic (PLEG): container finished" podID="f30f409e-2171-4b62-8e1b-e72ea450ec8d" containerID="dc4ae37edeb4ed98398c339380c86390f73246b108db79995cc4b50af1dce39e" exitCode=0 Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.053546 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b946c75cc-6b4w8" event={"ID":"f30f409e-2171-4b62-8e1b-e72ea450ec8d","Type":"ContainerDied","Data":"dc4ae37edeb4ed98398c339380c86390f73246b108db79995cc4b50af1dce39e"} Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.053576 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b946c75cc-6b4w8" event={"ID":"f30f409e-2171-4b62-8e1b-e72ea450ec8d","Type":"ContainerStarted","Data":"3424e309b384bb43f44fb7cfe36eefd8aba2084b1ec5aa95c1fc953ebd648626"} Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.114176 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=39.406242415 podStartE2EDuration="45.114160235s" podCreationTimestamp="2025-12-11 02:20:36 +0000 UTC" firstStartedPulling="2025-12-11 02:21:09.82032497 +0000 UTC m=+1211.509362349" lastFinishedPulling="2025-12-11 02:21:15.52824275 +0000 UTC m=+1217.217280169" observedRunningTime="2025-12-11 02:21:21.112041317 +0000 UTC m=+1222.801078716" watchObservedRunningTime="2025-12-11 02:21:21.114160235 +0000 UTC m=+1222.803197614" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.207485 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-qwpmm"] Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.208999 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-qwpmm" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.217337 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-qwpmm"] Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.255290 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-bce9-account-create-update-qnmss"] Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.256725 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-bce9-account-create-update-qnmss" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.263560 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.264687 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5dgh9\" (UniqueName: \"kubernetes.io/projected/20766632-27ff-42ab-9b3c-28aa8b8912bf-kube-api-access-5dgh9\") pod \"barbican-db-create-qwpmm\" (UID: \"20766632-27ff-42ab-9b3c-28aa8b8912bf\") " pod="openstack/barbican-db-create-qwpmm" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.264864 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20766632-27ff-42ab-9b3c-28aa8b8912bf-operator-scripts\") pod \"barbican-db-create-qwpmm\" (UID: \"20766632-27ff-42ab-9b3c-28aa8b8912bf\") " pod="openstack/barbican-db-create-qwpmm" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.278827 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-bce9-account-create-update-qnmss"] Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.323723 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-5a29-account-create-update-sb4qh"] Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.324926 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-5a29-account-create-update-sb4qh" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.327217 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.329413 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-xg279"] Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.330388 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-xg279" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.336544 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-5a29-account-create-update-sb4qh"] Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.347589 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-xg279"] Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.366129 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55qdq\" (UniqueName: \"kubernetes.io/projected/9d9a4a8f-d58e-4a90-9c33-3ffa6a7f5c36-kube-api-access-55qdq\") pod \"barbican-5a29-account-create-update-sb4qh\" (UID: \"9d9a4a8f-d58e-4a90-9c33-3ffa6a7f5c36\") " pod="openstack/barbican-5a29-account-create-update-sb4qh" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.366423 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9d9a4a8f-d58e-4a90-9c33-3ffa6a7f5c36-operator-scripts\") pod \"barbican-5a29-account-create-update-sb4qh\" (UID: \"9d9a4a8f-d58e-4a90-9c33-3ffa6a7f5c36\") " pod="openstack/barbican-5a29-account-create-update-sb4qh" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.366477 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/57686bda-9117-48cc-b29d-f5ff374937c9-operator-scripts\") pod \"cinder-db-create-xg279\" (UID: \"57686bda-9117-48cc-b29d-f5ff374937c9\") " pod="openstack/cinder-db-create-xg279" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.366522 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20766632-27ff-42ab-9b3c-28aa8b8912bf-operator-scripts\") pod \"barbican-db-create-qwpmm\" (UID: \"20766632-27ff-42ab-9b3c-28aa8b8912bf\") " pod="openstack/barbican-db-create-qwpmm" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.366540 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fd7g5\" (UniqueName: \"kubernetes.io/projected/793e155e-73e9-46d5-9de6-a4aebf3ca2b4-kube-api-access-fd7g5\") pod \"cinder-bce9-account-create-update-qnmss\" (UID: \"793e155e-73e9-46d5-9de6-a4aebf3ca2b4\") " pod="openstack/cinder-bce9-account-create-update-qnmss" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.366569 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/793e155e-73e9-46d5-9de6-a4aebf3ca2b4-operator-scripts\") pod \"cinder-bce9-account-create-update-qnmss\" (UID: \"793e155e-73e9-46d5-9de6-a4aebf3ca2b4\") " pod="openstack/cinder-bce9-account-create-update-qnmss" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.366588 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5dgh9\" (UniqueName: \"kubernetes.io/projected/20766632-27ff-42ab-9b3c-28aa8b8912bf-kube-api-access-5dgh9\") pod \"barbican-db-create-qwpmm\" (UID: \"20766632-27ff-42ab-9b3c-28aa8b8912bf\") " pod="openstack/barbican-db-create-qwpmm" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.366608 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2ndh\" (UniqueName: \"kubernetes.io/projected/57686bda-9117-48cc-b29d-f5ff374937c9-kube-api-access-c2ndh\") pod \"cinder-db-create-xg279\" (UID: \"57686bda-9117-48cc-b29d-f5ff374937c9\") " pod="openstack/cinder-db-create-xg279" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.367310 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20766632-27ff-42ab-9b3c-28aa8b8912bf-operator-scripts\") pod \"barbican-db-create-qwpmm\" (UID: \"20766632-27ff-42ab-9b3c-28aa8b8912bf\") " pod="openstack/barbican-db-create-qwpmm" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.391994 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-wdzl5"] Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.400345 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-wdzl5" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.400970 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5dgh9\" (UniqueName: \"kubernetes.io/projected/20766632-27ff-42ab-9b3c-28aa8b8912bf-kube-api-access-5dgh9\") pod \"barbican-db-create-qwpmm\" (UID: \"20766632-27ff-42ab-9b3c-28aa8b8912bf\") " pod="openstack/barbican-db-create-qwpmm" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.405975 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.406258 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.406260 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.406364 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-8rv79" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.434208 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-wdzl5"] Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.468083 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55qdq\" (UniqueName: \"kubernetes.io/projected/9d9a4a8f-d58e-4a90-9c33-3ffa6a7f5c36-kube-api-access-55qdq\") pod \"barbican-5a29-account-create-update-sb4qh\" (UID: \"9d9a4a8f-d58e-4a90-9c33-3ffa6a7f5c36\") " pod="openstack/barbican-5a29-account-create-update-sb4qh" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.468144 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9d9a4a8f-d58e-4a90-9c33-3ffa6a7f5c36-operator-scripts\") pod \"barbican-5a29-account-create-update-sb4qh\" (UID: \"9d9a4a8f-d58e-4a90-9c33-3ffa6a7f5c36\") " pod="openstack/barbican-5a29-account-create-update-sb4qh" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.468203 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/57686bda-9117-48cc-b29d-f5ff374937c9-operator-scripts\") pod \"cinder-db-create-xg279\" (UID: \"57686bda-9117-48cc-b29d-f5ff374937c9\") " pod="openstack/cinder-db-create-xg279" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.468243 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/922cc36d-5fd8-4cba-b6e3-5b16e6cb7531-config-data\") pod \"keystone-db-sync-wdzl5\" (UID: \"922cc36d-5fd8-4cba-b6e3-5b16e6cb7531\") " pod="openstack/keystone-db-sync-wdzl5" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.468268 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fd7g5\" (UniqueName: \"kubernetes.io/projected/793e155e-73e9-46d5-9de6-a4aebf3ca2b4-kube-api-access-fd7g5\") pod \"cinder-bce9-account-create-update-qnmss\" (UID: \"793e155e-73e9-46d5-9de6-a4aebf3ca2b4\") " pod="openstack/cinder-bce9-account-create-update-qnmss" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.468285 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/922cc36d-5fd8-4cba-b6e3-5b16e6cb7531-combined-ca-bundle\") pod \"keystone-db-sync-wdzl5\" (UID: \"922cc36d-5fd8-4cba-b6e3-5b16e6cb7531\") " pod="openstack/keystone-db-sync-wdzl5" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.468317 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/793e155e-73e9-46d5-9de6-a4aebf3ca2b4-operator-scripts\") pod \"cinder-bce9-account-create-update-qnmss\" (UID: \"793e155e-73e9-46d5-9de6-a4aebf3ca2b4\") " pod="openstack/cinder-bce9-account-create-update-qnmss" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.468334 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d84f5\" (UniqueName: \"kubernetes.io/projected/922cc36d-5fd8-4cba-b6e3-5b16e6cb7531-kube-api-access-d84f5\") pod \"keystone-db-sync-wdzl5\" (UID: \"922cc36d-5fd8-4cba-b6e3-5b16e6cb7531\") " pod="openstack/keystone-db-sync-wdzl5" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.468360 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2ndh\" (UniqueName: \"kubernetes.io/projected/57686bda-9117-48cc-b29d-f5ff374937c9-kube-api-access-c2ndh\") pod \"cinder-db-create-xg279\" (UID: \"57686bda-9117-48cc-b29d-f5ff374937c9\") " pod="openstack/cinder-db-create-xg279" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.469426 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9d9a4a8f-d58e-4a90-9c33-3ffa6a7f5c36-operator-scripts\") pod \"barbican-5a29-account-create-update-sb4qh\" (UID: \"9d9a4a8f-d58e-4a90-9c33-3ffa6a7f5c36\") " pod="openstack/barbican-5a29-account-create-update-sb4qh" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.469993 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/57686bda-9117-48cc-b29d-f5ff374937c9-operator-scripts\") pod \"cinder-db-create-xg279\" (UID: \"57686bda-9117-48cc-b29d-f5ff374937c9\") " pod="openstack/cinder-db-create-xg279" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.473633 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/793e155e-73e9-46d5-9de6-a4aebf3ca2b4-operator-scripts\") pod \"cinder-bce9-account-create-update-qnmss\" (UID: \"793e155e-73e9-46d5-9de6-a4aebf3ca2b4\") " pod="openstack/cinder-bce9-account-create-update-qnmss" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.485956 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fd7g5\" (UniqueName: \"kubernetes.io/projected/793e155e-73e9-46d5-9de6-a4aebf3ca2b4-kube-api-access-fd7g5\") pod \"cinder-bce9-account-create-update-qnmss\" (UID: \"793e155e-73e9-46d5-9de6-a4aebf3ca2b4\") " pod="openstack/cinder-bce9-account-create-update-qnmss" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.488431 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2ndh\" (UniqueName: \"kubernetes.io/projected/57686bda-9117-48cc-b29d-f5ff374937c9-kube-api-access-c2ndh\") pod \"cinder-db-create-xg279\" (UID: \"57686bda-9117-48cc-b29d-f5ff374937c9\") " pod="openstack/cinder-db-create-xg279" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.493040 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55qdq\" (UniqueName: \"kubernetes.io/projected/9d9a4a8f-d58e-4a90-9c33-3ffa6a7f5c36-kube-api-access-55qdq\") pod \"barbican-5a29-account-create-update-sb4qh\" (UID: \"9d9a4a8f-d58e-4a90-9c33-3ffa6a7f5c36\") " pod="openstack/barbican-5a29-account-create-update-sb4qh" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.501857 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-6b4w8"] Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.541924 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-qwpmm" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.548387 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-9hwb8"] Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.561258 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-9hwb8" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.575861 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/922cc36d-5fd8-4cba-b6e3-5b16e6cb7531-config-data\") pod \"keystone-db-sync-wdzl5\" (UID: \"922cc36d-5fd8-4cba-b6e3-5b16e6cb7531\") " pod="openstack/keystone-db-sync-wdzl5" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.575950 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/922cc36d-5fd8-4cba-b6e3-5b16e6cb7531-combined-ca-bundle\") pod \"keystone-db-sync-wdzl5\" (UID: \"922cc36d-5fd8-4cba-b6e3-5b16e6cb7531\") " pod="openstack/keystone-db-sync-wdzl5" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.576033 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d84f5\" (UniqueName: \"kubernetes.io/projected/922cc36d-5fd8-4cba-b6e3-5b16e6cb7531-kube-api-access-d84f5\") pod \"keystone-db-sync-wdzl5\" (UID: \"922cc36d-5fd8-4cba-b6e3-5b16e6cb7531\") " pod="openstack/keystone-db-sync-wdzl5" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.577179 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-bce9-account-create-update-qnmss" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.582826 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/922cc36d-5fd8-4cba-b6e3-5b16e6cb7531-combined-ca-bundle\") pod \"keystone-db-sync-wdzl5\" (UID: \"922cc36d-5fd8-4cba-b6e3-5b16e6cb7531\") " pod="openstack/keystone-db-sync-wdzl5" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.590905 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/922cc36d-5fd8-4cba-b6e3-5b16e6cb7531-config-data\") pod \"keystone-db-sync-wdzl5\" (UID: \"922cc36d-5fd8-4cba-b6e3-5b16e6cb7531\") " pod="openstack/keystone-db-sync-wdzl5" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.591019 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-9hwb8"] Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.614687 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d84f5\" (UniqueName: \"kubernetes.io/projected/922cc36d-5fd8-4cba-b6e3-5b16e6cb7531-kube-api-access-d84f5\") pod \"keystone-db-sync-wdzl5\" (UID: \"922cc36d-5fd8-4cba-b6e3-5b16e6cb7531\") " pod="openstack/keystone-db-sync-wdzl5" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.628521 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-4tss6"] Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.630095 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-4tss6" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.633914 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.640248 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-5a29-account-create-update-sb4qh" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.642190 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-4tss6"] Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.652504 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-xg279" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.679225 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j7gqg\" (UniqueName: \"kubernetes.io/projected/2bcb322d-a528-4d31-b639-367f5f7f3dd1-kube-api-access-j7gqg\") pod \"neutron-db-create-9hwb8\" (UID: \"2bcb322d-a528-4d31-b639-367f5f7f3dd1\") " pod="openstack/neutron-db-create-9hwb8" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.679361 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2bcb322d-a528-4d31-b639-367f5f7f3dd1-operator-scripts\") pod \"neutron-db-create-9hwb8\" (UID: \"2bcb322d-a528-4d31-b639-367f5f7f3dd1\") " pod="openstack/neutron-db-create-9hwb8" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.691763 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-70b2-account-create-update-fzkfv"] Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.693045 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-70b2-account-create-update-fzkfv" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.695144 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.698636 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-70b2-account-create-update-fzkfv"] Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.753693 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-wdzl5" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.780599 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ab730c6-95e4-438f-90d5-04812d26384f-operator-scripts\") pod \"neutron-70b2-account-create-update-fzkfv\" (UID: \"4ab730c6-95e4-438f-90d5-04812d26384f\") " pod="openstack/neutron-70b2-account-create-update-fzkfv" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.780652 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-4tss6\" (UID: \"5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00\") " pod="openstack/dnsmasq-dns-74f6bcbc87-4tss6" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.780693 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2bcb322d-a528-4d31-b639-367f5f7f3dd1-operator-scripts\") pod \"neutron-db-create-9hwb8\" (UID: \"2bcb322d-a528-4d31-b639-367f5f7f3dd1\") " pod="openstack/neutron-db-create-9hwb8" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.780741 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nklf2\" (UniqueName: \"kubernetes.io/projected/4ab730c6-95e4-438f-90d5-04812d26384f-kube-api-access-nklf2\") pod \"neutron-70b2-account-create-update-fzkfv\" (UID: \"4ab730c6-95e4-438f-90d5-04812d26384f\") " pod="openstack/neutron-70b2-account-create-update-fzkfv" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.780762 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j7gqg\" (UniqueName: \"kubernetes.io/projected/2bcb322d-a528-4d31-b639-367f5f7f3dd1-kube-api-access-j7gqg\") pod \"neutron-db-create-9hwb8\" (UID: \"2bcb322d-a528-4d31-b639-367f5f7f3dd1\") " pod="openstack/neutron-db-create-9hwb8" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.780782 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-4tss6\" (UID: \"5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00\") " pod="openstack/dnsmasq-dns-74f6bcbc87-4tss6" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.780801 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mmltc\" (UniqueName: \"kubernetes.io/projected/5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00-kube-api-access-mmltc\") pod \"dnsmasq-dns-74f6bcbc87-4tss6\" (UID: \"5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00\") " pod="openstack/dnsmasq-dns-74f6bcbc87-4tss6" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.780829 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-4tss6\" (UID: \"5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00\") " pod="openstack/dnsmasq-dns-74f6bcbc87-4tss6" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.780845 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-4tss6\" (UID: \"5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00\") " pod="openstack/dnsmasq-dns-74f6bcbc87-4tss6" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.780865 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00-config\") pod \"dnsmasq-dns-74f6bcbc87-4tss6\" (UID: \"5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00\") " pod="openstack/dnsmasq-dns-74f6bcbc87-4tss6" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.781560 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2bcb322d-a528-4d31-b639-367f5f7f3dd1-operator-scripts\") pod \"neutron-db-create-9hwb8\" (UID: \"2bcb322d-a528-4d31-b639-367f5f7f3dd1\") " pod="openstack/neutron-db-create-9hwb8" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.799794 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j7gqg\" (UniqueName: \"kubernetes.io/projected/2bcb322d-a528-4d31-b639-367f5f7f3dd1-kube-api-access-j7gqg\") pod \"neutron-db-create-9hwb8\" (UID: \"2bcb322d-a528-4d31-b639-367f5f7f3dd1\") " pod="openstack/neutron-db-create-9hwb8" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.882433 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ab730c6-95e4-438f-90d5-04812d26384f-operator-scripts\") pod \"neutron-70b2-account-create-update-fzkfv\" (UID: \"4ab730c6-95e4-438f-90d5-04812d26384f\") " pod="openstack/neutron-70b2-account-create-update-fzkfv" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.882762 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-4tss6\" (UID: \"5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00\") " pod="openstack/dnsmasq-dns-74f6bcbc87-4tss6" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.882833 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nklf2\" (UniqueName: \"kubernetes.io/projected/4ab730c6-95e4-438f-90d5-04812d26384f-kube-api-access-nklf2\") pod \"neutron-70b2-account-create-update-fzkfv\" (UID: \"4ab730c6-95e4-438f-90d5-04812d26384f\") " pod="openstack/neutron-70b2-account-create-update-fzkfv" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.882856 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-4tss6\" (UID: \"5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00\") " pod="openstack/dnsmasq-dns-74f6bcbc87-4tss6" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.882873 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mmltc\" (UniqueName: \"kubernetes.io/projected/5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00-kube-api-access-mmltc\") pod \"dnsmasq-dns-74f6bcbc87-4tss6\" (UID: \"5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00\") " pod="openstack/dnsmasq-dns-74f6bcbc87-4tss6" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.882900 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-4tss6\" (UID: \"5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00\") " pod="openstack/dnsmasq-dns-74f6bcbc87-4tss6" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.882932 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-4tss6\" (UID: \"5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00\") " pod="openstack/dnsmasq-dns-74f6bcbc87-4tss6" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.882952 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00-config\") pod \"dnsmasq-dns-74f6bcbc87-4tss6\" (UID: \"5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00\") " pod="openstack/dnsmasq-dns-74f6bcbc87-4tss6" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.883298 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ab730c6-95e4-438f-90d5-04812d26384f-operator-scripts\") pod \"neutron-70b2-account-create-update-fzkfv\" (UID: \"4ab730c6-95e4-438f-90d5-04812d26384f\") " pod="openstack/neutron-70b2-account-create-update-fzkfv" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.883719 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00-config\") pod \"dnsmasq-dns-74f6bcbc87-4tss6\" (UID: \"5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00\") " pod="openstack/dnsmasq-dns-74f6bcbc87-4tss6" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.883903 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-4tss6\" (UID: \"5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00\") " pod="openstack/dnsmasq-dns-74f6bcbc87-4tss6" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.884996 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-4tss6\" (UID: \"5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00\") " pod="openstack/dnsmasq-dns-74f6bcbc87-4tss6" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.885226 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-4tss6\" (UID: \"5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00\") " pod="openstack/dnsmasq-dns-74f6bcbc87-4tss6" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.885314 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-9hwb8" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.885361 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-4tss6\" (UID: \"5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00\") " pod="openstack/dnsmasq-dns-74f6bcbc87-4tss6" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.920001 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nklf2\" (UniqueName: \"kubernetes.io/projected/4ab730c6-95e4-438f-90d5-04812d26384f-kube-api-access-nklf2\") pod \"neutron-70b2-account-create-update-fzkfv\" (UID: \"4ab730c6-95e4-438f-90d5-04812d26384f\") " pod="openstack/neutron-70b2-account-create-update-fzkfv" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.928846 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mmltc\" (UniqueName: \"kubernetes.io/projected/5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00-kube-api-access-mmltc\") pod \"dnsmasq-dns-74f6bcbc87-4tss6\" (UID: \"5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00\") " pod="openstack/dnsmasq-dns-74f6bcbc87-4tss6" Dec 11 02:21:21 crc kubenswrapper[4824]: I1211 02:21:21.954850 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-4tss6" Dec 11 02:21:22 crc kubenswrapper[4824]: I1211 02:21:22.010160 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-70b2-account-create-update-fzkfv" Dec 11 02:21:22 crc kubenswrapper[4824]: I1211 02:21:22.101504 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b946c75cc-6b4w8" event={"ID":"f30f409e-2171-4b62-8e1b-e72ea450ec8d","Type":"ContainerStarted","Data":"53187289c93a8b376d356abc21a4121a086e61070d2fdf964c71278540a414ef"} Dec 11 02:21:22 crc kubenswrapper[4824]: I1211 02:21:22.101551 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5b946c75cc-6b4w8" Dec 11 02:21:22 crc kubenswrapper[4824]: I1211 02:21:22.234885 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5b946c75cc-6b4w8" podStartSLOduration=3.234869301 podStartE2EDuration="3.234869301s" podCreationTimestamp="2025-12-11 02:21:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:21:22.131554228 +0000 UTC m=+1223.820591607" watchObservedRunningTime="2025-12-11 02:21:22.234869301 +0000 UTC m=+1223.923906670" Dec 11 02:21:22 crc kubenswrapper[4824]: I1211 02:21:22.239045 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-5a29-account-create-update-sb4qh"] Dec 11 02:21:22 crc kubenswrapper[4824]: I1211 02:21:22.324646 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-bce9-account-create-update-qnmss"] Dec 11 02:21:22 crc kubenswrapper[4824]: I1211 02:21:22.361410 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-qwpmm"] Dec 11 02:21:22 crc kubenswrapper[4824]: I1211 02:21:22.511163 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-xg279"] Dec 11 02:21:22 crc kubenswrapper[4824]: I1211 02:21:22.554928 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-wdzl5"] Dec 11 02:21:22 crc kubenswrapper[4824]: I1211 02:21:22.670528 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-4tss6"] Dec 11 02:21:22 crc kubenswrapper[4824]: W1211 02:21:22.680149 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4ab730c6_95e4_438f_90d5_04812d26384f.slice/crio-57ad96f5ecfc92f344c308fa42e6c21c251626a570a7f09b6341936c5a23662b WatchSource:0}: Error finding container 57ad96f5ecfc92f344c308fa42e6c21c251626a570a7f09b6341936c5a23662b: Status 404 returned error can't find the container with id 57ad96f5ecfc92f344c308fa42e6c21c251626a570a7f09b6341936c5a23662b Dec 11 02:21:22 crc kubenswrapper[4824]: W1211 02:21:22.684574 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2bcb322d_a528_4d31_b639_367f5f7f3dd1.slice/crio-9a91daa221986bc252942a357521ffe55083096fc82f97958381f8aec1faad43 WatchSource:0}: Error finding container 9a91daa221986bc252942a357521ffe55083096fc82f97958381f8aec1faad43: Status 404 returned error can't find the container with id 9a91daa221986bc252942a357521ffe55083096fc82f97958381f8aec1faad43 Dec 11 02:21:22 crc kubenswrapper[4824]: W1211 02:21:22.686626 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5a1e4ee5_c52c_4c35_9978_2d03c2ba0d00.slice/crio-fedd41080434400a86cab37fd05ed9bf68b01384fa678fb9e8019221752f22cd WatchSource:0}: Error finding container fedd41080434400a86cab37fd05ed9bf68b01384fa678fb9e8019221752f22cd: Status 404 returned error can't find the container with id fedd41080434400a86cab37fd05ed9bf68b01384fa678fb9e8019221752f22cd Dec 11 02:21:22 crc kubenswrapper[4824]: I1211 02:21:22.695100 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-9hwb8"] Dec 11 02:21:22 crc kubenswrapper[4824]: I1211 02:21:22.704091 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-70b2-account-create-update-fzkfv"] Dec 11 02:21:23 crc kubenswrapper[4824]: I1211 02:21:23.107469 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-9hwb8" event={"ID":"2bcb322d-a528-4d31-b639-367f5f7f3dd1","Type":"ContainerStarted","Data":"0496ed10e37ac8801f313f31fb43102a538a86330c9c8f2f332bd62537eacb66"} Dec 11 02:21:23 crc kubenswrapper[4824]: I1211 02:21:23.107742 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-9hwb8" event={"ID":"2bcb322d-a528-4d31-b639-367f5f7f3dd1","Type":"ContainerStarted","Data":"9a91daa221986bc252942a357521ffe55083096fc82f97958381f8aec1faad43"} Dec 11 02:21:23 crc kubenswrapper[4824]: I1211 02:21:23.110430 4824 generic.go:334] "Generic (PLEG): container finished" podID="20766632-27ff-42ab-9b3c-28aa8b8912bf" containerID="4e15e7888c04c76b15f70f53b7ac680863bb27fccb96052650dd3487bf114c16" exitCode=0 Dec 11 02:21:23 crc kubenswrapper[4824]: I1211 02:21:23.110540 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-qwpmm" event={"ID":"20766632-27ff-42ab-9b3c-28aa8b8912bf","Type":"ContainerDied","Data":"4e15e7888c04c76b15f70f53b7ac680863bb27fccb96052650dd3487bf114c16"} Dec 11 02:21:23 crc kubenswrapper[4824]: I1211 02:21:23.110611 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-qwpmm" event={"ID":"20766632-27ff-42ab-9b3c-28aa8b8912bf","Type":"ContainerStarted","Data":"60044dfe8d9023325366ecbcaea7b6c8069f5987aa77797cac1add0cdba3da5f"} Dec 11 02:21:23 crc kubenswrapper[4824]: I1211 02:21:23.113582 4824 generic.go:334] "Generic (PLEG): container finished" podID="9d9a4a8f-d58e-4a90-9c33-3ffa6a7f5c36" containerID="6f33dbd6ce00a9d9546c84dba259368ec1dc821d57f101e08545e1808635435b" exitCode=0 Dec 11 02:21:23 crc kubenswrapper[4824]: I1211 02:21:23.113628 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-5a29-account-create-update-sb4qh" event={"ID":"9d9a4a8f-d58e-4a90-9c33-3ffa6a7f5c36","Type":"ContainerDied","Data":"6f33dbd6ce00a9d9546c84dba259368ec1dc821d57f101e08545e1808635435b"} Dec 11 02:21:23 crc kubenswrapper[4824]: I1211 02:21:23.113970 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-5a29-account-create-update-sb4qh" event={"ID":"9d9a4a8f-d58e-4a90-9c33-3ffa6a7f5c36","Type":"ContainerStarted","Data":"e98a02b221742cf37584a647bb19832b42875bf8bffdd1cd37fa182add099899"} Dec 11 02:21:23 crc kubenswrapper[4824]: I1211 02:21:23.115777 4824 generic.go:334] "Generic (PLEG): container finished" podID="5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00" containerID="63d0fc2a3cdbd5be8649bacd15b9dadb1f6545b630b7a004fdef91e677c73065" exitCode=0 Dec 11 02:21:23 crc kubenswrapper[4824]: I1211 02:21:23.115845 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-4tss6" event={"ID":"5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00","Type":"ContainerDied","Data":"63d0fc2a3cdbd5be8649bacd15b9dadb1f6545b630b7a004fdef91e677c73065"} Dec 11 02:21:23 crc kubenswrapper[4824]: I1211 02:21:23.115873 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-4tss6" event={"ID":"5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00","Type":"ContainerStarted","Data":"fedd41080434400a86cab37fd05ed9bf68b01384fa678fb9e8019221752f22cd"} Dec 11 02:21:23 crc kubenswrapper[4824]: I1211 02:21:23.117302 4824 generic.go:334] "Generic (PLEG): container finished" podID="793e155e-73e9-46d5-9de6-a4aebf3ca2b4" containerID="49694cc653f8c737dd7fc310645d2f7ea51a57f908795caeb53db67a847171bd" exitCode=0 Dec 11 02:21:23 crc kubenswrapper[4824]: I1211 02:21:23.117327 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-bce9-account-create-update-qnmss" event={"ID":"793e155e-73e9-46d5-9de6-a4aebf3ca2b4","Type":"ContainerDied","Data":"49694cc653f8c737dd7fc310645d2f7ea51a57f908795caeb53db67a847171bd"} Dec 11 02:21:23 crc kubenswrapper[4824]: I1211 02:21:23.117367 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-bce9-account-create-update-qnmss" event={"ID":"793e155e-73e9-46d5-9de6-a4aebf3ca2b4","Type":"ContainerStarted","Data":"6bbad84d1bfa2cbe0c9e56b4e71f15e0780ba6551c68f4cd4ea67c7dcc2fa8dc"} Dec 11 02:21:23 crc kubenswrapper[4824]: I1211 02:21:23.123322 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-70b2-account-create-update-fzkfv" event={"ID":"4ab730c6-95e4-438f-90d5-04812d26384f","Type":"ContainerStarted","Data":"b8e0d5fc75fb9a79e6229c9786439bf7ba7382203aec403b4081a22d8bdf8d57"} Dec 11 02:21:23 crc kubenswrapper[4824]: I1211 02:21:23.123409 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-70b2-account-create-update-fzkfv" event={"ID":"4ab730c6-95e4-438f-90d5-04812d26384f","Type":"ContainerStarted","Data":"57ad96f5ecfc92f344c308fa42e6c21c251626a570a7f09b6341936c5a23662b"} Dec 11 02:21:23 crc kubenswrapper[4824]: I1211 02:21:23.125486 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-wdzl5" event={"ID":"922cc36d-5fd8-4cba-b6e3-5b16e6cb7531","Type":"ContainerStarted","Data":"83180a0899238d7258d72d33fd50a17fd8f1406bc4480383825cf4102b7dd049"} Dec 11 02:21:23 crc kubenswrapper[4824]: I1211 02:21:23.126939 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5b946c75cc-6b4w8" podUID="f30f409e-2171-4b62-8e1b-e72ea450ec8d" containerName="dnsmasq-dns" containerID="cri-o://53187289c93a8b376d356abc21a4121a086e61070d2fdf964c71278540a414ef" gracePeriod=10 Dec 11 02:21:23 crc kubenswrapper[4824]: I1211 02:21:23.127242 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-xg279" event={"ID":"57686bda-9117-48cc-b29d-f5ff374937c9","Type":"ContainerStarted","Data":"4b2deaf1f85e37a911132c3ac53a75fb38c3326d25e7be47b5415a082094bb05"} Dec 11 02:21:23 crc kubenswrapper[4824]: I1211 02:21:23.127272 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-xg279" event={"ID":"57686bda-9117-48cc-b29d-f5ff374937c9","Type":"ContainerStarted","Data":"6ade820d4ca73a0cdca511e6910bfe6526df953b60cbba2960a74a78829d5ebf"} Dec 11 02:21:23 crc kubenswrapper[4824]: I1211 02:21:23.132844 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-create-9hwb8" podStartSLOduration=2.132826448 podStartE2EDuration="2.132826448s" podCreationTimestamp="2025-12-11 02:21:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:21:23.129758254 +0000 UTC m=+1224.818795633" watchObservedRunningTime="2025-12-11 02:21:23.132826448 +0000 UTC m=+1224.821863827" Dec 11 02:21:23 crc kubenswrapper[4824]: I1211 02:21:23.162172 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-70b2-account-create-update-fzkfv" podStartSLOduration=2.162152742 podStartE2EDuration="2.162152742s" podCreationTimestamp="2025-12-11 02:21:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:21:23.148736234 +0000 UTC m=+1224.837773613" watchObservedRunningTime="2025-12-11 02:21:23.162152742 +0000 UTC m=+1224.851190121" Dec 11 02:21:23 crc kubenswrapper[4824]: I1211 02:21:23.527900 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b946c75cc-6b4w8" Dec 11 02:21:23 crc kubenswrapper[4824]: I1211 02:21:23.612211 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f30f409e-2171-4b62-8e1b-e72ea450ec8d-ovsdbserver-sb\") pod \"f30f409e-2171-4b62-8e1b-e72ea450ec8d\" (UID: \"f30f409e-2171-4b62-8e1b-e72ea450ec8d\") " Dec 11 02:21:23 crc kubenswrapper[4824]: I1211 02:21:23.612312 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f30f409e-2171-4b62-8e1b-e72ea450ec8d-ovsdbserver-nb\") pod \"f30f409e-2171-4b62-8e1b-e72ea450ec8d\" (UID: \"f30f409e-2171-4b62-8e1b-e72ea450ec8d\") " Dec 11 02:21:23 crc kubenswrapper[4824]: I1211 02:21:23.612359 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f30f409e-2171-4b62-8e1b-e72ea450ec8d-config\") pod \"f30f409e-2171-4b62-8e1b-e72ea450ec8d\" (UID: \"f30f409e-2171-4b62-8e1b-e72ea450ec8d\") " Dec 11 02:21:23 crc kubenswrapper[4824]: I1211 02:21:23.612430 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-drwxg\" (UniqueName: \"kubernetes.io/projected/f30f409e-2171-4b62-8e1b-e72ea450ec8d-kube-api-access-drwxg\") pod \"f30f409e-2171-4b62-8e1b-e72ea450ec8d\" (UID: \"f30f409e-2171-4b62-8e1b-e72ea450ec8d\") " Dec 11 02:21:23 crc kubenswrapper[4824]: I1211 02:21:23.612529 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f30f409e-2171-4b62-8e1b-e72ea450ec8d-dns-svc\") pod \"f30f409e-2171-4b62-8e1b-e72ea450ec8d\" (UID: \"f30f409e-2171-4b62-8e1b-e72ea450ec8d\") " Dec 11 02:21:23 crc kubenswrapper[4824]: I1211 02:21:23.624388 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f30f409e-2171-4b62-8e1b-e72ea450ec8d-kube-api-access-drwxg" (OuterVolumeSpecName: "kube-api-access-drwxg") pod "f30f409e-2171-4b62-8e1b-e72ea450ec8d" (UID: "f30f409e-2171-4b62-8e1b-e72ea450ec8d"). InnerVolumeSpecName "kube-api-access-drwxg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:21:23 crc kubenswrapper[4824]: I1211 02:21:23.661819 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f30f409e-2171-4b62-8e1b-e72ea450ec8d-config" (OuterVolumeSpecName: "config") pod "f30f409e-2171-4b62-8e1b-e72ea450ec8d" (UID: "f30f409e-2171-4b62-8e1b-e72ea450ec8d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:21:23 crc kubenswrapper[4824]: I1211 02:21:23.666532 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f30f409e-2171-4b62-8e1b-e72ea450ec8d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f30f409e-2171-4b62-8e1b-e72ea450ec8d" (UID: "f30f409e-2171-4b62-8e1b-e72ea450ec8d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:21:23 crc kubenswrapper[4824]: I1211 02:21:23.668300 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f30f409e-2171-4b62-8e1b-e72ea450ec8d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f30f409e-2171-4b62-8e1b-e72ea450ec8d" (UID: "f30f409e-2171-4b62-8e1b-e72ea450ec8d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:21:23 crc kubenswrapper[4824]: I1211 02:21:23.673464 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f30f409e-2171-4b62-8e1b-e72ea450ec8d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f30f409e-2171-4b62-8e1b-e72ea450ec8d" (UID: "f30f409e-2171-4b62-8e1b-e72ea450ec8d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:21:23 crc kubenswrapper[4824]: I1211 02:21:23.714086 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f30f409e-2171-4b62-8e1b-e72ea450ec8d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 11 02:21:23 crc kubenswrapper[4824]: I1211 02:21:23.714126 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f30f409e-2171-4b62-8e1b-e72ea450ec8d-config\") on node \"crc\" DevicePath \"\"" Dec 11 02:21:23 crc kubenswrapper[4824]: I1211 02:21:23.714135 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-drwxg\" (UniqueName: \"kubernetes.io/projected/f30f409e-2171-4b62-8e1b-e72ea450ec8d-kube-api-access-drwxg\") on node \"crc\" DevicePath \"\"" Dec 11 02:21:23 crc kubenswrapper[4824]: I1211 02:21:23.714146 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f30f409e-2171-4b62-8e1b-e72ea450ec8d-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 11 02:21:23 crc kubenswrapper[4824]: I1211 02:21:23.714156 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f30f409e-2171-4b62-8e1b-e72ea450ec8d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 11 02:21:24 crc kubenswrapper[4824]: I1211 02:21:24.140106 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-4tss6" event={"ID":"5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00","Type":"ContainerStarted","Data":"ae65a91d118ac4d63800cb6c6bcfa72370d4dec5921ffa4c291a7fe609becdf5"} Dec 11 02:21:24 crc kubenswrapper[4824]: I1211 02:21:24.140257 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-74f6bcbc87-4tss6" Dec 11 02:21:24 crc kubenswrapper[4824]: I1211 02:21:24.142190 4824 generic.go:334] "Generic (PLEG): container finished" podID="4ab730c6-95e4-438f-90d5-04812d26384f" containerID="b8e0d5fc75fb9a79e6229c9786439bf7ba7382203aec403b4081a22d8bdf8d57" exitCode=0 Dec 11 02:21:24 crc kubenswrapper[4824]: I1211 02:21:24.142347 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-70b2-account-create-update-fzkfv" event={"ID":"4ab730c6-95e4-438f-90d5-04812d26384f","Type":"ContainerDied","Data":"b8e0d5fc75fb9a79e6229c9786439bf7ba7382203aec403b4081a22d8bdf8d57"} Dec 11 02:21:24 crc kubenswrapper[4824]: I1211 02:21:24.144006 4824 generic.go:334] "Generic (PLEG): container finished" podID="57686bda-9117-48cc-b29d-f5ff374937c9" containerID="4b2deaf1f85e37a911132c3ac53a75fb38c3326d25e7be47b5415a082094bb05" exitCode=0 Dec 11 02:21:24 crc kubenswrapper[4824]: I1211 02:21:24.144088 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-xg279" event={"ID":"57686bda-9117-48cc-b29d-f5ff374937c9","Type":"ContainerDied","Data":"4b2deaf1f85e37a911132c3ac53a75fb38c3326d25e7be47b5415a082094bb05"} Dec 11 02:21:24 crc kubenswrapper[4824]: I1211 02:21:24.146554 4824 generic.go:334] "Generic (PLEG): container finished" podID="2bcb322d-a528-4d31-b639-367f5f7f3dd1" containerID="0496ed10e37ac8801f313f31fb43102a538a86330c9c8f2f332bd62537eacb66" exitCode=0 Dec 11 02:21:24 crc kubenswrapper[4824]: I1211 02:21:24.146642 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-9hwb8" event={"ID":"2bcb322d-a528-4d31-b639-367f5f7f3dd1","Type":"ContainerDied","Data":"0496ed10e37ac8801f313f31fb43102a538a86330c9c8f2f332bd62537eacb66"} Dec 11 02:21:24 crc kubenswrapper[4824]: I1211 02:21:24.148321 4824 generic.go:334] "Generic (PLEG): container finished" podID="f30f409e-2171-4b62-8e1b-e72ea450ec8d" containerID="53187289c93a8b376d356abc21a4121a086e61070d2fdf964c71278540a414ef" exitCode=0 Dec 11 02:21:24 crc kubenswrapper[4824]: I1211 02:21:24.148496 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b946c75cc-6b4w8" Dec 11 02:21:24 crc kubenswrapper[4824]: I1211 02:21:24.150516 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b946c75cc-6b4w8" event={"ID":"f30f409e-2171-4b62-8e1b-e72ea450ec8d","Type":"ContainerDied","Data":"53187289c93a8b376d356abc21a4121a086e61070d2fdf964c71278540a414ef"} Dec 11 02:21:24 crc kubenswrapper[4824]: I1211 02:21:24.150550 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b946c75cc-6b4w8" event={"ID":"f30f409e-2171-4b62-8e1b-e72ea450ec8d","Type":"ContainerDied","Data":"3424e309b384bb43f44fb7cfe36eefd8aba2084b1ec5aa95c1fc953ebd648626"} Dec 11 02:21:24 crc kubenswrapper[4824]: I1211 02:21:24.150571 4824 scope.go:117] "RemoveContainer" containerID="53187289c93a8b376d356abc21a4121a086e61070d2fdf964c71278540a414ef" Dec 11 02:21:24 crc kubenswrapper[4824]: I1211 02:21:24.166648 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-74f6bcbc87-4tss6" podStartSLOduration=3.166570809 podStartE2EDuration="3.166570809s" podCreationTimestamp="2025-12-11 02:21:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:21:24.160181094 +0000 UTC m=+1225.849218473" watchObservedRunningTime="2025-12-11 02:21:24.166570809 +0000 UTC m=+1225.855608188" Dec 11 02:21:24 crc kubenswrapper[4824]: I1211 02:21:24.189484 4824 scope.go:117] "RemoveContainer" containerID="dc4ae37edeb4ed98398c339380c86390f73246b108db79995cc4b50af1dce39e" Dec 11 02:21:24 crc kubenswrapper[4824]: E1211 02:21:24.216548 4824 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 38.102.83.50:35002->38.102.83.50:42353: read tcp 38.102.83.50:35002->38.102.83.50:42353: read: connection reset by peer Dec 11 02:21:24 crc kubenswrapper[4824]: I1211 02:21:24.227331 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-6b4w8"] Dec 11 02:21:24 crc kubenswrapper[4824]: I1211 02:21:24.236712 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-6b4w8"] Dec 11 02:21:24 crc kubenswrapper[4824]: I1211 02:21:24.250268 4824 scope.go:117] "RemoveContainer" containerID="53187289c93a8b376d356abc21a4121a086e61070d2fdf964c71278540a414ef" Dec 11 02:21:24 crc kubenswrapper[4824]: E1211 02:21:24.250835 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"53187289c93a8b376d356abc21a4121a086e61070d2fdf964c71278540a414ef\": container with ID starting with 53187289c93a8b376d356abc21a4121a086e61070d2fdf964c71278540a414ef not found: ID does not exist" containerID="53187289c93a8b376d356abc21a4121a086e61070d2fdf964c71278540a414ef" Dec 11 02:21:24 crc kubenswrapper[4824]: I1211 02:21:24.250907 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53187289c93a8b376d356abc21a4121a086e61070d2fdf964c71278540a414ef"} err="failed to get container status \"53187289c93a8b376d356abc21a4121a086e61070d2fdf964c71278540a414ef\": rpc error: code = NotFound desc = could not find container \"53187289c93a8b376d356abc21a4121a086e61070d2fdf964c71278540a414ef\": container with ID starting with 53187289c93a8b376d356abc21a4121a086e61070d2fdf964c71278540a414ef not found: ID does not exist" Dec 11 02:21:24 crc kubenswrapper[4824]: I1211 02:21:24.250960 4824 scope.go:117] "RemoveContainer" containerID="dc4ae37edeb4ed98398c339380c86390f73246b108db79995cc4b50af1dce39e" Dec 11 02:21:24 crc kubenswrapper[4824]: E1211 02:21:24.253698 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc4ae37edeb4ed98398c339380c86390f73246b108db79995cc4b50af1dce39e\": container with ID starting with dc4ae37edeb4ed98398c339380c86390f73246b108db79995cc4b50af1dce39e not found: ID does not exist" containerID="dc4ae37edeb4ed98398c339380c86390f73246b108db79995cc4b50af1dce39e" Dec 11 02:21:24 crc kubenswrapper[4824]: I1211 02:21:24.253738 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc4ae37edeb4ed98398c339380c86390f73246b108db79995cc4b50af1dce39e"} err="failed to get container status \"dc4ae37edeb4ed98398c339380c86390f73246b108db79995cc4b50af1dce39e\": rpc error: code = NotFound desc = could not find container \"dc4ae37edeb4ed98398c339380c86390f73246b108db79995cc4b50af1dce39e\": container with ID starting with dc4ae37edeb4ed98398c339380c86390f73246b108db79995cc4b50af1dce39e not found: ID does not exist" Dec 11 02:21:24 crc kubenswrapper[4824]: I1211 02:21:24.540082 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-5a29-account-create-update-sb4qh" Dec 11 02:21:24 crc kubenswrapper[4824]: I1211 02:21:24.628666 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-55qdq\" (UniqueName: \"kubernetes.io/projected/9d9a4a8f-d58e-4a90-9c33-3ffa6a7f5c36-kube-api-access-55qdq\") pod \"9d9a4a8f-d58e-4a90-9c33-3ffa6a7f5c36\" (UID: \"9d9a4a8f-d58e-4a90-9c33-3ffa6a7f5c36\") " Dec 11 02:21:24 crc kubenswrapper[4824]: I1211 02:21:24.628734 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9d9a4a8f-d58e-4a90-9c33-3ffa6a7f5c36-operator-scripts\") pod \"9d9a4a8f-d58e-4a90-9c33-3ffa6a7f5c36\" (UID: \"9d9a4a8f-d58e-4a90-9c33-3ffa6a7f5c36\") " Dec 11 02:21:24 crc kubenswrapper[4824]: I1211 02:21:24.629166 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d9a4a8f-d58e-4a90-9c33-3ffa6a7f5c36-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9d9a4a8f-d58e-4a90-9c33-3ffa6a7f5c36" (UID: "9d9a4a8f-d58e-4a90-9c33-3ffa6a7f5c36"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:21:24 crc kubenswrapper[4824]: I1211 02:21:24.637348 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d9a4a8f-d58e-4a90-9c33-3ffa6a7f5c36-kube-api-access-55qdq" (OuterVolumeSpecName: "kube-api-access-55qdq") pod "9d9a4a8f-d58e-4a90-9c33-3ffa6a7f5c36" (UID: "9d9a4a8f-d58e-4a90-9c33-3ffa6a7f5c36"). InnerVolumeSpecName "kube-api-access-55qdq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:21:24 crc kubenswrapper[4824]: I1211 02:21:24.641087 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-bce9-account-create-update-qnmss" Dec 11 02:21:24 crc kubenswrapper[4824]: I1211 02:21:24.642838 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f30f409e-2171-4b62-8e1b-e72ea450ec8d" path="/var/lib/kubelet/pods/f30f409e-2171-4b62-8e1b-e72ea450ec8d/volumes" Dec 11 02:21:24 crc kubenswrapper[4824]: I1211 02:21:24.643435 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-qwpmm" Dec 11 02:21:24 crc kubenswrapper[4824]: I1211 02:21:24.650679 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-xg279" Dec 11 02:21:24 crc kubenswrapper[4824]: I1211 02:21:24.730588 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fd7g5\" (UniqueName: \"kubernetes.io/projected/793e155e-73e9-46d5-9de6-a4aebf3ca2b4-kube-api-access-fd7g5\") pod \"793e155e-73e9-46d5-9de6-a4aebf3ca2b4\" (UID: \"793e155e-73e9-46d5-9de6-a4aebf3ca2b4\") " Dec 11 02:21:24 crc kubenswrapper[4824]: I1211 02:21:24.730637 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/793e155e-73e9-46d5-9de6-a4aebf3ca2b4-operator-scripts\") pod \"793e155e-73e9-46d5-9de6-a4aebf3ca2b4\" (UID: \"793e155e-73e9-46d5-9de6-a4aebf3ca2b4\") " Dec 11 02:21:24 crc kubenswrapper[4824]: I1211 02:21:24.730707 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/57686bda-9117-48cc-b29d-f5ff374937c9-operator-scripts\") pod \"57686bda-9117-48cc-b29d-f5ff374937c9\" (UID: \"57686bda-9117-48cc-b29d-f5ff374937c9\") " Dec 11 02:21:24 crc kubenswrapper[4824]: I1211 02:21:24.730755 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5dgh9\" (UniqueName: \"kubernetes.io/projected/20766632-27ff-42ab-9b3c-28aa8b8912bf-kube-api-access-5dgh9\") pod \"20766632-27ff-42ab-9b3c-28aa8b8912bf\" (UID: \"20766632-27ff-42ab-9b3c-28aa8b8912bf\") " Dec 11 02:21:24 crc kubenswrapper[4824]: I1211 02:21:24.730794 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20766632-27ff-42ab-9b3c-28aa8b8912bf-operator-scripts\") pod \"20766632-27ff-42ab-9b3c-28aa8b8912bf\" (UID: \"20766632-27ff-42ab-9b3c-28aa8b8912bf\") " Dec 11 02:21:24 crc kubenswrapper[4824]: I1211 02:21:24.730846 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c2ndh\" (UniqueName: \"kubernetes.io/projected/57686bda-9117-48cc-b29d-f5ff374937c9-kube-api-access-c2ndh\") pod \"57686bda-9117-48cc-b29d-f5ff374937c9\" (UID: \"57686bda-9117-48cc-b29d-f5ff374937c9\") " Dec 11 02:21:24 crc kubenswrapper[4824]: I1211 02:21:24.731247 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/793e155e-73e9-46d5-9de6-a4aebf3ca2b4-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "793e155e-73e9-46d5-9de6-a4aebf3ca2b4" (UID: "793e155e-73e9-46d5-9de6-a4aebf3ca2b4"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:21:24 crc kubenswrapper[4824]: I1211 02:21:24.731309 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/20766632-27ff-42ab-9b3c-28aa8b8912bf-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "20766632-27ff-42ab-9b3c-28aa8b8912bf" (UID: "20766632-27ff-42ab-9b3c-28aa8b8912bf"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:21:24 crc kubenswrapper[4824]: I1211 02:21:24.731573 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/57686bda-9117-48cc-b29d-f5ff374937c9-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "57686bda-9117-48cc-b29d-f5ff374937c9" (UID: "57686bda-9117-48cc-b29d-f5ff374937c9"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:21:24 crc kubenswrapper[4824]: I1211 02:21:24.732028 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/793e155e-73e9-46d5-9de6-a4aebf3ca2b4-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 02:21:24 crc kubenswrapper[4824]: I1211 02:21:24.732045 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-55qdq\" (UniqueName: \"kubernetes.io/projected/9d9a4a8f-d58e-4a90-9c33-3ffa6a7f5c36-kube-api-access-55qdq\") on node \"crc\" DevicePath \"\"" Dec 11 02:21:24 crc kubenswrapper[4824]: I1211 02:21:24.732056 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/57686bda-9117-48cc-b29d-f5ff374937c9-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 02:21:24 crc kubenswrapper[4824]: I1211 02:21:24.732065 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9d9a4a8f-d58e-4a90-9c33-3ffa6a7f5c36-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 02:21:24 crc kubenswrapper[4824]: I1211 02:21:24.732073 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20766632-27ff-42ab-9b3c-28aa8b8912bf-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 02:21:24 crc kubenswrapper[4824]: I1211 02:21:24.734070 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/793e155e-73e9-46d5-9de6-a4aebf3ca2b4-kube-api-access-fd7g5" (OuterVolumeSpecName: "kube-api-access-fd7g5") pod "793e155e-73e9-46d5-9de6-a4aebf3ca2b4" (UID: "793e155e-73e9-46d5-9de6-a4aebf3ca2b4"). InnerVolumeSpecName "kube-api-access-fd7g5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:21:24 crc kubenswrapper[4824]: I1211 02:21:24.734577 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57686bda-9117-48cc-b29d-f5ff374937c9-kube-api-access-c2ndh" (OuterVolumeSpecName: "kube-api-access-c2ndh") pod "57686bda-9117-48cc-b29d-f5ff374937c9" (UID: "57686bda-9117-48cc-b29d-f5ff374937c9"). InnerVolumeSpecName "kube-api-access-c2ndh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:21:24 crc kubenswrapper[4824]: I1211 02:21:24.734992 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20766632-27ff-42ab-9b3c-28aa8b8912bf-kube-api-access-5dgh9" (OuterVolumeSpecName: "kube-api-access-5dgh9") pod "20766632-27ff-42ab-9b3c-28aa8b8912bf" (UID: "20766632-27ff-42ab-9b3c-28aa8b8912bf"). InnerVolumeSpecName "kube-api-access-5dgh9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:21:24 crc kubenswrapper[4824]: I1211 02:21:24.833150 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c2ndh\" (UniqueName: \"kubernetes.io/projected/57686bda-9117-48cc-b29d-f5ff374937c9-kube-api-access-c2ndh\") on node \"crc\" DevicePath \"\"" Dec 11 02:21:24 crc kubenswrapper[4824]: I1211 02:21:24.833176 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fd7g5\" (UniqueName: \"kubernetes.io/projected/793e155e-73e9-46d5-9de6-a4aebf3ca2b4-kube-api-access-fd7g5\") on node \"crc\" DevicePath \"\"" Dec 11 02:21:24 crc kubenswrapper[4824]: I1211 02:21:24.833185 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5dgh9\" (UniqueName: \"kubernetes.io/projected/20766632-27ff-42ab-9b3c-28aa8b8912bf-kube-api-access-5dgh9\") on node \"crc\" DevicePath \"\"" Dec 11 02:21:25 crc kubenswrapper[4824]: I1211 02:21:25.158879 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-qwpmm" event={"ID":"20766632-27ff-42ab-9b3c-28aa8b8912bf","Type":"ContainerDied","Data":"60044dfe8d9023325366ecbcaea7b6c8069f5987aa77797cac1add0cdba3da5f"} Dec 11 02:21:25 crc kubenswrapper[4824]: I1211 02:21:25.158923 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="60044dfe8d9023325366ecbcaea7b6c8069f5987aa77797cac1add0cdba3da5f" Dec 11 02:21:25 crc kubenswrapper[4824]: I1211 02:21:25.158991 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-qwpmm" Dec 11 02:21:25 crc kubenswrapper[4824]: I1211 02:21:25.165438 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-5a29-account-create-update-sb4qh" event={"ID":"9d9a4a8f-d58e-4a90-9c33-3ffa6a7f5c36","Type":"ContainerDied","Data":"e98a02b221742cf37584a647bb19832b42875bf8bffdd1cd37fa182add099899"} Dec 11 02:21:25 crc kubenswrapper[4824]: I1211 02:21:25.165660 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e98a02b221742cf37584a647bb19832b42875bf8bffdd1cd37fa182add099899" Dec 11 02:21:25 crc kubenswrapper[4824]: I1211 02:21:25.165783 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-5a29-account-create-update-sb4qh" Dec 11 02:21:25 crc kubenswrapper[4824]: I1211 02:21:25.172253 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-bce9-account-create-update-qnmss" event={"ID":"793e155e-73e9-46d5-9de6-a4aebf3ca2b4","Type":"ContainerDied","Data":"6bbad84d1bfa2cbe0c9e56b4e71f15e0780ba6551c68f4cd4ea67c7dcc2fa8dc"} Dec 11 02:21:25 crc kubenswrapper[4824]: I1211 02:21:25.172293 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6bbad84d1bfa2cbe0c9e56b4e71f15e0780ba6551c68f4cd4ea67c7dcc2fa8dc" Dec 11 02:21:25 crc kubenswrapper[4824]: I1211 02:21:25.172300 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-bce9-account-create-update-qnmss" Dec 11 02:21:25 crc kubenswrapper[4824]: I1211 02:21:25.175383 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-xg279" Dec 11 02:21:25 crc kubenswrapper[4824]: I1211 02:21:25.175395 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-xg279" event={"ID":"57686bda-9117-48cc-b29d-f5ff374937c9","Type":"ContainerDied","Data":"6ade820d4ca73a0cdca511e6910bfe6526df953b60cbba2960a74a78829d5ebf"} Dec 11 02:21:25 crc kubenswrapper[4824]: I1211 02:21:25.175417 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6ade820d4ca73a0cdca511e6910bfe6526df953b60cbba2960a74a78829d5ebf" Dec 11 02:21:28 crc kubenswrapper[4824]: I1211 02:21:28.028419 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-70b2-account-create-update-fzkfv" Dec 11 02:21:28 crc kubenswrapper[4824]: I1211 02:21:28.034517 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-9hwb8" Dec 11 02:21:28 crc kubenswrapper[4824]: I1211 02:21:28.106644 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j7gqg\" (UniqueName: \"kubernetes.io/projected/2bcb322d-a528-4d31-b639-367f5f7f3dd1-kube-api-access-j7gqg\") pod \"2bcb322d-a528-4d31-b639-367f5f7f3dd1\" (UID: \"2bcb322d-a528-4d31-b639-367f5f7f3dd1\") " Dec 11 02:21:28 crc kubenswrapper[4824]: I1211 02:21:28.106677 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ab730c6-95e4-438f-90d5-04812d26384f-operator-scripts\") pod \"4ab730c6-95e4-438f-90d5-04812d26384f\" (UID: \"4ab730c6-95e4-438f-90d5-04812d26384f\") " Dec 11 02:21:28 crc kubenswrapper[4824]: I1211 02:21:28.106738 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nklf2\" (UniqueName: \"kubernetes.io/projected/4ab730c6-95e4-438f-90d5-04812d26384f-kube-api-access-nklf2\") pod \"4ab730c6-95e4-438f-90d5-04812d26384f\" (UID: \"4ab730c6-95e4-438f-90d5-04812d26384f\") " Dec 11 02:21:28 crc kubenswrapper[4824]: I1211 02:21:28.106821 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2bcb322d-a528-4d31-b639-367f5f7f3dd1-operator-scripts\") pod \"2bcb322d-a528-4d31-b639-367f5f7f3dd1\" (UID: \"2bcb322d-a528-4d31-b639-367f5f7f3dd1\") " Dec 11 02:21:28 crc kubenswrapper[4824]: I1211 02:21:28.107682 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ab730c6-95e4-438f-90d5-04812d26384f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4ab730c6-95e4-438f-90d5-04812d26384f" (UID: "4ab730c6-95e4-438f-90d5-04812d26384f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:21:28 crc kubenswrapper[4824]: I1211 02:21:28.107742 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2bcb322d-a528-4d31-b639-367f5f7f3dd1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2bcb322d-a528-4d31-b639-367f5f7f3dd1" (UID: "2bcb322d-a528-4d31-b639-367f5f7f3dd1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:21:28 crc kubenswrapper[4824]: I1211 02:21:28.113884 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ab730c6-95e4-438f-90d5-04812d26384f-kube-api-access-nklf2" (OuterVolumeSpecName: "kube-api-access-nklf2") pod "4ab730c6-95e4-438f-90d5-04812d26384f" (UID: "4ab730c6-95e4-438f-90d5-04812d26384f"). InnerVolumeSpecName "kube-api-access-nklf2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:21:28 crc kubenswrapper[4824]: I1211 02:21:28.114376 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2bcb322d-a528-4d31-b639-367f5f7f3dd1-kube-api-access-j7gqg" (OuterVolumeSpecName: "kube-api-access-j7gqg") pod "2bcb322d-a528-4d31-b639-367f5f7f3dd1" (UID: "2bcb322d-a528-4d31-b639-367f5f7f3dd1"). InnerVolumeSpecName "kube-api-access-j7gqg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:21:28 crc kubenswrapper[4824]: I1211 02:21:28.208781 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nklf2\" (UniqueName: \"kubernetes.io/projected/4ab730c6-95e4-438f-90d5-04812d26384f-kube-api-access-nklf2\") on node \"crc\" DevicePath \"\"" Dec 11 02:21:28 crc kubenswrapper[4824]: I1211 02:21:28.208811 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2bcb322d-a528-4d31-b639-367f5f7f3dd1-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 02:21:28 crc kubenswrapper[4824]: I1211 02:21:28.208822 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j7gqg\" (UniqueName: \"kubernetes.io/projected/2bcb322d-a528-4d31-b639-367f5f7f3dd1-kube-api-access-j7gqg\") on node \"crc\" DevicePath \"\"" Dec 11 02:21:28 crc kubenswrapper[4824]: I1211 02:21:28.208831 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ab730c6-95e4-438f-90d5-04812d26384f-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 02:21:28 crc kubenswrapper[4824]: I1211 02:21:28.213212 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-70b2-account-create-update-fzkfv" Dec 11 02:21:28 crc kubenswrapper[4824]: I1211 02:21:28.213213 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-70b2-account-create-update-fzkfv" event={"ID":"4ab730c6-95e4-438f-90d5-04812d26384f","Type":"ContainerDied","Data":"57ad96f5ecfc92f344c308fa42e6c21c251626a570a7f09b6341936c5a23662b"} Dec 11 02:21:28 crc kubenswrapper[4824]: I1211 02:21:28.213559 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="57ad96f5ecfc92f344c308fa42e6c21c251626a570a7f09b6341936c5a23662b" Dec 11 02:21:28 crc kubenswrapper[4824]: I1211 02:21:28.216906 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-9hwb8" event={"ID":"2bcb322d-a528-4d31-b639-367f5f7f3dd1","Type":"ContainerDied","Data":"9a91daa221986bc252942a357521ffe55083096fc82f97958381f8aec1faad43"} Dec 11 02:21:28 crc kubenswrapper[4824]: I1211 02:21:28.216940 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9a91daa221986bc252942a357521ffe55083096fc82f97958381f8aec1faad43" Dec 11 02:21:28 crc kubenswrapper[4824]: I1211 02:21:28.217006 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-9hwb8" Dec 11 02:21:29 crc kubenswrapper[4824]: I1211 02:21:29.231213 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-wdzl5" event={"ID":"922cc36d-5fd8-4cba-b6e3-5b16e6cb7531","Type":"ContainerStarted","Data":"03ee038a6bf69c20fa36ea52fe0d3bec95eb293c2cc767dbb0d52a4f515cb1e2"} Dec 11 02:21:29 crc kubenswrapper[4824]: I1211 02:21:29.275931 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-wdzl5" podStartSLOduration=2.902196429 podStartE2EDuration="8.275903824s" podCreationTimestamp="2025-12-11 02:21:21 +0000 UTC" firstStartedPulling="2025-12-11 02:21:22.583155244 +0000 UTC m=+1224.272192623" lastFinishedPulling="2025-12-11 02:21:27.956862639 +0000 UTC m=+1229.645900018" observedRunningTime="2025-12-11 02:21:29.259628908 +0000 UTC m=+1230.948666327" watchObservedRunningTime="2025-12-11 02:21:29.275903824 +0000 UTC m=+1230.964941233" Dec 11 02:21:31 crc kubenswrapper[4824]: I1211 02:21:31.255795 4824 generic.go:334] "Generic (PLEG): container finished" podID="922cc36d-5fd8-4cba-b6e3-5b16e6cb7531" containerID="03ee038a6bf69c20fa36ea52fe0d3bec95eb293c2cc767dbb0d52a4f515cb1e2" exitCode=0 Dec 11 02:21:31 crc kubenswrapper[4824]: I1211 02:21:31.255914 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-wdzl5" event={"ID":"922cc36d-5fd8-4cba-b6e3-5b16e6cb7531","Type":"ContainerDied","Data":"03ee038a6bf69c20fa36ea52fe0d3bec95eb293c2cc767dbb0d52a4f515cb1e2"} Dec 11 02:21:31 crc kubenswrapper[4824]: I1211 02:21:31.957449 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-74f6bcbc87-4tss6" Dec 11 02:21:32 crc kubenswrapper[4824]: I1211 02:21:32.045278 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-5j6cw"] Dec 11 02:21:32 crc kubenswrapper[4824]: I1211 02:21:32.045630 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-698758b865-5j6cw" podUID="342737ab-7fa3-4654-9493-e826dacbaa6f" containerName="dnsmasq-dns" containerID="cri-o://60884c2e4f625ce1fc400158671d70c7dab41b7db11c1f8de1113467ae4206e9" gracePeriod=10 Dec 11 02:21:32 crc kubenswrapper[4824]: I1211 02:21:32.266953 4824 generic.go:334] "Generic (PLEG): container finished" podID="342737ab-7fa3-4654-9493-e826dacbaa6f" containerID="60884c2e4f625ce1fc400158671d70c7dab41b7db11c1f8de1113467ae4206e9" exitCode=0 Dec 11 02:21:32 crc kubenswrapper[4824]: I1211 02:21:32.267045 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-5j6cw" event={"ID":"342737ab-7fa3-4654-9493-e826dacbaa6f","Type":"ContainerDied","Data":"60884c2e4f625ce1fc400158671d70c7dab41b7db11c1f8de1113467ae4206e9"} Dec 11 02:21:32 crc kubenswrapper[4824]: I1211 02:21:32.518258 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-5j6cw" Dec 11 02:21:32 crc kubenswrapper[4824]: I1211 02:21:32.599648 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/342737ab-7fa3-4654-9493-e826dacbaa6f-ovsdbserver-sb\") pod \"342737ab-7fa3-4654-9493-e826dacbaa6f\" (UID: \"342737ab-7fa3-4654-9493-e826dacbaa6f\") " Dec 11 02:21:32 crc kubenswrapper[4824]: I1211 02:21:32.599728 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/342737ab-7fa3-4654-9493-e826dacbaa6f-dns-svc\") pod \"342737ab-7fa3-4654-9493-e826dacbaa6f\" (UID: \"342737ab-7fa3-4654-9493-e826dacbaa6f\") " Dec 11 02:21:32 crc kubenswrapper[4824]: I1211 02:21:32.599751 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/342737ab-7fa3-4654-9493-e826dacbaa6f-config\") pod \"342737ab-7fa3-4654-9493-e826dacbaa6f\" (UID: \"342737ab-7fa3-4654-9493-e826dacbaa6f\") " Dec 11 02:21:32 crc kubenswrapper[4824]: I1211 02:21:32.599785 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkx8k\" (UniqueName: \"kubernetes.io/projected/342737ab-7fa3-4654-9493-e826dacbaa6f-kube-api-access-zkx8k\") pod \"342737ab-7fa3-4654-9493-e826dacbaa6f\" (UID: \"342737ab-7fa3-4654-9493-e826dacbaa6f\") " Dec 11 02:21:32 crc kubenswrapper[4824]: I1211 02:21:32.599800 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/342737ab-7fa3-4654-9493-e826dacbaa6f-ovsdbserver-nb\") pod \"342737ab-7fa3-4654-9493-e826dacbaa6f\" (UID: \"342737ab-7fa3-4654-9493-e826dacbaa6f\") " Dec 11 02:21:32 crc kubenswrapper[4824]: I1211 02:21:32.608930 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/342737ab-7fa3-4654-9493-e826dacbaa6f-kube-api-access-zkx8k" (OuterVolumeSpecName: "kube-api-access-zkx8k") pod "342737ab-7fa3-4654-9493-e826dacbaa6f" (UID: "342737ab-7fa3-4654-9493-e826dacbaa6f"). InnerVolumeSpecName "kube-api-access-zkx8k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:21:32 crc kubenswrapper[4824]: I1211 02:21:32.644019 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/342737ab-7fa3-4654-9493-e826dacbaa6f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "342737ab-7fa3-4654-9493-e826dacbaa6f" (UID: "342737ab-7fa3-4654-9493-e826dacbaa6f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:21:32 crc kubenswrapper[4824]: I1211 02:21:32.648508 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/342737ab-7fa3-4654-9493-e826dacbaa6f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "342737ab-7fa3-4654-9493-e826dacbaa6f" (UID: "342737ab-7fa3-4654-9493-e826dacbaa6f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:21:32 crc kubenswrapper[4824]: I1211 02:21:32.655067 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-wdzl5" Dec 11 02:21:32 crc kubenswrapper[4824]: I1211 02:21:32.671607 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/342737ab-7fa3-4654-9493-e826dacbaa6f-config" (OuterVolumeSpecName: "config") pod "342737ab-7fa3-4654-9493-e826dacbaa6f" (UID: "342737ab-7fa3-4654-9493-e826dacbaa6f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:21:32 crc kubenswrapper[4824]: I1211 02:21:32.679826 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/342737ab-7fa3-4654-9493-e826dacbaa6f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "342737ab-7fa3-4654-9493-e826dacbaa6f" (UID: "342737ab-7fa3-4654-9493-e826dacbaa6f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:21:32 crc kubenswrapper[4824]: I1211 02:21:32.701349 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d84f5\" (UniqueName: \"kubernetes.io/projected/922cc36d-5fd8-4cba-b6e3-5b16e6cb7531-kube-api-access-d84f5\") pod \"922cc36d-5fd8-4cba-b6e3-5b16e6cb7531\" (UID: \"922cc36d-5fd8-4cba-b6e3-5b16e6cb7531\") " Dec 11 02:21:32 crc kubenswrapper[4824]: I1211 02:21:32.701645 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/922cc36d-5fd8-4cba-b6e3-5b16e6cb7531-config-data\") pod \"922cc36d-5fd8-4cba-b6e3-5b16e6cb7531\" (UID: \"922cc36d-5fd8-4cba-b6e3-5b16e6cb7531\") " Dec 11 02:21:32 crc kubenswrapper[4824]: I1211 02:21:32.701692 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/922cc36d-5fd8-4cba-b6e3-5b16e6cb7531-combined-ca-bundle\") pod \"922cc36d-5fd8-4cba-b6e3-5b16e6cb7531\" (UID: \"922cc36d-5fd8-4cba-b6e3-5b16e6cb7531\") " Dec 11 02:21:32 crc kubenswrapper[4824]: I1211 02:21:32.702506 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/342737ab-7fa3-4654-9493-e826dacbaa6f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 11 02:21:32 crc kubenswrapper[4824]: I1211 02:21:32.702546 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/342737ab-7fa3-4654-9493-e826dacbaa6f-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 11 02:21:32 crc kubenswrapper[4824]: I1211 02:21:32.702565 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/342737ab-7fa3-4654-9493-e826dacbaa6f-config\") on node \"crc\" DevicePath \"\"" Dec 11 02:21:32 crc kubenswrapper[4824]: I1211 02:21:32.702582 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkx8k\" (UniqueName: \"kubernetes.io/projected/342737ab-7fa3-4654-9493-e826dacbaa6f-kube-api-access-zkx8k\") on node \"crc\" DevicePath \"\"" Dec 11 02:21:32 crc kubenswrapper[4824]: I1211 02:21:32.702601 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/342737ab-7fa3-4654-9493-e826dacbaa6f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 11 02:21:32 crc kubenswrapper[4824]: I1211 02:21:32.705673 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/922cc36d-5fd8-4cba-b6e3-5b16e6cb7531-kube-api-access-d84f5" (OuterVolumeSpecName: "kube-api-access-d84f5") pod "922cc36d-5fd8-4cba-b6e3-5b16e6cb7531" (UID: "922cc36d-5fd8-4cba-b6e3-5b16e6cb7531"). InnerVolumeSpecName "kube-api-access-d84f5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:21:32 crc kubenswrapper[4824]: I1211 02:21:32.731673 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/922cc36d-5fd8-4cba-b6e3-5b16e6cb7531-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "922cc36d-5fd8-4cba-b6e3-5b16e6cb7531" (UID: "922cc36d-5fd8-4cba-b6e3-5b16e6cb7531"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:21:32 crc kubenswrapper[4824]: I1211 02:21:32.748269 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/922cc36d-5fd8-4cba-b6e3-5b16e6cb7531-config-data" (OuterVolumeSpecName: "config-data") pod "922cc36d-5fd8-4cba-b6e3-5b16e6cb7531" (UID: "922cc36d-5fd8-4cba-b6e3-5b16e6cb7531"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:21:32 crc kubenswrapper[4824]: I1211 02:21:32.803955 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/922cc36d-5fd8-4cba-b6e3-5b16e6cb7531-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 02:21:32 crc kubenswrapper[4824]: I1211 02:21:32.804018 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/922cc36d-5fd8-4cba-b6e3-5b16e6cb7531-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 02:21:32 crc kubenswrapper[4824]: I1211 02:21:32.804041 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d84f5\" (UniqueName: \"kubernetes.io/projected/922cc36d-5fd8-4cba-b6e3-5b16e6cb7531-kube-api-access-d84f5\") on node \"crc\" DevicePath \"\"" Dec 11 02:21:33 crc kubenswrapper[4824]: I1211 02:21:33.280555 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-5j6cw" event={"ID":"342737ab-7fa3-4654-9493-e826dacbaa6f","Type":"ContainerDied","Data":"f08525d1ef2fb0038ef806add621b4536a921ff2b14d914262cc47a0dc69773d"} Dec 11 02:21:33 crc kubenswrapper[4824]: I1211 02:21:33.280954 4824 scope.go:117] "RemoveContainer" containerID="60884c2e4f625ce1fc400158671d70c7dab41b7db11c1f8de1113467ae4206e9" Dec 11 02:21:33 crc kubenswrapper[4824]: I1211 02:21:33.280597 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-5j6cw" Dec 11 02:21:33 crc kubenswrapper[4824]: I1211 02:21:33.283685 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-wdzl5" event={"ID":"922cc36d-5fd8-4cba-b6e3-5b16e6cb7531","Type":"ContainerDied","Data":"83180a0899238d7258d72d33fd50a17fd8f1406bc4480383825cf4102b7dd049"} Dec 11 02:21:33 crc kubenswrapper[4824]: I1211 02:21:33.283724 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="83180a0899238d7258d72d33fd50a17fd8f1406bc4480383825cf4102b7dd049" Dec 11 02:21:33 crc kubenswrapper[4824]: I1211 02:21:33.283760 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-wdzl5" Dec 11 02:21:33 crc kubenswrapper[4824]: I1211 02:21:33.326038 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-5j6cw"] Dec 11 02:21:33 crc kubenswrapper[4824]: I1211 02:21:33.338783 4824 scope.go:117] "RemoveContainer" containerID="6e401da432977c68f6a09e8a58b770b58dccae00320a2be0b3ea4389be82ecb4" Dec 11 02:21:33 crc kubenswrapper[4824]: I1211 02:21:33.341745 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-698758b865-5j6cw"] Dec 11 02:21:33 crc kubenswrapper[4824]: I1211 02:21:33.458606 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-8xg6k"] Dec 11 02:21:33 crc kubenswrapper[4824]: E1211 02:21:33.458966 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ab730c6-95e4-438f-90d5-04812d26384f" containerName="mariadb-account-create-update" Dec 11 02:21:33 crc kubenswrapper[4824]: I1211 02:21:33.458989 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ab730c6-95e4-438f-90d5-04812d26384f" containerName="mariadb-account-create-update" Dec 11 02:21:33 crc kubenswrapper[4824]: E1211 02:21:33.459016 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="793e155e-73e9-46d5-9de6-a4aebf3ca2b4" containerName="mariadb-account-create-update" Dec 11 02:21:33 crc kubenswrapper[4824]: I1211 02:21:33.459024 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="793e155e-73e9-46d5-9de6-a4aebf3ca2b4" containerName="mariadb-account-create-update" Dec 11 02:21:33 crc kubenswrapper[4824]: E1211 02:21:33.459041 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20766632-27ff-42ab-9b3c-28aa8b8912bf" containerName="mariadb-database-create" Dec 11 02:21:33 crc kubenswrapper[4824]: I1211 02:21:33.459050 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="20766632-27ff-42ab-9b3c-28aa8b8912bf" containerName="mariadb-database-create" Dec 11 02:21:33 crc kubenswrapper[4824]: E1211 02:21:33.459064 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d9a4a8f-d58e-4a90-9c33-3ffa6a7f5c36" containerName="mariadb-account-create-update" Dec 11 02:21:33 crc kubenswrapper[4824]: I1211 02:21:33.459072 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d9a4a8f-d58e-4a90-9c33-3ffa6a7f5c36" containerName="mariadb-account-create-update" Dec 11 02:21:33 crc kubenswrapper[4824]: E1211 02:21:33.459092 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57686bda-9117-48cc-b29d-f5ff374937c9" containerName="mariadb-database-create" Dec 11 02:21:33 crc kubenswrapper[4824]: I1211 02:21:33.459100 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="57686bda-9117-48cc-b29d-f5ff374937c9" containerName="mariadb-database-create" Dec 11 02:21:33 crc kubenswrapper[4824]: E1211 02:21:33.464701 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f30f409e-2171-4b62-8e1b-e72ea450ec8d" containerName="dnsmasq-dns" Dec 11 02:21:33 crc kubenswrapper[4824]: I1211 02:21:33.464745 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f30f409e-2171-4b62-8e1b-e72ea450ec8d" containerName="dnsmasq-dns" Dec 11 02:21:33 crc kubenswrapper[4824]: E1211 02:21:33.464768 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="342737ab-7fa3-4654-9493-e826dacbaa6f" containerName="dnsmasq-dns" Dec 11 02:21:33 crc kubenswrapper[4824]: I1211 02:21:33.464778 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="342737ab-7fa3-4654-9493-e826dacbaa6f" containerName="dnsmasq-dns" Dec 11 02:21:33 crc kubenswrapper[4824]: E1211 02:21:33.464790 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="342737ab-7fa3-4654-9493-e826dacbaa6f" containerName="init" Dec 11 02:21:33 crc kubenswrapper[4824]: I1211 02:21:33.464797 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="342737ab-7fa3-4654-9493-e826dacbaa6f" containerName="init" Dec 11 02:21:33 crc kubenswrapper[4824]: E1211 02:21:33.464812 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bcb322d-a528-4d31-b639-367f5f7f3dd1" containerName="mariadb-database-create" Dec 11 02:21:33 crc kubenswrapper[4824]: I1211 02:21:33.464819 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bcb322d-a528-4d31-b639-367f5f7f3dd1" containerName="mariadb-database-create" Dec 11 02:21:33 crc kubenswrapper[4824]: E1211 02:21:33.464830 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f30f409e-2171-4b62-8e1b-e72ea450ec8d" containerName="init" Dec 11 02:21:33 crc kubenswrapper[4824]: I1211 02:21:33.464837 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f30f409e-2171-4b62-8e1b-e72ea450ec8d" containerName="init" Dec 11 02:21:33 crc kubenswrapper[4824]: E1211 02:21:33.464855 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="922cc36d-5fd8-4cba-b6e3-5b16e6cb7531" containerName="keystone-db-sync" Dec 11 02:21:33 crc kubenswrapper[4824]: I1211 02:21:33.464865 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="922cc36d-5fd8-4cba-b6e3-5b16e6cb7531" containerName="keystone-db-sync" Dec 11 02:21:33 crc kubenswrapper[4824]: I1211 02:21:33.465220 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="342737ab-7fa3-4654-9493-e826dacbaa6f" containerName="dnsmasq-dns" Dec 11 02:21:33 crc kubenswrapper[4824]: I1211 02:21:33.465246 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="57686bda-9117-48cc-b29d-f5ff374937c9" containerName="mariadb-database-create" Dec 11 02:21:33 crc kubenswrapper[4824]: I1211 02:21:33.465261 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d9a4a8f-d58e-4a90-9c33-3ffa6a7f5c36" containerName="mariadb-account-create-update" Dec 11 02:21:33 crc kubenswrapper[4824]: I1211 02:21:33.465279 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ab730c6-95e4-438f-90d5-04812d26384f" containerName="mariadb-account-create-update" Dec 11 02:21:33 crc kubenswrapper[4824]: I1211 02:21:33.465299 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f30f409e-2171-4b62-8e1b-e72ea450ec8d" containerName="dnsmasq-dns" Dec 11 02:21:33 crc kubenswrapper[4824]: I1211 02:21:33.465316 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="793e155e-73e9-46d5-9de6-a4aebf3ca2b4" containerName="mariadb-account-create-update" Dec 11 02:21:33 crc kubenswrapper[4824]: I1211 02:21:33.465330 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="922cc36d-5fd8-4cba-b6e3-5b16e6cb7531" containerName="keystone-db-sync" Dec 11 02:21:33 crc kubenswrapper[4824]: I1211 02:21:33.465340 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bcb322d-a528-4d31-b639-367f5f7f3dd1" containerName="mariadb-database-create" Dec 11 02:21:33 crc kubenswrapper[4824]: I1211 02:21:33.465349 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="20766632-27ff-42ab-9b3c-28aa8b8912bf" containerName="mariadb-database-create" Dec 11 02:21:33 crc kubenswrapper[4824]: I1211 02:21:33.466401 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-8xg6k" Dec 11 02:21:33 crc kubenswrapper[4824]: I1211 02:21:33.492013 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-8xg6k"] Dec 11 02:21:33 crc kubenswrapper[4824]: I1211 02:21:33.515038 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d8fffe44-22d1-4108-95e9-088db7c342c7-dns-svc\") pod \"dnsmasq-dns-847c4cc679-8xg6k\" (UID: \"d8fffe44-22d1-4108-95e9-088db7c342c7\") " pod="openstack/dnsmasq-dns-847c4cc679-8xg6k" Dec 11 02:21:33 crc kubenswrapper[4824]: I1211 02:21:33.515078 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d8fffe44-22d1-4108-95e9-088db7c342c7-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-8xg6k\" (UID: \"d8fffe44-22d1-4108-95e9-088db7c342c7\") " pod="openstack/dnsmasq-dns-847c4cc679-8xg6k" Dec 11 02:21:33 crc kubenswrapper[4824]: I1211 02:21:33.515126 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8fffe44-22d1-4108-95e9-088db7c342c7-config\") pod \"dnsmasq-dns-847c4cc679-8xg6k\" (UID: \"d8fffe44-22d1-4108-95e9-088db7c342c7\") " pod="openstack/dnsmasq-dns-847c4cc679-8xg6k" Dec 11 02:21:33 crc kubenswrapper[4824]: I1211 02:21:33.515223 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d8fffe44-22d1-4108-95e9-088db7c342c7-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-8xg6k\" (UID: \"d8fffe44-22d1-4108-95e9-088db7c342c7\") " pod="openstack/dnsmasq-dns-847c4cc679-8xg6k" Dec 11 02:21:33 crc kubenswrapper[4824]: I1211 02:21:33.515276 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6mxth\" (UniqueName: \"kubernetes.io/projected/d8fffe44-22d1-4108-95e9-088db7c342c7-kube-api-access-6mxth\") pod \"dnsmasq-dns-847c4cc679-8xg6k\" (UID: \"d8fffe44-22d1-4108-95e9-088db7c342c7\") " pod="openstack/dnsmasq-dns-847c4cc679-8xg6k" Dec 11 02:21:33 crc kubenswrapper[4824]: I1211 02:21:33.515297 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d8fffe44-22d1-4108-95e9-088db7c342c7-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-8xg6k\" (UID: \"d8fffe44-22d1-4108-95e9-088db7c342c7\") " pod="openstack/dnsmasq-dns-847c4cc679-8xg6k" Dec 11 02:21:33 crc kubenswrapper[4824]: I1211 02:21:33.528407 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-l45dd"] Dec 11 02:21:33 crc kubenswrapper[4824]: I1211 02:21:33.529303 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-l45dd" Dec 11 02:21:33 crc kubenswrapper[4824]: I1211 02:21:33.536046 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 11 02:21:33 crc kubenswrapper[4824]: I1211 02:21:33.536194 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 11 02:21:33 crc kubenswrapper[4824]: I1211 02:21:33.536418 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 11 02:21:33 crc kubenswrapper[4824]: I1211 02:21:33.536565 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 11 02:21:33 crc kubenswrapper[4824]: I1211 02:21:33.536878 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-8rv79" Dec 11 02:21:33 crc kubenswrapper[4824]: I1211 02:21:33.544047 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-l45dd"] Dec 11 02:21:33 crc kubenswrapper[4824]: I1211 02:21:33.618926 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f-fernet-keys\") pod \"keystone-bootstrap-l45dd\" (UID: \"daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f\") " pod="openstack/keystone-bootstrap-l45dd" Dec 11 02:21:33 crc kubenswrapper[4824]: I1211 02:21:33.618981 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d8fffe44-22d1-4108-95e9-088db7c342c7-dns-svc\") pod \"dnsmasq-dns-847c4cc679-8xg6k\" (UID: \"d8fffe44-22d1-4108-95e9-088db7c342c7\") " pod="openstack/dnsmasq-dns-847c4cc679-8xg6k" Dec 11 02:21:33 crc kubenswrapper[4824]: I1211 02:21:33.619006 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d8fffe44-22d1-4108-95e9-088db7c342c7-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-8xg6k\" (UID: \"d8fffe44-22d1-4108-95e9-088db7c342c7\") " pod="openstack/dnsmasq-dns-847c4cc679-8xg6k" Dec 11 02:21:33 crc kubenswrapper[4824]: I1211 02:21:33.619031 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8fffe44-22d1-4108-95e9-088db7c342c7-config\") pod \"dnsmasq-dns-847c4cc679-8xg6k\" (UID: \"d8fffe44-22d1-4108-95e9-088db7c342c7\") " pod="openstack/dnsmasq-dns-847c4cc679-8xg6k" Dec 11 02:21:33 crc kubenswrapper[4824]: I1211 02:21:33.619083 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d8fffe44-22d1-4108-95e9-088db7c342c7-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-8xg6k\" (UID: \"d8fffe44-22d1-4108-95e9-088db7c342c7\") " pod="openstack/dnsmasq-dns-847c4cc679-8xg6k" Dec 11 02:21:33 crc kubenswrapper[4824]: I1211 02:21:33.619121 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f-scripts\") pod \"keystone-bootstrap-l45dd\" (UID: \"daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f\") " pod="openstack/keystone-bootstrap-l45dd" Dec 11 02:21:33 crc kubenswrapper[4824]: I1211 02:21:33.619138 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f-config-data\") pod \"keystone-bootstrap-l45dd\" (UID: \"daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f\") " pod="openstack/keystone-bootstrap-l45dd" Dec 11 02:21:33 crc kubenswrapper[4824]: I1211 02:21:33.619154 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f-combined-ca-bundle\") pod \"keystone-bootstrap-l45dd\" (UID: \"daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f\") " pod="openstack/keystone-bootstrap-l45dd" Dec 11 02:21:33 crc kubenswrapper[4824]: I1211 02:21:33.619174 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mp5wm\" (UniqueName: \"kubernetes.io/projected/daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f-kube-api-access-mp5wm\") pod \"keystone-bootstrap-l45dd\" (UID: \"daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f\") " pod="openstack/keystone-bootstrap-l45dd" Dec 11 02:21:33 crc kubenswrapper[4824]: I1211 02:21:33.619209 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6mxth\" (UniqueName: \"kubernetes.io/projected/d8fffe44-22d1-4108-95e9-088db7c342c7-kube-api-access-6mxth\") pod \"dnsmasq-dns-847c4cc679-8xg6k\" (UID: \"d8fffe44-22d1-4108-95e9-088db7c342c7\") " pod="openstack/dnsmasq-dns-847c4cc679-8xg6k" Dec 11 02:21:33 crc kubenswrapper[4824]: I1211 02:21:33.619228 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d8fffe44-22d1-4108-95e9-088db7c342c7-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-8xg6k\" (UID: \"d8fffe44-22d1-4108-95e9-088db7c342c7\") " pod="openstack/dnsmasq-dns-847c4cc679-8xg6k" Dec 11 02:21:33 crc kubenswrapper[4824]: I1211 02:21:33.619265 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f-credential-keys\") pod \"keystone-bootstrap-l45dd\" (UID: \"daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f\") " pod="openstack/keystone-bootstrap-l45dd" Dec 11 02:21:33 crc kubenswrapper[4824]: I1211 02:21:33.620148 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d8fffe44-22d1-4108-95e9-088db7c342c7-dns-svc\") pod \"dnsmasq-dns-847c4cc679-8xg6k\" (UID: \"d8fffe44-22d1-4108-95e9-088db7c342c7\") " pod="openstack/dnsmasq-dns-847c4cc679-8xg6k" Dec 11 02:21:33 crc kubenswrapper[4824]: I1211 02:21:33.620706 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d8fffe44-22d1-4108-95e9-088db7c342c7-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-8xg6k\" (UID: \"d8fffe44-22d1-4108-95e9-088db7c342c7\") " pod="openstack/dnsmasq-dns-847c4cc679-8xg6k" Dec 11 02:21:33 crc kubenswrapper[4824]: I1211 02:21:33.621217 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8fffe44-22d1-4108-95e9-088db7c342c7-config\") pod \"dnsmasq-dns-847c4cc679-8xg6k\" (UID: \"d8fffe44-22d1-4108-95e9-088db7c342c7\") " pod="openstack/dnsmasq-dns-847c4cc679-8xg6k" Dec 11 02:21:33 crc kubenswrapper[4824]: I1211 02:21:33.621699 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d8fffe44-22d1-4108-95e9-088db7c342c7-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-8xg6k\" (UID: \"d8fffe44-22d1-4108-95e9-088db7c342c7\") " pod="openstack/dnsmasq-dns-847c4cc679-8xg6k" Dec 11 02:21:33 crc kubenswrapper[4824]: I1211 02:21:33.622469 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d8fffe44-22d1-4108-95e9-088db7c342c7-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-8xg6k\" (UID: \"d8fffe44-22d1-4108-95e9-088db7c342c7\") " pod="openstack/dnsmasq-dns-847c4cc679-8xg6k" Dec 11 02:21:33 crc kubenswrapper[4824]: I1211 02:21:33.661139 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6mxth\" (UniqueName: \"kubernetes.io/projected/d8fffe44-22d1-4108-95e9-088db7c342c7-kube-api-access-6mxth\") pod \"dnsmasq-dns-847c4cc679-8xg6k\" (UID: \"d8fffe44-22d1-4108-95e9-088db7c342c7\") " pod="openstack/dnsmasq-dns-847c4cc679-8xg6k" Dec 11 02:21:33 crc kubenswrapper[4824]: I1211 02:21:33.720998 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f-credential-keys\") pod \"keystone-bootstrap-l45dd\" (UID: \"daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f\") " pod="openstack/keystone-bootstrap-l45dd" Dec 11 02:21:33 crc kubenswrapper[4824]: I1211 02:21:33.721049 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f-fernet-keys\") pod \"keystone-bootstrap-l45dd\" (UID: \"daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f\") " pod="openstack/keystone-bootstrap-l45dd" Dec 11 02:21:33 crc kubenswrapper[4824]: I1211 02:21:33.721160 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f-scripts\") pod \"keystone-bootstrap-l45dd\" (UID: \"daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f\") " pod="openstack/keystone-bootstrap-l45dd" Dec 11 02:21:33 crc kubenswrapper[4824]: I1211 02:21:33.721178 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f-config-data\") pod \"keystone-bootstrap-l45dd\" (UID: \"daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f\") " pod="openstack/keystone-bootstrap-l45dd" Dec 11 02:21:33 crc kubenswrapper[4824]: I1211 02:21:33.721196 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f-combined-ca-bundle\") pod \"keystone-bootstrap-l45dd\" (UID: \"daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f\") " pod="openstack/keystone-bootstrap-l45dd" Dec 11 02:21:33 crc kubenswrapper[4824]: I1211 02:21:33.721216 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mp5wm\" (UniqueName: \"kubernetes.io/projected/daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f-kube-api-access-mp5wm\") pod \"keystone-bootstrap-l45dd\" (UID: \"daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f\") " pod="openstack/keystone-bootstrap-l45dd" Dec 11 02:21:33 crc kubenswrapper[4824]: I1211 02:21:33.728780 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f-credential-keys\") pod \"keystone-bootstrap-l45dd\" (UID: \"daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f\") " pod="openstack/keystone-bootstrap-l45dd" Dec 11 02:21:33 crc kubenswrapper[4824]: I1211 02:21:33.734736 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f-fernet-keys\") pod \"keystone-bootstrap-l45dd\" (UID: \"daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f\") " pod="openstack/keystone-bootstrap-l45dd" Dec 11 02:21:33 crc kubenswrapper[4824]: I1211 02:21:33.734787 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f-scripts\") pod \"keystone-bootstrap-l45dd\" (UID: \"daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f\") " pod="openstack/keystone-bootstrap-l45dd" Dec 11 02:21:33 crc kubenswrapper[4824]: I1211 02:21:33.735409 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f-config-data\") pod \"keystone-bootstrap-l45dd\" (UID: \"daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f\") " pod="openstack/keystone-bootstrap-l45dd" Dec 11 02:21:33 crc kubenswrapper[4824]: I1211 02:21:33.735770 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f-combined-ca-bundle\") pod \"keystone-bootstrap-l45dd\" (UID: \"daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f\") " pod="openstack/keystone-bootstrap-l45dd" Dec 11 02:21:33 crc kubenswrapper[4824]: I1211 02:21:33.786830 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-8xg6k" Dec 11 02:21:33 crc kubenswrapper[4824]: I1211 02:21:33.834271 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mp5wm\" (UniqueName: \"kubernetes.io/projected/daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f-kube-api-access-mp5wm\") pod \"keystone-bootstrap-l45dd\" (UID: \"daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f\") " pod="openstack/keystone-bootstrap-l45dd" Dec 11 02:21:33 crc kubenswrapper[4824]: I1211 02:21:33.852446 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-l45dd" Dec 11 02:21:33 crc kubenswrapper[4824]: I1211 02:21:33.951968 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7656dcd5fc-pvdtd"] Dec 11 02:21:33 crc kubenswrapper[4824]: I1211 02:21:33.961398 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7656dcd5fc-pvdtd" Dec 11 02:21:33 crc kubenswrapper[4824]: I1211 02:21:33.985491 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Dec 11 02:21:33 crc kubenswrapper[4824]: I1211 02:21:33.985955 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-xvxxc" Dec 11 02:21:33 crc kubenswrapper[4824]: I1211 02:21:33.986097 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Dec 11 02:21:33 crc kubenswrapper[4824]: I1211 02:21:33.986257 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.006418 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-nm5bg"] Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.007448 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-nm5bg" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.026293 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.027311 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fcst2\" (UniqueName: \"kubernetes.io/projected/97c61c1d-8fc0-46f3-9bb0-583a8c4aba1f-kube-api-access-fcst2\") pod \"horizon-7656dcd5fc-pvdtd\" (UID: \"97c61c1d-8fc0-46f3-9bb0-583a8c4aba1f\") " pod="openstack/horizon-7656dcd5fc-pvdtd" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.027374 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/97c61c1d-8fc0-46f3-9bb0-583a8c4aba1f-scripts\") pod \"horizon-7656dcd5fc-pvdtd\" (UID: \"97c61c1d-8fc0-46f3-9bb0-583a8c4aba1f\") " pod="openstack/horizon-7656dcd5fc-pvdtd" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.027457 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/97c61c1d-8fc0-46f3-9bb0-583a8c4aba1f-logs\") pod \"horizon-7656dcd5fc-pvdtd\" (UID: \"97c61c1d-8fc0-46f3-9bb0-583a8c4aba1f\") " pod="openstack/horizon-7656dcd5fc-pvdtd" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.027488 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/97c61c1d-8fc0-46f3-9bb0-583a8c4aba1f-horizon-secret-key\") pod \"horizon-7656dcd5fc-pvdtd\" (UID: \"97c61c1d-8fc0-46f3-9bb0-583a8c4aba1f\") " pod="openstack/horizon-7656dcd5fc-pvdtd" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.027509 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/97c61c1d-8fc0-46f3-9bb0-583a8c4aba1f-config-data\") pod \"horizon-7656dcd5fc-pvdtd\" (UID: \"97c61c1d-8fc0-46f3-9bb0-583a8c4aba1f\") " pod="openstack/horizon-7656dcd5fc-pvdtd" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.027887 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.028004 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7656dcd5fc-pvdtd"] Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.048030 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-nm5bg"] Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.067424 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-lpwl8"] Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.068713 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-lpwl8" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.078573 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.078651 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-b7xg9" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.078787 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-c5jks" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.082931 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.095288 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-tn658"] Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.096735 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-tn658" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.100223 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.100453 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.100579 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-xsjrn" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.110787 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-lpwl8"] Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.141707 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fcst2\" (UniqueName: \"kubernetes.io/projected/97c61c1d-8fc0-46f3-9bb0-583a8c4aba1f-kube-api-access-fcst2\") pod \"horizon-7656dcd5fc-pvdtd\" (UID: \"97c61c1d-8fc0-46f3-9bb0-583a8c4aba1f\") " pod="openstack/horizon-7656dcd5fc-pvdtd" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.141764 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07b76bc0-c213-45d8-bebd-06b3e70322ab-combined-ca-bundle\") pod \"neutron-db-sync-nm5bg\" (UID: \"07b76bc0-c213-45d8-bebd-06b3e70322ab\") " pod="openstack/neutron-db-sync-nm5bg" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.141794 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/97c61c1d-8fc0-46f3-9bb0-583a8c4aba1f-scripts\") pod \"horizon-7656dcd5fc-pvdtd\" (UID: \"97c61c1d-8fc0-46f3-9bb0-583a8c4aba1f\") " pod="openstack/horizon-7656dcd5fc-pvdtd" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.141831 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/07b76bc0-c213-45d8-bebd-06b3e70322ab-config\") pod \"neutron-db-sync-nm5bg\" (UID: \"07b76bc0-c213-45d8-bebd-06b3e70322ab\") " pod="openstack/neutron-db-sync-nm5bg" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.141854 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4jdvp\" (UniqueName: \"kubernetes.io/projected/07b76bc0-c213-45d8-bebd-06b3e70322ab-kube-api-access-4jdvp\") pod \"neutron-db-sync-nm5bg\" (UID: \"07b76bc0-c213-45d8-bebd-06b3e70322ab\") " pod="openstack/neutron-db-sync-nm5bg" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.141877 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7-db-sync-config-data\") pod \"cinder-db-sync-tn658\" (UID: \"8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7\") " pod="openstack/cinder-db-sync-tn658" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.141894 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7-combined-ca-bundle\") pod \"cinder-db-sync-tn658\" (UID: \"8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7\") " pod="openstack/cinder-db-sync-tn658" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.141909 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n22sp\" (UniqueName: \"kubernetes.io/projected/68cc858a-9296-4a34-8322-b2dabc8ac1a2-kube-api-access-n22sp\") pod \"placement-db-sync-lpwl8\" (UID: \"68cc858a-9296-4a34-8322-b2dabc8ac1a2\") " pod="openstack/placement-db-sync-lpwl8" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.141931 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68cc858a-9296-4a34-8322-b2dabc8ac1a2-combined-ca-bundle\") pod \"placement-db-sync-lpwl8\" (UID: \"68cc858a-9296-4a34-8322-b2dabc8ac1a2\") " pod="openstack/placement-db-sync-lpwl8" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.141951 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68cc858a-9296-4a34-8322-b2dabc8ac1a2-config-data\") pod \"placement-db-sync-lpwl8\" (UID: \"68cc858a-9296-4a34-8322-b2dabc8ac1a2\") " pod="openstack/placement-db-sync-lpwl8" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.141994 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/97c61c1d-8fc0-46f3-9bb0-583a8c4aba1f-logs\") pod \"horizon-7656dcd5fc-pvdtd\" (UID: \"97c61c1d-8fc0-46f3-9bb0-583a8c4aba1f\") " pod="openstack/horizon-7656dcd5fc-pvdtd" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.142018 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vgswv\" (UniqueName: \"kubernetes.io/projected/8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7-kube-api-access-vgswv\") pod \"cinder-db-sync-tn658\" (UID: \"8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7\") " pod="openstack/cinder-db-sync-tn658" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.142043 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/97c61c1d-8fc0-46f3-9bb0-583a8c4aba1f-horizon-secret-key\") pod \"horizon-7656dcd5fc-pvdtd\" (UID: \"97c61c1d-8fc0-46f3-9bb0-583a8c4aba1f\") " pod="openstack/horizon-7656dcd5fc-pvdtd" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.142068 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/97c61c1d-8fc0-46f3-9bb0-583a8c4aba1f-config-data\") pod \"horizon-7656dcd5fc-pvdtd\" (UID: \"97c61c1d-8fc0-46f3-9bb0-583a8c4aba1f\") " pod="openstack/horizon-7656dcd5fc-pvdtd" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.142084 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/68cc858a-9296-4a34-8322-b2dabc8ac1a2-logs\") pod \"placement-db-sync-lpwl8\" (UID: \"68cc858a-9296-4a34-8322-b2dabc8ac1a2\") " pod="openstack/placement-db-sync-lpwl8" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.142103 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/68cc858a-9296-4a34-8322-b2dabc8ac1a2-scripts\") pod \"placement-db-sync-lpwl8\" (UID: \"68cc858a-9296-4a34-8322-b2dabc8ac1a2\") " pod="openstack/placement-db-sync-lpwl8" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.142135 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7-etc-machine-id\") pod \"cinder-db-sync-tn658\" (UID: \"8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7\") " pod="openstack/cinder-db-sync-tn658" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.142580 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/97c61c1d-8fc0-46f3-9bb0-583a8c4aba1f-logs\") pod \"horizon-7656dcd5fc-pvdtd\" (UID: \"97c61c1d-8fc0-46f3-9bb0-583a8c4aba1f\") " pod="openstack/horizon-7656dcd5fc-pvdtd" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.143000 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/97c61c1d-8fc0-46f3-9bb0-583a8c4aba1f-scripts\") pod \"horizon-7656dcd5fc-pvdtd\" (UID: \"97c61c1d-8fc0-46f3-9bb0-583a8c4aba1f\") " pod="openstack/horizon-7656dcd5fc-pvdtd" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.143228 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7-scripts\") pod \"cinder-db-sync-tn658\" (UID: \"8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7\") " pod="openstack/cinder-db-sync-tn658" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.143249 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7-config-data\") pod \"cinder-db-sync-tn658\" (UID: \"8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7\") " pod="openstack/cinder-db-sync-tn658" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.144204 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/97c61c1d-8fc0-46f3-9bb0-583a8c4aba1f-config-data\") pod \"horizon-7656dcd5fc-pvdtd\" (UID: \"97c61c1d-8fc0-46f3-9bb0-583a8c4aba1f\") " pod="openstack/horizon-7656dcd5fc-pvdtd" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.153693 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/97c61c1d-8fc0-46f3-9bb0-583a8c4aba1f-horizon-secret-key\") pod \"horizon-7656dcd5fc-pvdtd\" (UID: \"97c61c1d-8fc0-46f3-9bb0-583a8c4aba1f\") " pod="openstack/horizon-7656dcd5fc-pvdtd" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.153761 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-tn658"] Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.173170 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-8xg6k"] Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.202064 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fcst2\" (UniqueName: \"kubernetes.io/projected/97c61c1d-8fc0-46f3-9bb0-583a8c4aba1f-kube-api-access-fcst2\") pod \"horizon-7656dcd5fc-pvdtd\" (UID: \"97c61c1d-8fc0-46f3-9bb0-583a8c4aba1f\") " pod="openstack/horizon-7656dcd5fc-pvdtd" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.217181 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-mxg84"] Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.226991 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-mxg84" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.245691 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07b76bc0-c213-45d8-bebd-06b3e70322ab-combined-ca-bundle\") pod \"neutron-db-sync-nm5bg\" (UID: \"07b76bc0-c213-45d8-bebd-06b3e70322ab\") " pod="openstack/neutron-db-sync-nm5bg" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.245758 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/07b76bc0-c213-45d8-bebd-06b3e70322ab-config\") pod \"neutron-db-sync-nm5bg\" (UID: \"07b76bc0-c213-45d8-bebd-06b3e70322ab\") " pod="openstack/neutron-db-sync-nm5bg" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.245782 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4jdvp\" (UniqueName: \"kubernetes.io/projected/07b76bc0-c213-45d8-bebd-06b3e70322ab-kube-api-access-4jdvp\") pod \"neutron-db-sync-nm5bg\" (UID: \"07b76bc0-c213-45d8-bebd-06b3e70322ab\") " pod="openstack/neutron-db-sync-nm5bg" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.245807 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7-db-sync-config-data\") pod \"cinder-db-sync-tn658\" (UID: \"8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7\") " pod="openstack/cinder-db-sync-tn658" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.245832 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7-combined-ca-bundle\") pod \"cinder-db-sync-tn658\" (UID: \"8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7\") " pod="openstack/cinder-db-sync-tn658" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.245852 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n22sp\" (UniqueName: \"kubernetes.io/projected/68cc858a-9296-4a34-8322-b2dabc8ac1a2-kube-api-access-n22sp\") pod \"placement-db-sync-lpwl8\" (UID: \"68cc858a-9296-4a34-8322-b2dabc8ac1a2\") " pod="openstack/placement-db-sync-lpwl8" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.245907 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68cc858a-9296-4a34-8322-b2dabc8ac1a2-combined-ca-bundle\") pod \"placement-db-sync-lpwl8\" (UID: \"68cc858a-9296-4a34-8322-b2dabc8ac1a2\") " pod="openstack/placement-db-sync-lpwl8" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.245932 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68cc858a-9296-4a34-8322-b2dabc8ac1a2-config-data\") pod \"placement-db-sync-lpwl8\" (UID: \"68cc858a-9296-4a34-8322-b2dabc8ac1a2\") " pod="openstack/placement-db-sync-lpwl8" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.245970 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vgswv\" (UniqueName: \"kubernetes.io/projected/8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7-kube-api-access-vgswv\") pod \"cinder-db-sync-tn658\" (UID: \"8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7\") " pod="openstack/cinder-db-sync-tn658" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.246011 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/68cc858a-9296-4a34-8322-b2dabc8ac1a2-logs\") pod \"placement-db-sync-lpwl8\" (UID: \"68cc858a-9296-4a34-8322-b2dabc8ac1a2\") " pod="openstack/placement-db-sync-lpwl8" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.246039 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/68cc858a-9296-4a34-8322-b2dabc8ac1a2-scripts\") pod \"placement-db-sync-lpwl8\" (UID: \"68cc858a-9296-4a34-8322-b2dabc8ac1a2\") " pod="openstack/placement-db-sync-lpwl8" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.246064 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7-etc-machine-id\") pod \"cinder-db-sync-tn658\" (UID: \"8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7\") " pod="openstack/cinder-db-sync-tn658" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.246091 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7-config-data\") pod \"cinder-db-sync-tn658\" (UID: \"8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7\") " pod="openstack/cinder-db-sync-tn658" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.246123 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7-scripts\") pod \"cinder-db-sync-tn658\" (UID: \"8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7\") " pod="openstack/cinder-db-sync-tn658" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.260027 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-mxg84"] Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.264093 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7-etc-machine-id\") pod \"cinder-db-sync-tn658\" (UID: \"8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7\") " pod="openstack/cinder-db-sync-tn658" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.264870 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/68cc858a-9296-4a34-8322-b2dabc8ac1a2-logs\") pod \"placement-db-sync-lpwl8\" (UID: \"68cc858a-9296-4a34-8322-b2dabc8ac1a2\") " pod="openstack/placement-db-sync-lpwl8" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.272813 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68cc858a-9296-4a34-8322-b2dabc8ac1a2-config-data\") pod \"placement-db-sync-lpwl8\" (UID: \"68cc858a-9296-4a34-8322-b2dabc8ac1a2\") " pod="openstack/placement-db-sync-lpwl8" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.279490 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7-config-data\") pod \"cinder-db-sync-tn658\" (UID: \"8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7\") " pod="openstack/cinder-db-sync-tn658" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.285460 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7-scripts\") pod \"cinder-db-sync-tn658\" (UID: \"8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7\") " pod="openstack/cinder-db-sync-tn658" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.285800 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68cc858a-9296-4a34-8322-b2dabc8ac1a2-combined-ca-bundle\") pod \"placement-db-sync-lpwl8\" (UID: \"68cc858a-9296-4a34-8322-b2dabc8ac1a2\") " pod="openstack/placement-db-sync-lpwl8" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.296876 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07b76bc0-c213-45d8-bebd-06b3e70322ab-combined-ca-bundle\") pod \"neutron-db-sync-nm5bg\" (UID: \"07b76bc0-c213-45d8-bebd-06b3e70322ab\") " pod="openstack/neutron-db-sync-nm5bg" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.302522 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/68cc858a-9296-4a34-8322-b2dabc8ac1a2-scripts\") pod \"placement-db-sync-lpwl8\" (UID: \"68cc858a-9296-4a34-8322-b2dabc8ac1a2\") " pod="openstack/placement-db-sync-lpwl8" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.303052 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7-combined-ca-bundle\") pod \"cinder-db-sync-tn658\" (UID: \"8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7\") " pod="openstack/cinder-db-sync-tn658" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.304281 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/07b76bc0-c213-45d8-bebd-06b3e70322ab-config\") pod \"neutron-db-sync-nm5bg\" (UID: \"07b76bc0-c213-45d8-bebd-06b3e70322ab\") " pod="openstack/neutron-db-sync-nm5bg" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.312984 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-55cc49576f-sct4f"] Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.325792 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-55cc49576f-sct4f" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.347441 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7656dcd5fc-pvdtd" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.391567 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7-db-sync-config-data\") pod \"cinder-db-sync-tn658\" (UID: \"8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7\") " pod="openstack/cinder-db-sync-tn658" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.392200 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4jdvp\" (UniqueName: \"kubernetes.io/projected/07b76bc0-c213-45d8-bebd-06b3e70322ab-kube-api-access-4jdvp\") pod \"neutron-db-sync-nm5bg\" (UID: \"07b76bc0-c213-45d8-bebd-06b3e70322ab\") " pod="openstack/neutron-db-sync-nm5bg" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.392353 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n22sp\" (UniqueName: \"kubernetes.io/projected/68cc858a-9296-4a34-8322-b2dabc8ac1a2-kube-api-access-n22sp\") pod \"placement-db-sync-lpwl8\" (UID: \"68cc858a-9296-4a34-8322-b2dabc8ac1a2\") " pod="openstack/placement-db-sync-lpwl8" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.393418 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vgswv\" (UniqueName: \"kubernetes.io/projected/8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7-kube-api-access-vgswv\") pod \"cinder-db-sync-tn658\" (UID: \"8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7\") " pod="openstack/cinder-db-sync-tn658" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.410316 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-nm5bg" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.472444 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-55cc49576f-sct4f"] Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.477599 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-lpwl8" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.480322 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-tn658" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.480725 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a36c6797-0e71-4dd1-b2ae-fb540f63e794-config\") pod \"dnsmasq-dns-785d8bcb8c-mxg84\" (UID: \"a36c6797-0e71-4dd1-b2ae-fb540f63e794\") " pod="openstack/dnsmasq-dns-785d8bcb8c-mxg84" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.480801 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a36c6797-0e71-4dd1-b2ae-fb540f63e794-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-mxg84\" (UID: \"a36c6797-0e71-4dd1-b2ae-fb540f63e794\") " pod="openstack/dnsmasq-dns-785d8bcb8c-mxg84" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.480870 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a36c6797-0e71-4dd1-b2ae-fb540f63e794-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-mxg84\" (UID: \"a36c6797-0e71-4dd1-b2ae-fb540f63e794\") " pod="openstack/dnsmasq-dns-785d8bcb8c-mxg84" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.480912 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/539eb117-6cac-4c4c-abe6-7840021f5233-scripts\") pod \"horizon-55cc49576f-sct4f\" (UID: \"539eb117-6cac-4c4c-abe6-7840021f5233\") " pod="openstack/horizon-55cc49576f-sct4f" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.481037 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a36c6797-0e71-4dd1-b2ae-fb540f63e794-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-mxg84\" (UID: \"a36c6797-0e71-4dd1-b2ae-fb540f63e794\") " pod="openstack/dnsmasq-dns-785d8bcb8c-mxg84" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.481063 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/539eb117-6cac-4c4c-abe6-7840021f5233-logs\") pod \"horizon-55cc49576f-sct4f\" (UID: \"539eb117-6cac-4c4c-abe6-7840021f5233\") " pod="openstack/horizon-55cc49576f-sct4f" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.481146 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/539eb117-6cac-4c4c-abe6-7840021f5233-horizon-secret-key\") pod \"horizon-55cc49576f-sct4f\" (UID: \"539eb117-6cac-4c4c-abe6-7840021f5233\") " pod="openstack/horizon-55cc49576f-sct4f" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.481181 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xrq8w\" (UniqueName: \"kubernetes.io/projected/a36c6797-0e71-4dd1-b2ae-fb540f63e794-kube-api-access-xrq8w\") pod \"dnsmasq-dns-785d8bcb8c-mxg84\" (UID: \"a36c6797-0e71-4dd1-b2ae-fb540f63e794\") " pod="openstack/dnsmasq-dns-785d8bcb8c-mxg84" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.481218 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a36c6797-0e71-4dd1-b2ae-fb540f63e794-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-mxg84\" (UID: \"a36c6797-0e71-4dd1-b2ae-fb540f63e794\") " pod="openstack/dnsmasq-dns-785d8bcb8c-mxg84" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.481248 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x7b6b\" (UniqueName: \"kubernetes.io/projected/539eb117-6cac-4c4c-abe6-7840021f5233-kube-api-access-x7b6b\") pod \"horizon-55cc49576f-sct4f\" (UID: \"539eb117-6cac-4c4c-abe6-7840021f5233\") " pod="openstack/horizon-55cc49576f-sct4f" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.481304 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/539eb117-6cac-4c4c-abe6-7840021f5233-config-data\") pod \"horizon-55cc49576f-sct4f\" (UID: \"539eb117-6cac-4c4c-abe6-7840021f5233\") " pod="openstack/horizon-55cc49576f-sct4f" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.498532 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-pbsrf"] Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.499569 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-pbsrf" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.510243 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.510432 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-bdr5s" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.534593 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.536495 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.540036 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-pbsrf"] Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.541571 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.541750 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.541861 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-5plrn" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.541981 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.575456 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.582103 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a36c6797-0e71-4dd1-b2ae-fb540f63e794-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-mxg84\" (UID: \"a36c6797-0e71-4dd1-b2ae-fb540f63e794\") " pod="openstack/dnsmasq-dns-785d8bcb8c-mxg84" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.582155 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a36c6797-0e71-4dd1-b2ae-fb540f63e794-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-mxg84\" (UID: \"a36c6797-0e71-4dd1-b2ae-fb540f63e794\") " pod="openstack/dnsmasq-dns-785d8bcb8c-mxg84" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.582183 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/539eb117-6cac-4c4c-abe6-7840021f5233-scripts\") pod \"horizon-55cc49576f-sct4f\" (UID: \"539eb117-6cac-4c4c-abe6-7840021f5233\") " pod="openstack/horizon-55cc49576f-sct4f" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.582207 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a36c6797-0e71-4dd1-b2ae-fb540f63e794-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-mxg84\" (UID: \"a36c6797-0e71-4dd1-b2ae-fb540f63e794\") " pod="openstack/dnsmasq-dns-785d8bcb8c-mxg84" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.582223 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/539eb117-6cac-4c4c-abe6-7840021f5233-logs\") pod \"horizon-55cc49576f-sct4f\" (UID: \"539eb117-6cac-4c4c-abe6-7840021f5233\") " pod="openstack/horizon-55cc49576f-sct4f" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.582266 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/539eb117-6cac-4c4c-abe6-7840021f5233-horizon-secret-key\") pod \"horizon-55cc49576f-sct4f\" (UID: \"539eb117-6cac-4c4c-abe6-7840021f5233\") " pod="openstack/horizon-55cc49576f-sct4f" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.582288 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xrq8w\" (UniqueName: \"kubernetes.io/projected/a36c6797-0e71-4dd1-b2ae-fb540f63e794-kube-api-access-xrq8w\") pod \"dnsmasq-dns-785d8bcb8c-mxg84\" (UID: \"a36c6797-0e71-4dd1-b2ae-fb540f63e794\") " pod="openstack/dnsmasq-dns-785d8bcb8c-mxg84" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.582316 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a36c6797-0e71-4dd1-b2ae-fb540f63e794-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-mxg84\" (UID: \"a36c6797-0e71-4dd1-b2ae-fb540f63e794\") " pod="openstack/dnsmasq-dns-785d8bcb8c-mxg84" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.582343 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x7b6b\" (UniqueName: \"kubernetes.io/projected/539eb117-6cac-4c4c-abe6-7840021f5233-kube-api-access-x7b6b\") pod \"horizon-55cc49576f-sct4f\" (UID: \"539eb117-6cac-4c4c-abe6-7840021f5233\") " pod="openstack/horizon-55cc49576f-sct4f" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.582378 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/539eb117-6cac-4c4c-abe6-7840021f5233-config-data\") pod \"horizon-55cc49576f-sct4f\" (UID: \"539eb117-6cac-4c4c-abe6-7840021f5233\") " pod="openstack/horizon-55cc49576f-sct4f" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.582399 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a36c6797-0e71-4dd1-b2ae-fb540f63e794-config\") pod \"dnsmasq-dns-785d8bcb8c-mxg84\" (UID: \"a36c6797-0e71-4dd1-b2ae-fb540f63e794\") " pod="openstack/dnsmasq-dns-785d8bcb8c-mxg84" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.583146 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a36c6797-0e71-4dd1-b2ae-fb540f63e794-config\") pod \"dnsmasq-dns-785d8bcb8c-mxg84\" (UID: \"a36c6797-0e71-4dd1-b2ae-fb540f63e794\") " pod="openstack/dnsmasq-dns-785d8bcb8c-mxg84" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.583631 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a36c6797-0e71-4dd1-b2ae-fb540f63e794-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-mxg84\" (UID: \"a36c6797-0e71-4dd1-b2ae-fb540f63e794\") " pod="openstack/dnsmasq-dns-785d8bcb8c-mxg84" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.586323 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a36c6797-0e71-4dd1-b2ae-fb540f63e794-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-mxg84\" (UID: \"a36c6797-0e71-4dd1-b2ae-fb540f63e794\") " pod="openstack/dnsmasq-dns-785d8bcb8c-mxg84" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.589451 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/539eb117-6cac-4c4c-abe6-7840021f5233-scripts\") pod \"horizon-55cc49576f-sct4f\" (UID: \"539eb117-6cac-4c4c-abe6-7840021f5233\") " pod="openstack/horizon-55cc49576f-sct4f" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.590056 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a36c6797-0e71-4dd1-b2ae-fb540f63e794-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-mxg84\" (UID: \"a36c6797-0e71-4dd1-b2ae-fb540f63e794\") " pod="openstack/dnsmasq-dns-785d8bcb8c-mxg84" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.590282 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/539eb117-6cac-4c4c-abe6-7840021f5233-logs\") pod \"horizon-55cc49576f-sct4f\" (UID: \"539eb117-6cac-4c4c-abe6-7840021f5233\") " pod="openstack/horizon-55cc49576f-sct4f" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.590993 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a36c6797-0e71-4dd1-b2ae-fb540f63e794-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-mxg84\" (UID: \"a36c6797-0e71-4dd1-b2ae-fb540f63e794\") " pod="openstack/dnsmasq-dns-785d8bcb8c-mxg84" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.592014 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/539eb117-6cac-4c4c-abe6-7840021f5233-config-data\") pod \"horizon-55cc49576f-sct4f\" (UID: \"539eb117-6cac-4c4c-abe6-7840021f5233\") " pod="openstack/horizon-55cc49576f-sct4f" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.597369 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/539eb117-6cac-4c4c-abe6-7840021f5233-horizon-secret-key\") pod \"horizon-55cc49576f-sct4f\" (UID: \"539eb117-6cac-4c4c-abe6-7840021f5233\") " pod="openstack/horizon-55cc49576f-sct4f" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.607354 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.612126 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.612685 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.614652 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.614971 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.617341 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x7b6b\" (UniqueName: \"kubernetes.io/projected/539eb117-6cac-4c4c-abe6-7840021f5233-kube-api-access-x7b6b\") pod \"horizon-55cc49576f-sct4f\" (UID: \"539eb117-6cac-4c4c-abe6-7840021f5233\") " pod="openstack/horizon-55cc49576f-sct4f" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.621465 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.621851 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xrq8w\" (UniqueName: \"kubernetes.io/projected/a36c6797-0e71-4dd1-b2ae-fb540f63e794-kube-api-access-xrq8w\") pod \"dnsmasq-dns-785d8bcb8c-mxg84\" (UID: \"a36c6797-0e71-4dd1-b2ae-fb540f63e794\") " pod="openstack/dnsmasq-dns-785d8bcb8c-mxg84" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.630393 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.637566 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.643874 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.662496 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="342737ab-7fa3-4654-9493-e826dacbaa6f" path="/var/lib/kubelet/pods/342737ab-7fa3-4654-9493-e826dacbaa6f/volumes" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.664277 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.684969 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"6d658bd2-e722-433e-825c-b5f4068c7ae2\") " pod="openstack/glance-default-external-api-0" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.685207 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d658bd2-e722-433e-825c-b5f4068c7ae2-config-data\") pod \"glance-default-external-api-0\" (UID: \"6d658bd2-e722-433e-825c-b5f4068c7ae2\") " pod="openstack/glance-default-external-api-0" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.685245 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6d658bd2-e722-433e-825c-b5f4068c7ae2-scripts\") pod \"glance-default-external-api-0\" (UID: \"6d658bd2-e722-433e-825c-b5f4068c7ae2\") " pod="openstack/glance-default-external-api-0" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.685324 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jh9z2\" (UniqueName: \"kubernetes.io/projected/d536a1b1-baa2-497a-ba0c-ad03599168cc-kube-api-access-jh9z2\") pod \"barbican-db-sync-pbsrf\" (UID: \"d536a1b1-baa2-497a-ba0c-ad03599168cc\") " pod="openstack/barbican-db-sync-pbsrf" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.685349 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6d658bd2-e722-433e-825c-b5f4068c7ae2-logs\") pod \"glance-default-external-api-0\" (UID: \"6d658bd2-e722-433e-825c-b5f4068c7ae2\") " pod="openstack/glance-default-external-api-0" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.685390 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d536a1b1-baa2-497a-ba0c-ad03599168cc-db-sync-config-data\") pod \"barbican-db-sync-pbsrf\" (UID: \"d536a1b1-baa2-497a-ba0c-ad03599168cc\") " pod="openstack/barbican-db-sync-pbsrf" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.685451 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6d658bd2-e722-433e-825c-b5f4068c7ae2-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"6d658bd2-e722-433e-825c-b5f4068c7ae2\") " pod="openstack/glance-default-external-api-0" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.685547 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v92r7\" (UniqueName: \"kubernetes.io/projected/6d658bd2-e722-433e-825c-b5f4068c7ae2-kube-api-access-v92r7\") pod \"glance-default-external-api-0\" (UID: \"6d658bd2-e722-433e-825c-b5f4068c7ae2\") " pod="openstack/glance-default-external-api-0" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.685608 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d658bd2-e722-433e-825c-b5f4068c7ae2-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"6d658bd2-e722-433e-825c-b5f4068c7ae2\") " pod="openstack/glance-default-external-api-0" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.685653 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d536a1b1-baa2-497a-ba0c-ad03599168cc-combined-ca-bundle\") pod \"barbican-db-sync-pbsrf\" (UID: \"d536a1b1-baa2-497a-ba0c-ad03599168cc\") " pod="openstack/barbican-db-sync-pbsrf" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.685717 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d658bd2-e722-433e-825c-b5f4068c7ae2-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"6d658bd2-e722-433e-825c-b5f4068c7ae2\") " pod="openstack/glance-default-external-api-0" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.755740 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-8xg6k"] Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.787361 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d658bd2-e722-433e-825c-b5f4068c7ae2-config-data\") pod \"glance-default-external-api-0\" (UID: \"6d658bd2-e722-433e-825c-b5f4068c7ae2\") " pod="openstack/glance-default-external-api-0" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.787406 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6d658bd2-e722-433e-825c-b5f4068c7ae2-scripts\") pod \"glance-default-external-api-0\" (UID: \"6d658bd2-e722-433e-825c-b5f4068c7ae2\") " pod="openstack/glance-default-external-api-0" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.787435 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5bd5bd9b-501f-43e4-88da-7b4f5c591b58-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5bd5bd9b-501f-43e4-88da-7b4f5c591b58\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.787453 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5bd5bd9b-501f-43e4-88da-7b4f5c591b58-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5bd5bd9b-501f-43e4-88da-7b4f5c591b58\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.787474 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"5bd5bd9b-501f-43e4-88da-7b4f5c591b58\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.787493 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bd5bd9b-501f-43e4-88da-7b4f5c591b58-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5bd5bd9b-501f-43e4-88da-7b4f5c591b58\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.787515 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6d658bd2-e722-433e-825c-b5f4068c7ae2-logs\") pod \"glance-default-external-api-0\" (UID: \"6d658bd2-e722-433e-825c-b5f4068c7ae2\") " pod="openstack/glance-default-external-api-0" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.787530 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jh9z2\" (UniqueName: \"kubernetes.io/projected/d536a1b1-baa2-497a-ba0c-ad03599168cc-kube-api-access-jh9z2\") pod \"barbican-db-sync-pbsrf\" (UID: \"d536a1b1-baa2-497a-ba0c-ad03599168cc\") " pod="openstack/barbican-db-sync-pbsrf" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.787550 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/efa62db6-8875-4ae6-99ea-007e08be02bd-scripts\") pod \"ceilometer-0\" (UID: \"efa62db6-8875-4ae6-99ea-007e08be02bd\") " pod="openstack/ceilometer-0" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.787577 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d536a1b1-baa2-497a-ba0c-ad03599168cc-db-sync-config-data\") pod \"barbican-db-sync-pbsrf\" (UID: \"d536a1b1-baa2-497a-ba0c-ad03599168cc\") " pod="openstack/barbican-db-sync-pbsrf" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.787598 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5bd5bd9b-501f-43e4-88da-7b4f5c591b58-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"5bd5bd9b-501f-43e4-88da-7b4f5c591b58\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.787628 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6d658bd2-e722-433e-825c-b5f4068c7ae2-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"6d658bd2-e722-433e-825c-b5f4068c7ae2\") " pod="openstack/glance-default-external-api-0" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.787659 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/efa62db6-8875-4ae6-99ea-007e08be02bd-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"efa62db6-8875-4ae6-99ea-007e08be02bd\") " pod="openstack/ceilometer-0" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.787676 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v92r7\" (UniqueName: \"kubernetes.io/projected/6d658bd2-e722-433e-825c-b5f4068c7ae2-kube-api-access-v92r7\") pod \"glance-default-external-api-0\" (UID: \"6d658bd2-e722-433e-825c-b5f4068c7ae2\") " pod="openstack/glance-default-external-api-0" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.787695 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/efa62db6-8875-4ae6-99ea-007e08be02bd-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"efa62db6-8875-4ae6-99ea-007e08be02bd\") " pod="openstack/ceilometer-0" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.787710 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5bd5bd9b-501f-43e4-88da-7b4f5c591b58-logs\") pod \"glance-default-internal-api-0\" (UID: \"5bd5bd9b-501f-43e4-88da-7b4f5c591b58\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.787733 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d658bd2-e722-433e-825c-b5f4068c7ae2-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"6d658bd2-e722-433e-825c-b5f4068c7ae2\") " pod="openstack/glance-default-external-api-0" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.787760 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cqww2\" (UniqueName: \"kubernetes.io/projected/efa62db6-8875-4ae6-99ea-007e08be02bd-kube-api-access-cqww2\") pod \"ceilometer-0\" (UID: \"efa62db6-8875-4ae6-99ea-007e08be02bd\") " pod="openstack/ceilometer-0" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.787779 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d536a1b1-baa2-497a-ba0c-ad03599168cc-combined-ca-bundle\") pod \"barbican-db-sync-pbsrf\" (UID: \"d536a1b1-baa2-497a-ba0c-ad03599168cc\") " pod="openstack/barbican-db-sync-pbsrf" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.787798 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-88rkb\" (UniqueName: \"kubernetes.io/projected/5bd5bd9b-501f-43e4-88da-7b4f5c591b58-kube-api-access-88rkb\") pod \"glance-default-internal-api-0\" (UID: \"5bd5bd9b-501f-43e4-88da-7b4f5c591b58\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.787828 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d658bd2-e722-433e-825c-b5f4068c7ae2-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"6d658bd2-e722-433e-825c-b5f4068c7ae2\") " pod="openstack/glance-default-external-api-0" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.787849 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/efa62db6-8875-4ae6-99ea-007e08be02bd-log-httpd\") pod \"ceilometer-0\" (UID: \"efa62db6-8875-4ae6-99ea-007e08be02bd\") " pod="openstack/ceilometer-0" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.787873 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/efa62db6-8875-4ae6-99ea-007e08be02bd-run-httpd\") pod \"ceilometer-0\" (UID: \"efa62db6-8875-4ae6-99ea-007e08be02bd\") " pod="openstack/ceilometer-0" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.787904 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5bd5bd9b-501f-43e4-88da-7b4f5c591b58-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5bd5bd9b-501f-43e4-88da-7b4f5c591b58\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.787921 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/efa62db6-8875-4ae6-99ea-007e08be02bd-config-data\") pod \"ceilometer-0\" (UID: \"efa62db6-8875-4ae6-99ea-007e08be02bd\") " pod="openstack/ceilometer-0" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.787937 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"6d658bd2-e722-433e-825c-b5f4068c7ae2\") " pod="openstack/glance-default-external-api-0" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.788245 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"6d658bd2-e722-433e-825c-b5f4068c7ae2\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/glance-default-external-api-0" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.791050 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6d658bd2-e722-433e-825c-b5f4068c7ae2-logs\") pod \"glance-default-external-api-0\" (UID: \"6d658bd2-e722-433e-825c-b5f4068c7ae2\") " pod="openstack/glance-default-external-api-0" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.792467 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6d658bd2-e722-433e-825c-b5f4068c7ae2-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"6d658bd2-e722-433e-825c-b5f4068c7ae2\") " pod="openstack/glance-default-external-api-0" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.798615 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6d658bd2-e722-433e-825c-b5f4068c7ae2-scripts\") pod \"glance-default-external-api-0\" (UID: \"6d658bd2-e722-433e-825c-b5f4068c7ae2\") " pod="openstack/glance-default-external-api-0" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.799728 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d536a1b1-baa2-497a-ba0c-ad03599168cc-combined-ca-bundle\") pod \"barbican-db-sync-pbsrf\" (UID: \"d536a1b1-baa2-497a-ba0c-ad03599168cc\") " pod="openstack/barbican-db-sync-pbsrf" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.800164 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-mxg84" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.800841 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d536a1b1-baa2-497a-ba0c-ad03599168cc-db-sync-config-data\") pod \"barbican-db-sync-pbsrf\" (UID: \"d536a1b1-baa2-497a-ba0c-ad03599168cc\") " pod="openstack/barbican-db-sync-pbsrf" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.808396 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d658bd2-e722-433e-825c-b5f4068c7ae2-config-data\") pod \"glance-default-external-api-0\" (UID: \"6d658bd2-e722-433e-825c-b5f4068c7ae2\") " pod="openstack/glance-default-external-api-0" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.816387 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v92r7\" (UniqueName: \"kubernetes.io/projected/6d658bd2-e722-433e-825c-b5f4068c7ae2-kube-api-access-v92r7\") pod \"glance-default-external-api-0\" (UID: \"6d658bd2-e722-433e-825c-b5f4068c7ae2\") " pod="openstack/glance-default-external-api-0" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.817410 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d658bd2-e722-433e-825c-b5f4068c7ae2-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"6d658bd2-e722-433e-825c-b5f4068c7ae2\") " pod="openstack/glance-default-external-api-0" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.820390 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-55cc49576f-sct4f" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.820771 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d658bd2-e722-433e-825c-b5f4068c7ae2-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"6d658bd2-e722-433e-825c-b5f4068c7ae2\") " pod="openstack/glance-default-external-api-0" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.821825 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jh9z2\" (UniqueName: \"kubernetes.io/projected/d536a1b1-baa2-497a-ba0c-ad03599168cc-kube-api-access-jh9z2\") pod \"barbican-db-sync-pbsrf\" (UID: \"d536a1b1-baa2-497a-ba0c-ad03599168cc\") " pod="openstack/barbican-db-sync-pbsrf" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.834834 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"6d658bd2-e722-433e-825c-b5f4068c7ae2\") " pod="openstack/glance-default-external-api-0" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.845335 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-pbsrf" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.868208 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 11 02:21:34 crc kubenswrapper[4824]: W1211 02:21:34.870999 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd8fffe44_22d1_4108_95e9_088db7c342c7.slice/crio-332d003b20e747c04355340803a8cae24cfa7a91c8d897bb9f0ae15a50122344 WatchSource:0}: Error finding container 332d003b20e747c04355340803a8cae24cfa7a91c8d897bb9f0ae15a50122344: Status 404 returned error can't find the container with id 332d003b20e747c04355340803a8cae24cfa7a91c8d897bb9f0ae15a50122344 Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.889517 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5bd5bd9b-501f-43e4-88da-7b4f5c591b58-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5bd5bd9b-501f-43e4-88da-7b4f5c591b58\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.889566 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/efa62db6-8875-4ae6-99ea-007e08be02bd-config-data\") pod \"ceilometer-0\" (UID: \"efa62db6-8875-4ae6-99ea-007e08be02bd\") " pod="openstack/ceilometer-0" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.889612 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5bd5bd9b-501f-43e4-88da-7b4f5c591b58-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5bd5bd9b-501f-43e4-88da-7b4f5c591b58\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.889632 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5bd5bd9b-501f-43e4-88da-7b4f5c591b58-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5bd5bd9b-501f-43e4-88da-7b4f5c591b58\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.889652 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"5bd5bd9b-501f-43e4-88da-7b4f5c591b58\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.889668 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bd5bd9b-501f-43e4-88da-7b4f5c591b58-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5bd5bd9b-501f-43e4-88da-7b4f5c591b58\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.889691 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/efa62db6-8875-4ae6-99ea-007e08be02bd-scripts\") pod \"ceilometer-0\" (UID: \"efa62db6-8875-4ae6-99ea-007e08be02bd\") " pod="openstack/ceilometer-0" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.889722 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5bd5bd9b-501f-43e4-88da-7b4f5c591b58-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"5bd5bd9b-501f-43e4-88da-7b4f5c591b58\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.889746 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/efa62db6-8875-4ae6-99ea-007e08be02bd-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"efa62db6-8875-4ae6-99ea-007e08be02bd\") " pod="openstack/ceilometer-0" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.889768 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5bd5bd9b-501f-43e4-88da-7b4f5c591b58-logs\") pod \"glance-default-internal-api-0\" (UID: \"5bd5bd9b-501f-43e4-88da-7b4f5c591b58\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.889783 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/efa62db6-8875-4ae6-99ea-007e08be02bd-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"efa62db6-8875-4ae6-99ea-007e08be02bd\") " pod="openstack/ceilometer-0" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.889805 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqww2\" (UniqueName: \"kubernetes.io/projected/efa62db6-8875-4ae6-99ea-007e08be02bd-kube-api-access-cqww2\") pod \"ceilometer-0\" (UID: \"efa62db6-8875-4ae6-99ea-007e08be02bd\") " pod="openstack/ceilometer-0" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.889827 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-88rkb\" (UniqueName: \"kubernetes.io/projected/5bd5bd9b-501f-43e4-88da-7b4f5c591b58-kube-api-access-88rkb\") pod \"glance-default-internal-api-0\" (UID: \"5bd5bd9b-501f-43e4-88da-7b4f5c591b58\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.889862 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/efa62db6-8875-4ae6-99ea-007e08be02bd-log-httpd\") pod \"ceilometer-0\" (UID: \"efa62db6-8875-4ae6-99ea-007e08be02bd\") " pod="openstack/ceilometer-0" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.889886 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/efa62db6-8875-4ae6-99ea-007e08be02bd-run-httpd\") pod \"ceilometer-0\" (UID: \"efa62db6-8875-4ae6-99ea-007e08be02bd\") " pod="openstack/ceilometer-0" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.890368 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/efa62db6-8875-4ae6-99ea-007e08be02bd-run-httpd\") pod \"ceilometer-0\" (UID: \"efa62db6-8875-4ae6-99ea-007e08be02bd\") " pod="openstack/ceilometer-0" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.890676 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5bd5bd9b-501f-43e4-88da-7b4f5c591b58-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5bd5bd9b-501f-43e4-88da-7b4f5c591b58\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.891329 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/efa62db6-8875-4ae6-99ea-007e08be02bd-log-httpd\") pod \"ceilometer-0\" (UID: \"efa62db6-8875-4ae6-99ea-007e08be02bd\") " pod="openstack/ceilometer-0" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.891839 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5bd5bd9b-501f-43e4-88da-7b4f5c591b58-logs\") pod \"glance-default-internal-api-0\" (UID: \"5bd5bd9b-501f-43e4-88da-7b4f5c591b58\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.892237 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"5bd5bd9b-501f-43e4-88da-7b4f5c591b58\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-internal-api-0" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.896538 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5bd5bd9b-501f-43e4-88da-7b4f5c591b58-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5bd5bd9b-501f-43e4-88da-7b4f5c591b58\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.897486 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/efa62db6-8875-4ae6-99ea-007e08be02bd-config-data\") pod \"ceilometer-0\" (UID: \"efa62db6-8875-4ae6-99ea-007e08be02bd\") " pod="openstack/ceilometer-0" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.899831 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5bd5bd9b-501f-43e4-88da-7b4f5c591b58-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"5bd5bd9b-501f-43e4-88da-7b4f5c591b58\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.900799 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/efa62db6-8875-4ae6-99ea-007e08be02bd-scripts\") pod \"ceilometer-0\" (UID: \"efa62db6-8875-4ae6-99ea-007e08be02bd\") " pod="openstack/ceilometer-0" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.902020 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/efa62db6-8875-4ae6-99ea-007e08be02bd-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"efa62db6-8875-4ae6-99ea-007e08be02bd\") " pod="openstack/ceilometer-0" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.902297 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5bd5bd9b-501f-43e4-88da-7b4f5c591b58-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5bd5bd9b-501f-43e4-88da-7b4f5c591b58\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.905123 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bd5bd9b-501f-43e4-88da-7b4f5c591b58-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5bd5bd9b-501f-43e4-88da-7b4f5c591b58\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.909302 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqww2\" (UniqueName: \"kubernetes.io/projected/efa62db6-8875-4ae6-99ea-007e08be02bd-kube-api-access-cqww2\") pod \"ceilometer-0\" (UID: \"efa62db6-8875-4ae6-99ea-007e08be02bd\") " pod="openstack/ceilometer-0" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.909383 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-88rkb\" (UniqueName: \"kubernetes.io/projected/5bd5bd9b-501f-43e4-88da-7b4f5c591b58-kube-api-access-88rkb\") pod \"glance-default-internal-api-0\" (UID: \"5bd5bd9b-501f-43e4-88da-7b4f5c591b58\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.918645 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/efa62db6-8875-4ae6-99ea-007e08be02bd-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"efa62db6-8875-4ae6-99ea-007e08be02bd\") " pod="openstack/ceilometer-0" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.921514 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"5bd5bd9b-501f-43e4-88da-7b4f5c591b58\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.956641 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 11 02:21:34 crc kubenswrapper[4824]: I1211 02:21:34.994391 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 11 02:21:35 crc kubenswrapper[4824]: I1211 02:21:35.096163 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-l45dd"] Dec 11 02:21:35 crc kubenswrapper[4824]: I1211 02:21:35.176829 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7656dcd5fc-pvdtd"] Dec 11 02:21:35 crc kubenswrapper[4824]: I1211 02:21:35.188880 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-nm5bg"] Dec 11 02:21:35 crc kubenswrapper[4824]: I1211 02:21:35.390092 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-lpwl8"] Dec 11 02:21:35 crc kubenswrapper[4824]: I1211 02:21:35.394327 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-nm5bg" event={"ID":"07b76bc0-c213-45d8-bebd-06b3e70322ab","Type":"ContainerStarted","Data":"abbb752d7aa8e3149ae4ec2b431e8cdf4d21d3617f5f01e0405c39ecb5e0a48d"} Dec 11 02:21:35 crc kubenswrapper[4824]: I1211 02:21:35.405744 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-8xg6k" event={"ID":"d8fffe44-22d1-4108-95e9-088db7c342c7","Type":"ContainerStarted","Data":"332d003b20e747c04355340803a8cae24cfa7a91c8d897bb9f0ae15a50122344"} Dec 11 02:21:35 crc kubenswrapper[4824]: I1211 02:21:35.406804 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-l45dd" event={"ID":"daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f","Type":"ContainerStarted","Data":"2d3935a901d32fee9fab9c6f749b8abde6996842fb034062a8a60344334460df"} Dec 11 02:21:35 crc kubenswrapper[4824]: I1211 02:21:35.413742 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7656dcd5fc-pvdtd" event={"ID":"97c61c1d-8fc0-46f3-9bb0-583a8c4aba1f","Type":"ContainerStarted","Data":"3e28db81b0a324868b2b39a3b900d1a684da2db2d4786ac105c94fc9b19062f7"} Dec 11 02:21:35 crc kubenswrapper[4824]: I1211 02:21:35.585175 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-55cc49576f-sct4f"] Dec 11 02:21:35 crc kubenswrapper[4824]: I1211 02:21:35.613049 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-tn658"] Dec 11 02:21:35 crc kubenswrapper[4824]: I1211 02:21:35.713079 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-mxg84"] Dec 11 02:21:35 crc kubenswrapper[4824]: I1211 02:21:35.799180 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-pbsrf"] Dec 11 02:21:35 crc kubenswrapper[4824]: I1211 02:21:35.942616 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 11 02:21:35 crc kubenswrapper[4824]: I1211 02:21:35.966025 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 11 02:21:36 crc kubenswrapper[4824]: I1211 02:21:36.051622 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 11 02:21:36 crc kubenswrapper[4824]: I1211 02:21:36.083928 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7656dcd5fc-pvdtd"] Dec 11 02:21:36 crc kubenswrapper[4824]: I1211 02:21:36.131092 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 11 02:21:36 crc kubenswrapper[4824]: I1211 02:21:36.152209 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 11 02:21:36 crc kubenswrapper[4824]: I1211 02:21:36.182353 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-5f68f577dc-nj2zp"] Dec 11 02:21:36 crc kubenswrapper[4824]: I1211 02:21:36.183702 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5f68f577dc-nj2zp" Dec 11 02:21:36 crc kubenswrapper[4824]: I1211 02:21:36.218064 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 11 02:21:36 crc kubenswrapper[4824]: I1211 02:21:36.226485 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5f68f577dc-nj2zp"] Dec 11 02:21:36 crc kubenswrapper[4824]: I1211 02:21:36.348958 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rwmt9\" (UniqueName: \"kubernetes.io/projected/d7637730-463d-4742-a5fb-4ed6029edd61-kube-api-access-rwmt9\") pod \"horizon-5f68f577dc-nj2zp\" (UID: \"d7637730-463d-4742-a5fb-4ed6029edd61\") " pod="openstack/horizon-5f68f577dc-nj2zp" Dec 11 02:21:36 crc kubenswrapper[4824]: I1211 02:21:36.349015 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d7637730-463d-4742-a5fb-4ed6029edd61-scripts\") pod \"horizon-5f68f577dc-nj2zp\" (UID: \"d7637730-463d-4742-a5fb-4ed6029edd61\") " pod="openstack/horizon-5f68f577dc-nj2zp" Dec 11 02:21:36 crc kubenswrapper[4824]: I1211 02:21:36.349043 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d7637730-463d-4742-a5fb-4ed6029edd61-logs\") pod \"horizon-5f68f577dc-nj2zp\" (UID: \"d7637730-463d-4742-a5fb-4ed6029edd61\") " pod="openstack/horizon-5f68f577dc-nj2zp" Dec 11 02:21:36 crc kubenswrapper[4824]: I1211 02:21:36.349099 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d7637730-463d-4742-a5fb-4ed6029edd61-horizon-secret-key\") pod \"horizon-5f68f577dc-nj2zp\" (UID: \"d7637730-463d-4742-a5fb-4ed6029edd61\") " pod="openstack/horizon-5f68f577dc-nj2zp" Dec 11 02:21:36 crc kubenswrapper[4824]: I1211 02:21:36.349154 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d7637730-463d-4742-a5fb-4ed6029edd61-config-data\") pod \"horizon-5f68f577dc-nj2zp\" (UID: \"d7637730-463d-4742-a5fb-4ed6029edd61\") " pod="openstack/horizon-5f68f577dc-nj2zp" Dec 11 02:21:36 crc kubenswrapper[4824]: I1211 02:21:36.453016 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d7637730-463d-4742-a5fb-4ed6029edd61-logs\") pod \"horizon-5f68f577dc-nj2zp\" (UID: \"d7637730-463d-4742-a5fb-4ed6029edd61\") " pod="openstack/horizon-5f68f577dc-nj2zp" Dec 11 02:21:36 crc kubenswrapper[4824]: I1211 02:21:36.453122 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d7637730-463d-4742-a5fb-4ed6029edd61-horizon-secret-key\") pod \"horizon-5f68f577dc-nj2zp\" (UID: \"d7637730-463d-4742-a5fb-4ed6029edd61\") " pod="openstack/horizon-5f68f577dc-nj2zp" Dec 11 02:21:36 crc kubenswrapper[4824]: I1211 02:21:36.453166 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d7637730-463d-4742-a5fb-4ed6029edd61-config-data\") pod \"horizon-5f68f577dc-nj2zp\" (UID: \"d7637730-463d-4742-a5fb-4ed6029edd61\") " pod="openstack/horizon-5f68f577dc-nj2zp" Dec 11 02:21:36 crc kubenswrapper[4824]: I1211 02:21:36.453221 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rwmt9\" (UniqueName: \"kubernetes.io/projected/d7637730-463d-4742-a5fb-4ed6029edd61-kube-api-access-rwmt9\") pod \"horizon-5f68f577dc-nj2zp\" (UID: \"d7637730-463d-4742-a5fb-4ed6029edd61\") " pod="openstack/horizon-5f68f577dc-nj2zp" Dec 11 02:21:36 crc kubenswrapper[4824]: I1211 02:21:36.453253 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d7637730-463d-4742-a5fb-4ed6029edd61-scripts\") pod \"horizon-5f68f577dc-nj2zp\" (UID: \"d7637730-463d-4742-a5fb-4ed6029edd61\") " pod="openstack/horizon-5f68f577dc-nj2zp" Dec 11 02:21:36 crc kubenswrapper[4824]: I1211 02:21:36.453966 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d7637730-463d-4742-a5fb-4ed6029edd61-logs\") pod \"horizon-5f68f577dc-nj2zp\" (UID: \"d7637730-463d-4742-a5fb-4ed6029edd61\") " pod="openstack/horizon-5f68f577dc-nj2zp" Dec 11 02:21:36 crc kubenswrapper[4824]: I1211 02:21:36.455032 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d7637730-463d-4742-a5fb-4ed6029edd61-scripts\") pod \"horizon-5f68f577dc-nj2zp\" (UID: \"d7637730-463d-4742-a5fb-4ed6029edd61\") " pod="openstack/horizon-5f68f577dc-nj2zp" Dec 11 02:21:36 crc kubenswrapper[4824]: I1211 02:21:36.461097 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d7637730-463d-4742-a5fb-4ed6029edd61-horizon-secret-key\") pod \"horizon-5f68f577dc-nj2zp\" (UID: \"d7637730-463d-4742-a5fb-4ed6029edd61\") " pod="openstack/horizon-5f68f577dc-nj2zp" Dec 11 02:21:36 crc kubenswrapper[4824]: I1211 02:21:36.461723 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d7637730-463d-4742-a5fb-4ed6029edd61-config-data\") pod \"horizon-5f68f577dc-nj2zp\" (UID: \"d7637730-463d-4742-a5fb-4ed6029edd61\") " pod="openstack/horizon-5f68f577dc-nj2zp" Dec 11 02:21:36 crc kubenswrapper[4824]: I1211 02:21:36.468576 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"efa62db6-8875-4ae6-99ea-007e08be02bd","Type":"ContainerStarted","Data":"7f115c2b8b658ee4f32895ec61a442af614b8c9df49d6fcd4f69ec87080063f2"} Dec 11 02:21:36 crc kubenswrapper[4824]: I1211 02:21:36.476379 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6d658bd2-e722-433e-825c-b5f4068c7ae2","Type":"ContainerStarted","Data":"b563b0b55b6d8a1a5d784d6cb06c4c178f1c83027829be3a6820eb0ea5c05871"} Dec 11 02:21:36 crc kubenswrapper[4824]: I1211 02:21:36.476754 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rwmt9\" (UniqueName: \"kubernetes.io/projected/d7637730-463d-4742-a5fb-4ed6029edd61-kube-api-access-rwmt9\") pod \"horizon-5f68f577dc-nj2zp\" (UID: \"d7637730-463d-4742-a5fb-4ed6029edd61\") " pod="openstack/horizon-5f68f577dc-nj2zp" Dec 11 02:21:36 crc kubenswrapper[4824]: I1211 02:21:36.485257 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-pbsrf" event={"ID":"d536a1b1-baa2-497a-ba0c-ad03599168cc","Type":"ContainerStarted","Data":"74fb14189c9001a125a5b8284f126c27ce67a62e029fe6f80e77f8d47a0075b9"} Dec 11 02:21:36 crc kubenswrapper[4824]: I1211 02:21:36.490349 4824 generic.go:334] "Generic (PLEG): container finished" podID="d8fffe44-22d1-4108-95e9-088db7c342c7" containerID="50e6bcbeea57c3f998dd2fbdb1888e9a906eb3fbe8df7f6bbdc3d74d9a50108d" exitCode=0 Dec 11 02:21:36 crc kubenswrapper[4824]: I1211 02:21:36.490534 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-8xg6k" event={"ID":"d8fffe44-22d1-4108-95e9-088db7c342c7","Type":"ContainerDied","Data":"50e6bcbeea57c3f998dd2fbdb1888e9a906eb3fbe8df7f6bbdc3d74d9a50108d"} Dec 11 02:21:36 crc kubenswrapper[4824]: I1211 02:21:36.506369 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-55cc49576f-sct4f" event={"ID":"539eb117-6cac-4c4c-abe6-7840021f5233","Type":"ContainerStarted","Data":"346db3dfa2a85b752217e9c56567fb1b73235c430431b12b2bba32d0558c6dd6"} Dec 11 02:21:36 crc kubenswrapper[4824]: I1211 02:21:36.531312 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-nm5bg" event={"ID":"07b76bc0-c213-45d8-bebd-06b3e70322ab","Type":"ContainerStarted","Data":"d38997cd7f66152cf930c9deb794b8e30589e769a2b2d85afbc34bbe959d1c6b"} Dec 11 02:21:36 crc kubenswrapper[4824]: I1211 02:21:36.534032 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5bd5bd9b-501f-43e4-88da-7b4f5c591b58","Type":"ContainerStarted","Data":"bcacac34cf2cbf245a937445b260ede2ab3250b3c50923d970602e2f8303c42e"} Dec 11 02:21:36 crc kubenswrapper[4824]: I1211 02:21:36.542563 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-lpwl8" event={"ID":"68cc858a-9296-4a34-8322-b2dabc8ac1a2","Type":"ContainerStarted","Data":"06c8ade0aaed7f6fe86d287e1a891c7a424aa4e4c2af950bc4f08db03eb1c99d"} Dec 11 02:21:36 crc kubenswrapper[4824]: I1211 02:21:36.546239 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-tn658" event={"ID":"8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7","Type":"ContainerStarted","Data":"d84bd6f07a5f4326184457e98968c257a95a27bc22c12f9d7c4a5db2ad7953a7"} Dec 11 02:21:36 crc kubenswrapper[4824]: I1211 02:21:36.553342 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-l45dd" event={"ID":"daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f","Type":"ContainerStarted","Data":"3ad722591b153d578fa20c1ff377018b683efae930cebb1a49b0e62ef80c0355"} Dec 11 02:21:36 crc kubenswrapper[4824]: I1211 02:21:36.558433 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-nm5bg" podStartSLOduration=3.558419428 podStartE2EDuration="3.558419428s" podCreationTimestamp="2025-12-11 02:21:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:21:36.555063616 +0000 UTC m=+1238.244100995" watchObservedRunningTime="2025-12-11 02:21:36.558419428 +0000 UTC m=+1238.247456807" Dec 11 02:21:36 crc kubenswrapper[4824]: I1211 02:21:36.583203 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-l45dd" podStartSLOduration=3.583186757 podStartE2EDuration="3.583186757s" podCreationTimestamp="2025-12-11 02:21:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:21:36.581565203 +0000 UTC m=+1238.270602582" watchObservedRunningTime="2025-12-11 02:21:36.583186757 +0000 UTC m=+1238.272224126" Dec 11 02:21:36 crc kubenswrapper[4824]: I1211 02:21:36.586608 4824 generic.go:334] "Generic (PLEG): container finished" podID="a36c6797-0e71-4dd1-b2ae-fb540f63e794" containerID="634f409a880778b883a02c8949b0f0ac6c137e947d48e01593f6dc39b1311df4" exitCode=0 Dec 11 02:21:36 crc kubenswrapper[4824]: I1211 02:21:36.586657 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-mxg84" event={"ID":"a36c6797-0e71-4dd1-b2ae-fb540f63e794","Type":"ContainerDied","Data":"634f409a880778b883a02c8949b0f0ac6c137e947d48e01593f6dc39b1311df4"} Dec 11 02:21:36 crc kubenswrapper[4824]: I1211 02:21:36.586681 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-mxg84" event={"ID":"a36c6797-0e71-4dd1-b2ae-fb540f63e794","Type":"ContainerStarted","Data":"0acb45c451d715906cd7ce451913139c9c04e01e1d5c35766d7143e2e63cade4"} Dec 11 02:21:36 crc kubenswrapper[4824]: I1211 02:21:36.693232 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5f68f577dc-nj2zp" Dec 11 02:21:37 crc kubenswrapper[4824]: I1211 02:21:37.163025 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-8xg6k" Dec 11 02:21:37 crc kubenswrapper[4824]: I1211 02:21:37.270037 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8fffe44-22d1-4108-95e9-088db7c342c7-config\") pod \"d8fffe44-22d1-4108-95e9-088db7c342c7\" (UID: \"d8fffe44-22d1-4108-95e9-088db7c342c7\") " Dec 11 02:21:37 crc kubenswrapper[4824]: I1211 02:21:37.270094 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6mxth\" (UniqueName: \"kubernetes.io/projected/d8fffe44-22d1-4108-95e9-088db7c342c7-kube-api-access-6mxth\") pod \"d8fffe44-22d1-4108-95e9-088db7c342c7\" (UID: \"d8fffe44-22d1-4108-95e9-088db7c342c7\") " Dec 11 02:21:37 crc kubenswrapper[4824]: I1211 02:21:37.270190 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d8fffe44-22d1-4108-95e9-088db7c342c7-dns-swift-storage-0\") pod \"d8fffe44-22d1-4108-95e9-088db7c342c7\" (UID: \"d8fffe44-22d1-4108-95e9-088db7c342c7\") " Dec 11 02:21:37 crc kubenswrapper[4824]: I1211 02:21:37.270215 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d8fffe44-22d1-4108-95e9-088db7c342c7-ovsdbserver-nb\") pod \"d8fffe44-22d1-4108-95e9-088db7c342c7\" (UID: \"d8fffe44-22d1-4108-95e9-088db7c342c7\") " Dec 11 02:21:37 crc kubenswrapper[4824]: I1211 02:21:37.270265 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d8fffe44-22d1-4108-95e9-088db7c342c7-ovsdbserver-sb\") pod \"d8fffe44-22d1-4108-95e9-088db7c342c7\" (UID: \"d8fffe44-22d1-4108-95e9-088db7c342c7\") " Dec 11 02:21:37 crc kubenswrapper[4824]: I1211 02:21:37.270364 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d8fffe44-22d1-4108-95e9-088db7c342c7-dns-svc\") pod \"d8fffe44-22d1-4108-95e9-088db7c342c7\" (UID: \"d8fffe44-22d1-4108-95e9-088db7c342c7\") " Dec 11 02:21:37 crc kubenswrapper[4824]: I1211 02:21:37.295279 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d8fffe44-22d1-4108-95e9-088db7c342c7-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "d8fffe44-22d1-4108-95e9-088db7c342c7" (UID: "d8fffe44-22d1-4108-95e9-088db7c342c7"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:21:37 crc kubenswrapper[4824]: I1211 02:21:37.310240 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d8fffe44-22d1-4108-95e9-088db7c342c7-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d8fffe44-22d1-4108-95e9-088db7c342c7" (UID: "d8fffe44-22d1-4108-95e9-088db7c342c7"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:21:37 crc kubenswrapper[4824]: I1211 02:21:37.319768 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d8fffe44-22d1-4108-95e9-088db7c342c7-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d8fffe44-22d1-4108-95e9-088db7c342c7" (UID: "d8fffe44-22d1-4108-95e9-088db7c342c7"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:21:37 crc kubenswrapper[4824]: I1211 02:21:37.325513 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8fffe44-22d1-4108-95e9-088db7c342c7-kube-api-access-6mxth" (OuterVolumeSpecName: "kube-api-access-6mxth") pod "d8fffe44-22d1-4108-95e9-088db7c342c7" (UID: "d8fffe44-22d1-4108-95e9-088db7c342c7"). InnerVolumeSpecName "kube-api-access-6mxth". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:21:37 crc kubenswrapper[4824]: I1211 02:21:37.332642 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d8fffe44-22d1-4108-95e9-088db7c342c7-config" (OuterVolumeSpecName: "config") pod "d8fffe44-22d1-4108-95e9-088db7c342c7" (UID: "d8fffe44-22d1-4108-95e9-088db7c342c7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:21:37 crc kubenswrapper[4824]: I1211 02:21:37.349037 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d8fffe44-22d1-4108-95e9-088db7c342c7-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d8fffe44-22d1-4108-95e9-088db7c342c7" (UID: "d8fffe44-22d1-4108-95e9-088db7c342c7"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:21:37 crc kubenswrapper[4824]: I1211 02:21:37.385408 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d8fffe44-22d1-4108-95e9-088db7c342c7-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 11 02:21:37 crc kubenswrapper[4824]: I1211 02:21:37.385624 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d8fffe44-22d1-4108-95e9-088db7c342c7-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 11 02:21:37 crc kubenswrapper[4824]: I1211 02:21:37.385647 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8fffe44-22d1-4108-95e9-088db7c342c7-config\") on node \"crc\" DevicePath \"\"" Dec 11 02:21:37 crc kubenswrapper[4824]: I1211 02:21:37.385659 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6mxth\" (UniqueName: \"kubernetes.io/projected/d8fffe44-22d1-4108-95e9-088db7c342c7-kube-api-access-6mxth\") on node \"crc\" DevicePath \"\"" Dec 11 02:21:37 crc kubenswrapper[4824]: I1211 02:21:37.385678 4824 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d8fffe44-22d1-4108-95e9-088db7c342c7-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 11 02:21:37 crc kubenswrapper[4824]: I1211 02:21:37.385688 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d8fffe44-22d1-4108-95e9-088db7c342c7-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 11 02:21:37 crc kubenswrapper[4824]: I1211 02:21:37.399159 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5f68f577dc-nj2zp"] Dec 11 02:21:37 crc kubenswrapper[4824]: I1211 02:21:37.623974 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5f68f577dc-nj2zp" event={"ID":"d7637730-463d-4742-a5fb-4ed6029edd61","Type":"ContainerStarted","Data":"cd385f1cb3484d0e9da65069df64c3791209f38b68d60f06475a2e9df506d95f"} Dec 11 02:21:37 crc kubenswrapper[4824]: I1211 02:21:37.631373 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-mxg84" event={"ID":"a36c6797-0e71-4dd1-b2ae-fb540f63e794","Type":"ContainerStarted","Data":"c3f9dde6135d2febf970553b4531a0212d07687850754b4751ff75ce07523c00"} Dec 11 02:21:37 crc kubenswrapper[4824]: I1211 02:21:37.632262 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-785d8bcb8c-mxg84" Dec 11 02:21:37 crc kubenswrapper[4824]: I1211 02:21:37.641646 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-8xg6k" event={"ID":"d8fffe44-22d1-4108-95e9-088db7c342c7","Type":"ContainerDied","Data":"332d003b20e747c04355340803a8cae24cfa7a91c8d897bb9f0ae15a50122344"} Dec 11 02:21:37 crc kubenswrapper[4824]: I1211 02:21:37.641687 4824 scope.go:117] "RemoveContainer" containerID="50e6bcbeea57c3f998dd2fbdb1888e9a906eb3fbe8df7f6bbdc3d74d9a50108d" Dec 11 02:21:37 crc kubenswrapper[4824]: I1211 02:21:37.641782 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-8xg6k" Dec 11 02:21:37 crc kubenswrapper[4824]: I1211 02:21:37.678396 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5bd5bd9b-501f-43e4-88da-7b4f5c591b58","Type":"ContainerStarted","Data":"50ee438d3d0bff2e98d14fa97e8da5909280aa6fe3eaf9c27b9e6c59e5c12125"} Dec 11 02:21:37 crc kubenswrapper[4824]: I1211 02:21:37.692196 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6d658bd2-e722-433e-825c-b5f4068c7ae2","Type":"ContainerStarted","Data":"0a8851eabb0da1bc53638228e1741f6f6ce45fdf946a6207eaf2c47b8e46f6dc"} Dec 11 02:21:37 crc kubenswrapper[4824]: I1211 02:21:37.718972 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-785d8bcb8c-mxg84" podStartSLOduration=3.718950106 podStartE2EDuration="3.718950106s" podCreationTimestamp="2025-12-11 02:21:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:21:37.657240704 +0000 UTC m=+1239.346278083" watchObservedRunningTime="2025-12-11 02:21:37.718950106 +0000 UTC m=+1239.407987485" Dec 11 02:21:37 crc kubenswrapper[4824]: I1211 02:21:37.736159 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-8xg6k"] Dec 11 02:21:37 crc kubenswrapper[4824]: I1211 02:21:37.744968 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-8xg6k"] Dec 11 02:21:38 crc kubenswrapper[4824]: I1211 02:21:38.666995 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d8fffe44-22d1-4108-95e9-088db7c342c7" path="/var/lib/kubelet/pods/d8fffe44-22d1-4108-95e9-088db7c342c7/volumes" Dec 11 02:21:38 crc kubenswrapper[4824]: I1211 02:21:38.716776 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6d658bd2-e722-433e-825c-b5f4068c7ae2","Type":"ContainerStarted","Data":"cf5110cb19c13deb1b1e0dcf9c2f0ba6163d78866f03c52187c7b51fa1b4cff5"} Dec 11 02:21:38 crc kubenswrapper[4824]: I1211 02:21:38.716913 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="6d658bd2-e722-433e-825c-b5f4068c7ae2" containerName="glance-log" containerID="cri-o://0a8851eabb0da1bc53638228e1741f6f6ce45fdf946a6207eaf2c47b8e46f6dc" gracePeriod=30 Dec 11 02:21:38 crc kubenswrapper[4824]: I1211 02:21:38.717714 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="6d658bd2-e722-433e-825c-b5f4068c7ae2" containerName="glance-httpd" containerID="cri-o://cf5110cb19c13deb1b1e0dcf9c2f0ba6163d78866f03c52187c7b51fa1b4cff5" gracePeriod=30 Dec 11 02:21:38 crc kubenswrapper[4824]: I1211 02:21:38.754589 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="5bd5bd9b-501f-43e4-88da-7b4f5c591b58" containerName="glance-log" containerID="cri-o://50ee438d3d0bff2e98d14fa97e8da5909280aa6fe3eaf9c27b9e6c59e5c12125" gracePeriod=30 Dec 11 02:21:38 crc kubenswrapper[4824]: I1211 02:21:38.754728 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="5bd5bd9b-501f-43e4-88da-7b4f5c591b58" containerName="glance-httpd" containerID="cri-o://7bc894eb0df73abeaa4452cb22e48e8207ff47741a99a1f8b733b49245f935d6" gracePeriod=30 Dec 11 02:21:38 crc kubenswrapper[4824]: I1211 02:21:38.754899 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5bd5bd9b-501f-43e4-88da-7b4f5c591b58","Type":"ContainerStarted","Data":"7bc894eb0df73abeaa4452cb22e48e8207ff47741a99a1f8b733b49245f935d6"} Dec 11 02:21:38 crc kubenswrapper[4824]: I1211 02:21:38.854870 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.854851988 podStartE2EDuration="4.854851988s" podCreationTimestamp="2025-12-11 02:21:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:21:38.831080917 +0000 UTC m=+1240.520118286" watchObservedRunningTime="2025-12-11 02:21:38.854851988 +0000 UTC m=+1240.543889367" Dec 11 02:21:38 crc kubenswrapper[4824]: I1211 02:21:38.884621 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.884603925 podStartE2EDuration="4.884603925s" podCreationTimestamp="2025-12-11 02:21:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:21:38.871134795 +0000 UTC m=+1240.560172184" watchObservedRunningTime="2025-12-11 02:21:38.884603925 +0000 UTC m=+1240.573641304" Dec 11 02:21:39 crc kubenswrapper[4824]: I1211 02:21:39.477098 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 11 02:21:39 crc kubenswrapper[4824]: I1211 02:21:39.633085 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5bd5bd9b-501f-43e4-88da-7b4f5c591b58-internal-tls-certs\") pod \"5bd5bd9b-501f-43e4-88da-7b4f5c591b58\" (UID: \"5bd5bd9b-501f-43e4-88da-7b4f5c591b58\") " Dec 11 02:21:39 crc kubenswrapper[4824]: I1211 02:21:39.633146 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5bd5bd9b-501f-43e4-88da-7b4f5c591b58-config-data\") pod \"5bd5bd9b-501f-43e4-88da-7b4f5c591b58\" (UID: \"5bd5bd9b-501f-43e4-88da-7b4f5c591b58\") " Dec 11 02:21:39 crc kubenswrapper[4824]: I1211 02:21:39.633165 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5bd5bd9b-501f-43e4-88da-7b4f5c591b58-httpd-run\") pod \"5bd5bd9b-501f-43e4-88da-7b4f5c591b58\" (UID: \"5bd5bd9b-501f-43e4-88da-7b4f5c591b58\") " Dec 11 02:21:39 crc kubenswrapper[4824]: I1211 02:21:39.633198 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5bd5bd9b-501f-43e4-88da-7b4f5c591b58-logs\") pod \"5bd5bd9b-501f-43e4-88da-7b4f5c591b58\" (UID: \"5bd5bd9b-501f-43e4-88da-7b4f5c591b58\") " Dec 11 02:21:39 crc kubenswrapper[4824]: I1211 02:21:39.633240 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5bd5bd9b-501f-43e4-88da-7b4f5c591b58-scripts\") pod \"5bd5bd9b-501f-43e4-88da-7b4f5c591b58\" (UID: \"5bd5bd9b-501f-43e4-88da-7b4f5c591b58\") " Dec 11 02:21:39 crc kubenswrapper[4824]: I1211 02:21:39.633264 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bd5bd9b-501f-43e4-88da-7b4f5c591b58-combined-ca-bundle\") pod \"5bd5bd9b-501f-43e4-88da-7b4f5c591b58\" (UID: \"5bd5bd9b-501f-43e4-88da-7b4f5c591b58\") " Dec 11 02:21:39 crc kubenswrapper[4824]: I1211 02:21:39.633330 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"5bd5bd9b-501f-43e4-88da-7b4f5c591b58\" (UID: \"5bd5bd9b-501f-43e4-88da-7b4f5c591b58\") " Dec 11 02:21:39 crc kubenswrapper[4824]: I1211 02:21:39.633432 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-88rkb\" (UniqueName: \"kubernetes.io/projected/5bd5bd9b-501f-43e4-88da-7b4f5c591b58-kube-api-access-88rkb\") pod \"5bd5bd9b-501f-43e4-88da-7b4f5c591b58\" (UID: \"5bd5bd9b-501f-43e4-88da-7b4f5c591b58\") " Dec 11 02:21:39 crc kubenswrapper[4824]: I1211 02:21:39.636635 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5bd5bd9b-501f-43e4-88da-7b4f5c591b58-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "5bd5bd9b-501f-43e4-88da-7b4f5c591b58" (UID: "5bd5bd9b-501f-43e4-88da-7b4f5c591b58"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:21:39 crc kubenswrapper[4824]: I1211 02:21:39.638670 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5bd5bd9b-501f-43e4-88da-7b4f5c591b58-logs" (OuterVolumeSpecName: "logs") pod "5bd5bd9b-501f-43e4-88da-7b4f5c591b58" (UID: "5bd5bd9b-501f-43e4-88da-7b4f5c591b58"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:21:39 crc kubenswrapper[4824]: I1211 02:21:39.643373 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5bd5bd9b-501f-43e4-88da-7b4f5c591b58-scripts" (OuterVolumeSpecName: "scripts") pod "5bd5bd9b-501f-43e4-88da-7b4f5c591b58" (UID: "5bd5bd9b-501f-43e4-88da-7b4f5c591b58"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:21:39 crc kubenswrapper[4824]: I1211 02:21:39.645933 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "5bd5bd9b-501f-43e4-88da-7b4f5c591b58" (UID: "5bd5bd9b-501f-43e4-88da-7b4f5c591b58"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 11 02:21:39 crc kubenswrapper[4824]: I1211 02:21:39.661922 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5bd5bd9b-501f-43e4-88da-7b4f5c591b58-kube-api-access-88rkb" (OuterVolumeSpecName: "kube-api-access-88rkb") pod "5bd5bd9b-501f-43e4-88da-7b4f5c591b58" (UID: "5bd5bd9b-501f-43e4-88da-7b4f5c591b58"). InnerVolumeSpecName "kube-api-access-88rkb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:21:39 crc kubenswrapper[4824]: I1211 02:21:39.675452 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5bd5bd9b-501f-43e4-88da-7b4f5c591b58-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5bd5bd9b-501f-43e4-88da-7b4f5c591b58" (UID: "5bd5bd9b-501f-43e4-88da-7b4f5c591b58"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:21:39 crc kubenswrapper[4824]: I1211 02:21:39.716254 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5bd5bd9b-501f-43e4-88da-7b4f5c591b58-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "5bd5bd9b-501f-43e4-88da-7b4f5c591b58" (UID: "5bd5bd9b-501f-43e4-88da-7b4f5c591b58"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:21:39 crc kubenswrapper[4824]: I1211 02:21:39.721610 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5bd5bd9b-501f-43e4-88da-7b4f5c591b58-config-data" (OuterVolumeSpecName: "config-data") pod "5bd5bd9b-501f-43e4-88da-7b4f5c591b58" (UID: "5bd5bd9b-501f-43e4-88da-7b4f5c591b58"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:21:39 crc kubenswrapper[4824]: I1211 02:21:39.734801 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5bd5bd9b-501f-43e4-88da-7b4f5c591b58-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 02:21:39 crc kubenswrapper[4824]: I1211 02:21:39.734824 4824 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5bd5bd9b-501f-43e4-88da-7b4f5c591b58-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 11 02:21:39 crc kubenswrapper[4824]: I1211 02:21:39.734833 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5bd5bd9b-501f-43e4-88da-7b4f5c591b58-logs\") on node \"crc\" DevicePath \"\"" Dec 11 02:21:39 crc kubenswrapper[4824]: I1211 02:21:39.734840 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5bd5bd9b-501f-43e4-88da-7b4f5c591b58-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 02:21:39 crc kubenswrapper[4824]: I1211 02:21:39.734848 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bd5bd9b-501f-43e4-88da-7b4f5c591b58-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 02:21:39 crc kubenswrapper[4824]: I1211 02:21:39.734870 4824 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Dec 11 02:21:39 crc kubenswrapper[4824]: I1211 02:21:39.734878 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-88rkb\" (UniqueName: \"kubernetes.io/projected/5bd5bd9b-501f-43e4-88da-7b4f5c591b58-kube-api-access-88rkb\") on node \"crc\" DevicePath \"\"" Dec 11 02:21:39 crc kubenswrapper[4824]: I1211 02:21:39.734890 4824 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5bd5bd9b-501f-43e4-88da-7b4f5c591b58-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 11 02:21:39 crc kubenswrapper[4824]: I1211 02:21:39.763404 4824 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Dec 11 02:21:39 crc kubenswrapper[4824]: I1211 02:21:39.787152 4824 generic.go:334] "Generic (PLEG): container finished" podID="5bd5bd9b-501f-43e4-88da-7b4f5c591b58" containerID="7bc894eb0df73abeaa4452cb22e48e8207ff47741a99a1f8b733b49245f935d6" exitCode=143 Dec 11 02:21:39 crc kubenswrapper[4824]: I1211 02:21:39.787184 4824 generic.go:334] "Generic (PLEG): container finished" podID="5bd5bd9b-501f-43e4-88da-7b4f5c591b58" containerID="50ee438d3d0bff2e98d14fa97e8da5909280aa6fe3eaf9c27b9e6c59e5c12125" exitCode=143 Dec 11 02:21:39 crc kubenswrapper[4824]: I1211 02:21:39.787214 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 11 02:21:39 crc kubenswrapper[4824]: I1211 02:21:39.787264 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5bd5bd9b-501f-43e4-88da-7b4f5c591b58","Type":"ContainerDied","Data":"7bc894eb0df73abeaa4452cb22e48e8207ff47741a99a1f8b733b49245f935d6"} Dec 11 02:21:39 crc kubenswrapper[4824]: I1211 02:21:39.787323 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5bd5bd9b-501f-43e4-88da-7b4f5c591b58","Type":"ContainerDied","Data":"50ee438d3d0bff2e98d14fa97e8da5909280aa6fe3eaf9c27b9e6c59e5c12125"} Dec 11 02:21:39 crc kubenswrapper[4824]: I1211 02:21:39.787335 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5bd5bd9b-501f-43e4-88da-7b4f5c591b58","Type":"ContainerDied","Data":"bcacac34cf2cbf245a937445b260ede2ab3250b3c50923d970602e2f8303c42e"} Dec 11 02:21:39 crc kubenswrapper[4824]: I1211 02:21:39.787352 4824 scope.go:117] "RemoveContainer" containerID="7bc894eb0df73abeaa4452cb22e48e8207ff47741a99a1f8b733b49245f935d6" Dec 11 02:21:39 crc kubenswrapper[4824]: I1211 02:21:39.790253 4824 generic.go:334] "Generic (PLEG): container finished" podID="6d658bd2-e722-433e-825c-b5f4068c7ae2" containerID="cf5110cb19c13deb1b1e0dcf9c2f0ba6163d78866f03c52187c7b51fa1b4cff5" exitCode=0 Dec 11 02:21:39 crc kubenswrapper[4824]: I1211 02:21:39.790266 4824 generic.go:334] "Generic (PLEG): container finished" podID="6d658bd2-e722-433e-825c-b5f4068c7ae2" containerID="0a8851eabb0da1bc53638228e1741f6f6ce45fdf946a6207eaf2c47b8e46f6dc" exitCode=143 Dec 11 02:21:39 crc kubenswrapper[4824]: I1211 02:21:39.790457 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6d658bd2-e722-433e-825c-b5f4068c7ae2","Type":"ContainerDied","Data":"cf5110cb19c13deb1b1e0dcf9c2f0ba6163d78866f03c52187c7b51fa1b4cff5"} Dec 11 02:21:39 crc kubenswrapper[4824]: I1211 02:21:39.790502 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6d658bd2-e722-433e-825c-b5f4068c7ae2","Type":"ContainerDied","Data":"0a8851eabb0da1bc53638228e1741f6f6ce45fdf946a6207eaf2c47b8e46f6dc"} Dec 11 02:21:39 crc kubenswrapper[4824]: I1211 02:21:39.836862 4824 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Dec 11 02:21:39 crc kubenswrapper[4824]: I1211 02:21:39.860540 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 11 02:21:39 crc kubenswrapper[4824]: I1211 02:21:39.869249 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 11 02:21:39 crc kubenswrapper[4824]: I1211 02:21:39.895727 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 11 02:21:39 crc kubenswrapper[4824]: E1211 02:21:39.896099 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bd5bd9b-501f-43e4-88da-7b4f5c591b58" containerName="glance-httpd" Dec 11 02:21:39 crc kubenswrapper[4824]: I1211 02:21:39.896128 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bd5bd9b-501f-43e4-88da-7b4f5c591b58" containerName="glance-httpd" Dec 11 02:21:39 crc kubenswrapper[4824]: E1211 02:21:39.896139 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bd5bd9b-501f-43e4-88da-7b4f5c591b58" containerName="glance-log" Dec 11 02:21:39 crc kubenswrapper[4824]: I1211 02:21:39.896145 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bd5bd9b-501f-43e4-88da-7b4f5c591b58" containerName="glance-log" Dec 11 02:21:39 crc kubenswrapper[4824]: E1211 02:21:39.896174 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8fffe44-22d1-4108-95e9-088db7c342c7" containerName="init" Dec 11 02:21:39 crc kubenswrapper[4824]: I1211 02:21:39.896181 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8fffe44-22d1-4108-95e9-088db7c342c7" containerName="init" Dec 11 02:21:39 crc kubenswrapper[4824]: I1211 02:21:39.896361 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8fffe44-22d1-4108-95e9-088db7c342c7" containerName="init" Dec 11 02:21:39 crc kubenswrapper[4824]: I1211 02:21:39.896389 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="5bd5bd9b-501f-43e4-88da-7b4f5c591b58" containerName="glance-httpd" Dec 11 02:21:39 crc kubenswrapper[4824]: I1211 02:21:39.896398 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="5bd5bd9b-501f-43e4-88da-7b4f5c591b58" containerName="glance-log" Dec 11 02:21:39 crc kubenswrapper[4824]: I1211 02:21:39.897963 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 11 02:21:39 crc kubenswrapper[4824]: I1211 02:21:39.901834 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 11 02:21:39 crc kubenswrapper[4824]: I1211 02:21:39.902045 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 11 02:21:39 crc kubenswrapper[4824]: I1211 02:21:39.907296 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 11 02:21:40 crc kubenswrapper[4824]: I1211 02:21:40.039331 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/189f7197-9da3-402f-9c9b-042e34ca6556-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"189f7197-9da3-402f-9c9b-042e34ca6556\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:21:40 crc kubenswrapper[4824]: I1211 02:21:40.039426 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/189f7197-9da3-402f-9c9b-042e34ca6556-scripts\") pod \"glance-default-internal-api-0\" (UID: \"189f7197-9da3-402f-9c9b-042e34ca6556\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:21:40 crc kubenswrapper[4824]: I1211 02:21:40.039454 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"189f7197-9da3-402f-9c9b-042e34ca6556\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:21:40 crc kubenswrapper[4824]: I1211 02:21:40.039486 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tnnmb\" (UniqueName: \"kubernetes.io/projected/189f7197-9da3-402f-9c9b-042e34ca6556-kube-api-access-tnnmb\") pod \"glance-default-internal-api-0\" (UID: \"189f7197-9da3-402f-9c9b-042e34ca6556\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:21:40 crc kubenswrapper[4824]: I1211 02:21:40.039520 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/189f7197-9da3-402f-9c9b-042e34ca6556-logs\") pod \"glance-default-internal-api-0\" (UID: \"189f7197-9da3-402f-9c9b-042e34ca6556\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:21:40 crc kubenswrapper[4824]: I1211 02:21:40.039549 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/189f7197-9da3-402f-9c9b-042e34ca6556-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"189f7197-9da3-402f-9c9b-042e34ca6556\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:21:40 crc kubenswrapper[4824]: I1211 02:21:40.039647 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/189f7197-9da3-402f-9c9b-042e34ca6556-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"189f7197-9da3-402f-9c9b-042e34ca6556\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:21:40 crc kubenswrapper[4824]: I1211 02:21:40.039679 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/189f7197-9da3-402f-9c9b-042e34ca6556-config-data\") pod \"glance-default-internal-api-0\" (UID: \"189f7197-9da3-402f-9c9b-042e34ca6556\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:21:40 crc kubenswrapper[4824]: I1211 02:21:40.141215 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/189f7197-9da3-402f-9c9b-042e34ca6556-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"189f7197-9da3-402f-9c9b-042e34ca6556\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:21:40 crc kubenswrapper[4824]: I1211 02:21:40.141393 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/189f7197-9da3-402f-9c9b-042e34ca6556-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"189f7197-9da3-402f-9c9b-042e34ca6556\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:21:40 crc kubenswrapper[4824]: I1211 02:21:40.141429 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/189f7197-9da3-402f-9c9b-042e34ca6556-config-data\") pod \"glance-default-internal-api-0\" (UID: \"189f7197-9da3-402f-9c9b-042e34ca6556\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:21:40 crc kubenswrapper[4824]: I1211 02:21:40.141466 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/189f7197-9da3-402f-9c9b-042e34ca6556-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"189f7197-9da3-402f-9c9b-042e34ca6556\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:21:40 crc kubenswrapper[4824]: I1211 02:21:40.141512 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/189f7197-9da3-402f-9c9b-042e34ca6556-scripts\") pod \"glance-default-internal-api-0\" (UID: \"189f7197-9da3-402f-9c9b-042e34ca6556\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:21:40 crc kubenswrapper[4824]: I1211 02:21:40.141535 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"189f7197-9da3-402f-9c9b-042e34ca6556\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:21:40 crc kubenswrapper[4824]: I1211 02:21:40.141569 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tnnmb\" (UniqueName: \"kubernetes.io/projected/189f7197-9da3-402f-9c9b-042e34ca6556-kube-api-access-tnnmb\") pod \"glance-default-internal-api-0\" (UID: \"189f7197-9da3-402f-9c9b-042e34ca6556\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:21:40 crc kubenswrapper[4824]: I1211 02:21:40.141604 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/189f7197-9da3-402f-9c9b-042e34ca6556-logs\") pod \"glance-default-internal-api-0\" (UID: \"189f7197-9da3-402f-9c9b-042e34ca6556\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:21:40 crc kubenswrapper[4824]: I1211 02:21:40.142773 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"189f7197-9da3-402f-9c9b-042e34ca6556\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-internal-api-0" Dec 11 02:21:40 crc kubenswrapper[4824]: I1211 02:21:40.143607 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/189f7197-9da3-402f-9c9b-042e34ca6556-logs\") pod \"glance-default-internal-api-0\" (UID: \"189f7197-9da3-402f-9c9b-042e34ca6556\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:21:40 crc kubenswrapper[4824]: I1211 02:21:40.143624 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/189f7197-9da3-402f-9c9b-042e34ca6556-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"189f7197-9da3-402f-9c9b-042e34ca6556\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:21:40 crc kubenswrapper[4824]: I1211 02:21:40.147460 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/189f7197-9da3-402f-9c9b-042e34ca6556-config-data\") pod \"glance-default-internal-api-0\" (UID: \"189f7197-9da3-402f-9c9b-042e34ca6556\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:21:40 crc kubenswrapper[4824]: I1211 02:21:40.148053 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/189f7197-9da3-402f-9c9b-042e34ca6556-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"189f7197-9da3-402f-9c9b-042e34ca6556\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:21:40 crc kubenswrapper[4824]: I1211 02:21:40.148082 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/189f7197-9da3-402f-9c9b-042e34ca6556-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"189f7197-9da3-402f-9c9b-042e34ca6556\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:21:40 crc kubenswrapper[4824]: I1211 02:21:40.148317 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/189f7197-9da3-402f-9c9b-042e34ca6556-scripts\") pod \"glance-default-internal-api-0\" (UID: \"189f7197-9da3-402f-9c9b-042e34ca6556\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:21:40 crc kubenswrapper[4824]: I1211 02:21:40.157674 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tnnmb\" (UniqueName: \"kubernetes.io/projected/189f7197-9da3-402f-9c9b-042e34ca6556-kube-api-access-tnnmb\") pod \"glance-default-internal-api-0\" (UID: \"189f7197-9da3-402f-9c9b-042e34ca6556\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:21:40 crc kubenswrapper[4824]: I1211 02:21:40.168915 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"189f7197-9da3-402f-9c9b-042e34ca6556\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:21:40 crc kubenswrapper[4824]: I1211 02:21:40.242220 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 11 02:21:40 crc kubenswrapper[4824]: I1211 02:21:40.643045 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5bd5bd9b-501f-43e4-88da-7b4f5c591b58" path="/var/lib/kubelet/pods/5bd5bd9b-501f-43e4-88da-7b4f5c591b58/volumes" Dec 11 02:21:40 crc kubenswrapper[4824]: I1211 02:21:40.800395 4824 generic.go:334] "Generic (PLEG): container finished" podID="daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f" containerID="3ad722591b153d578fa20c1ff377018b683efae930cebb1a49b0e62ef80c0355" exitCode=0 Dec 11 02:21:40 crc kubenswrapper[4824]: I1211 02:21:40.800452 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-l45dd" event={"ID":"daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f","Type":"ContainerDied","Data":"3ad722591b153d578fa20c1ff377018b683efae930cebb1a49b0e62ef80c0355"} Dec 11 02:21:41 crc kubenswrapper[4824]: I1211 02:21:41.953866 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 11 02:21:42 crc kubenswrapper[4824]: I1211 02:21:42.080343 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"6d658bd2-e722-433e-825c-b5f4068c7ae2\" (UID: \"6d658bd2-e722-433e-825c-b5f4068c7ae2\") " Dec 11 02:21:42 crc kubenswrapper[4824]: I1211 02:21:42.080766 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6d658bd2-e722-433e-825c-b5f4068c7ae2-scripts\") pod \"6d658bd2-e722-433e-825c-b5f4068c7ae2\" (UID: \"6d658bd2-e722-433e-825c-b5f4068c7ae2\") " Dec 11 02:21:42 crc kubenswrapper[4824]: I1211 02:21:42.080822 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6d658bd2-e722-433e-825c-b5f4068c7ae2-httpd-run\") pod \"6d658bd2-e722-433e-825c-b5f4068c7ae2\" (UID: \"6d658bd2-e722-433e-825c-b5f4068c7ae2\") " Dec 11 02:21:42 crc kubenswrapper[4824]: I1211 02:21:42.080855 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v92r7\" (UniqueName: \"kubernetes.io/projected/6d658bd2-e722-433e-825c-b5f4068c7ae2-kube-api-access-v92r7\") pod \"6d658bd2-e722-433e-825c-b5f4068c7ae2\" (UID: \"6d658bd2-e722-433e-825c-b5f4068c7ae2\") " Dec 11 02:21:42 crc kubenswrapper[4824]: I1211 02:21:42.080930 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d658bd2-e722-433e-825c-b5f4068c7ae2-public-tls-certs\") pod \"6d658bd2-e722-433e-825c-b5f4068c7ae2\" (UID: \"6d658bd2-e722-433e-825c-b5f4068c7ae2\") " Dec 11 02:21:42 crc kubenswrapper[4824]: I1211 02:21:42.080954 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d658bd2-e722-433e-825c-b5f4068c7ae2-config-data\") pod \"6d658bd2-e722-433e-825c-b5f4068c7ae2\" (UID: \"6d658bd2-e722-433e-825c-b5f4068c7ae2\") " Dec 11 02:21:42 crc kubenswrapper[4824]: I1211 02:21:42.081184 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6d658bd2-e722-433e-825c-b5f4068c7ae2-logs\") pod \"6d658bd2-e722-433e-825c-b5f4068c7ae2\" (UID: \"6d658bd2-e722-433e-825c-b5f4068c7ae2\") " Dec 11 02:21:42 crc kubenswrapper[4824]: I1211 02:21:42.081290 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6d658bd2-e722-433e-825c-b5f4068c7ae2-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "6d658bd2-e722-433e-825c-b5f4068c7ae2" (UID: "6d658bd2-e722-433e-825c-b5f4068c7ae2"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:21:42 crc kubenswrapper[4824]: I1211 02:21:42.081580 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d658bd2-e722-433e-825c-b5f4068c7ae2-combined-ca-bundle\") pod \"6d658bd2-e722-433e-825c-b5f4068c7ae2\" (UID: \"6d658bd2-e722-433e-825c-b5f4068c7ae2\") " Dec 11 02:21:42 crc kubenswrapper[4824]: I1211 02:21:42.081579 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6d658bd2-e722-433e-825c-b5f4068c7ae2-logs" (OuterVolumeSpecName: "logs") pod "6d658bd2-e722-433e-825c-b5f4068c7ae2" (UID: "6d658bd2-e722-433e-825c-b5f4068c7ae2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:21:42 crc kubenswrapper[4824]: I1211 02:21:42.082196 4824 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6d658bd2-e722-433e-825c-b5f4068c7ae2-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 11 02:21:42 crc kubenswrapper[4824]: I1211 02:21:42.082219 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6d658bd2-e722-433e-825c-b5f4068c7ae2-logs\") on node \"crc\" DevicePath \"\"" Dec 11 02:21:42 crc kubenswrapper[4824]: I1211 02:21:42.086780 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d658bd2-e722-433e-825c-b5f4068c7ae2-scripts" (OuterVolumeSpecName: "scripts") pod "6d658bd2-e722-433e-825c-b5f4068c7ae2" (UID: "6d658bd2-e722-433e-825c-b5f4068c7ae2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:21:42 crc kubenswrapper[4824]: I1211 02:21:42.089527 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "glance") pod "6d658bd2-e722-433e-825c-b5f4068c7ae2" (UID: "6d658bd2-e722-433e-825c-b5f4068c7ae2"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 11 02:21:42 crc kubenswrapper[4824]: I1211 02:21:42.111469 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d658bd2-e722-433e-825c-b5f4068c7ae2-kube-api-access-v92r7" (OuterVolumeSpecName: "kube-api-access-v92r7") pod "6d658bd2-e722-433e-825c-b5f4068c7ae2" (UID: "6d658bd2-e722-433e-825c-b5f4068c7ae2"). InnerVolumeSpecName "kube-api-access-v92r7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:21:42 crc kubenswrapper[4824]: I1211 02:21:42.123229 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d658bd2-e722-433e-825c-b5f4068c7ae2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6d658bd2-e722-433e-825c-b5f4068c7ae2" (UID: "6d658bd2-e722-433e-825c-b5f4068c7ae2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:21:42 crc kubenswrapper[4824]: I1211 02:21:42.160695 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d658bd2-e722-433e-825c-b5f4068c7ae2-config-data" (OuterVolumeSpecName: "config-data") pod "6d658bd2-e722-433e-825c-b5f4068c7ae2" (UID: "6d658bd2-e722-433e-825c-b5f4068c7ae2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:21:42 crc kubenswrapper[4824]: I1211 02:21:42.170573 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d658bd2-e722-433e-825c-b5f4068c7ae2-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "6d658bd2-e722-433e-825c-b5f4068c7ae2" (UID: "6d658bd2-e722-433e-825c-b5f4068c7ae2"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:21:42 crc kubenswrapper[4824]: I1211 02:21:42.185729 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d658bd2-e722-433e-825c-b5f4068c7ae2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 02:21:42 crc kubenswrapper[4824]: I1211 02:21:42.185801 4824 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Dec 11 02:21:42 crc kubenswrapper[4824]: I1211 02:21:42.185812 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6d658bd2-e722-433e-825c-b5f4068c7ae2-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 02:21:42 crc kubenswrapper[4824]: I1211 02:21:42.185823 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v92r7\" (UniqueName: \"kubernetes.io/projected/6d658bd2-e722-433e-825c-b5f4068c7ae2-kube-api-access-v92r7\") on node \"crc\" DevicePath \"\"" Dec 11 02:21:42 crc kubenswrapper[4824]: I1211 02:21:42.185851 4824 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d658bd2-e722-433e-825c-b5f4068c7ae2-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 11 02:21:42 crc kubenswrapper[4824]: I1211 02:21:42.185860 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d658bd2-e722-433e-825c-b5f4068c7ae2-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 02:21:42 crc kubenswrapper[4824]: I1211 02:21:42.205066 4824 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Dec 11 02:21:42 crc kubenswrapper[4824]: I1211 02:21:42.287024 4824 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Dec 11 02:21:42 crc kubenswrapper[4824]: I1211 02:21:42.816764 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6d658bd2-e722-433e-825c-b5f4068c7ae2","Type":"ContainerDied","Data":"b563b0b55b6d8a1a5d784d6cb06c4c178f1c83027829be3a6820eb0ea5c05871"} Dec 11 02:21:42 crc kubenswrapper[4824]: I1211 02:21:42.816817 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 11 02:21:42 crc kubenswrapper[4824]: I1211 02:21:42.843819 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 11 02:21:42 crc kubenswrapper[4824]: I1211 02:21:42.850527 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 11 02:21:42 crc kubenswrapper[4824]: I1211 02:21:42.879504 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 11 02:21:42 crc kubenswrapper[4824]: E1211 02:21:42.879968 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d658bd2-e722-433e-825c-b5f4068c7ae2" containerName="glance-log" Dec 11 02:21:42 crc kubenswrapper[4824]: I1211 02:21:42.879990 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d658bd2-e722-433e-825c-b5f4068c7ae2" containerName="glance-log" Dec 11 02:21:42 crc kubenswrapper[4824]: E1211 02:21:42.880012 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d658bd2-e722-433e-825c-b5f4068c7ae2" containerName="glance-httpd" Dec 11 02:21:42 crc kubenswrapper[4824]: I1211 02:21:42.880020 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d658bd2-e722-433e-825c-b5f4068c7ae2" containerName="glance-httpd" Dec 11 02:21:42 crc kubenswrapper[4824]: I1211 02:21:42.880240 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d658bd2-e722-433e-825c-b5f4068c7ae2" containerName="glance-httpd" Dec 11 02:21:42 crc kubenswrapper[4824]: I1211 02:21:42.880270 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d658bd2-e722-433e-825c-b5f4068c7ae2" containerName="glance-log" Dec 11 02:21:42 crc kubenswrapper[4824]: I1211 02:21:42.882877 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 11 02:21:42 crc kubenswrapper[4824]: I1211 02:21:42.893446 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 11 02:21:42 crc kubenswrapper[4824]: I1211 02:21:42.896680 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 11 02:21:42 crc kubenswrapper[4824]: I1211 02:21:42.898030 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 11 02:21:43 crc kubenswrapper[4824]: I1211 02:21:43.005207 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"dfa654c8-1f3b-43f6-8a1f-709a05020b4e\") " pod="openstack/glance-default-external-api-0" Dec 11 02:21:43 crc kubenswrapper[4824]: I1211 02:21:43.005263 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/dfa654c8-1f3b-43f6-8a1f-709a05020b4e-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"dfa654c8-1f3b-43f6-8a1f-709a05020b4e\") " pod="openstack/glance-default-external-api-0" Dec 11 02:21:43 crc kubenswrapper[4824]: I1211 02:21:43.005284 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q6vwj\" (UniqueName: \"kubernetes.io/projected/dfa654c8-1f3b-43f6-8a1f-709a05020b4e-kube-api-access-q6vwj\") pod \"glance-default-external-api-0\" (UID: \"dfa654c8-1f3b-43f6-8a1f-709a05020b4e\") " pod="openstack/glance-default-external-api-0" Dec 11 02:21:43 crc kubenswrapper[4824]: I1211 02:21:43.005342 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dfa654c8-1f3b-43f6-8a1f-709a05020b4e-logs\") pod \"glance-default-external-api-0\" (UID: \"dfa654c8-1f3b-43f6-8a1f-709a05020b4e\") " pod="openstack/glance-default-external-api-0" Dec 11 02:21:43 crc kubenswrapper[4824]: I1211 02:21:43.005380 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfa654c8-1f3b-43f6-8a1f-709a05020b4e-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"dfa654c8-1f3b-43f6-8a1f-709a05020b4e\") " pod="openstack/glance-default-external-api-0" Dec 11 02:21:43 crc kubenswrapper[4824]: I1211 02:21:43.005426 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dfa654c8-1f3b-43f6-8a1f-709a05020b4e-scripts\") pod \"glance-default-external-api-0\" (UID: \"dfa654c8-1f3b-43f6-8a1f-709a05020b4e\") " pod="openstack/glance-default-external-api-0" Dec 11 02:21:43 crc kubenswrapper[4824]: I1211 02:21:43.005445 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dfa654c8-1f3b-43f6-8a1f-709a05020b4e-config-data\") pod \"glance-default-external-api-0\" (UID: \"dfa654c8-1f3b-43f6-8a1f-709a05020b4e\") " pod="openstack/glance-default-external-api-0" Dec 11 02:21:43 crc kubenswrapper[4824]: I1211 02:21:43.005498 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dfa654c8-1f3b-43f6-8a1f-709a05020b4e-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"dfa654c8-1f3b-43f6-8a1f-709a05020b4e\") " pod="openstack/glance-default-external-api-0" Dec 11 02:21:43 crc kubenswrapper[4824]: I1211 02:21:43.107308 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dfa654c8-1f3b-43f6-8a1f-709a05020b4e-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"dfa654c8-1f3b-43f6-8a1f-709a05020b4e\") " pod="openstack/glance-default-external-api-0" Dec 11 02:21:43 crc kubenswrapper[4824]: I1211 02:21:43.107650 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"dfa654c8-1f3b-43f6-8a1f-709a05020b4e\") " pod="openstack/glance-default-external-api-0" Dec 11 02:21:43 crc kubenswrapper[4824]: I1211 02:21:43.107698 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/dfa654c8-1f3b-43f6-8a1f-709a05020b4e-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"dfa654c8-1f3b-43f6-8a1f-709a05020b4e\") " pod="openstack/glance-default-external-api-0" Dec 11 02:21:43 crc kubenswrapper[4824]: I1211 02:21:43.107719 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q6vwj\" (UniqueName: \"kubernetes.io/projected/dfa654c8-1f3b-43f6-8a1f-709a05020b4e-kube-api-access-q6vwj\") pod \"glance-default-external-api-0\" (UID: \"dfa654c8-1f3b-43f6-8a1f-709a05020b4e\") " pod="openstack/glance-default-external-api-0" Dec 11 02:21:43 crc kubenswrapper[4824]: I1211 02:21:43.107755 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dfa654c8-1f3b-43f6-8a1f-709a05020b4e-logs\") pod \"glance-default-external-api-0\" (UID: \"dfa654c8-1f3b-43f6-8a1f-709a05020b4e\") " pod="openstack/glance-default-external-api-0" Dec 11 02:21:43 crc kubenswrapper[4824]: I1211 02:21:43.107806 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfa654c8-1f3b-43f6-8a1f-709a05020b4e-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"dfa654c8-1f3b-43f6-8a1f-709a05020b4e\") " pod="openstack/glance-default-external-api-0" Dec 11 02:21:43 crc kubenswrapper[4824]: I1211 02:21:43.107869 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dfa654c8-1f3b-43f6-8a1f-709a05020b4e-scripts\") pod \"glance-default-external-api-0\" (UID: \"dfa654c8-1f3b-43f6-8a1f-709a05020b4e\") " pod="openstack/glance-default-external-api-0" Dec 11 02:21:43 crc kubenswrapper[4824]: I1211 02:21:43.107911 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dfa654c8-1f3b-43f6-8a1f-709a05020b4e-config-data\") pod \"glance-default-external-api-0\" (UID: \"dfa654c8-1f3b-43f6-8a1f-709a05020b4e\") " pod="openstack/glance-default-external-api-0" Dec 11 02:21:43 crc kubenswrapper[4824]: I1211 02:21:43.109618 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-55cc49576f-sct4f"] Dec 11 02:21:43 crc kubenswrapper[4824]: I1211 02:21:43.113806 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dfa654c8-1f3b-43f6-8a1f-709a05020b4e-config-data\") pod \"glance-default-external-api-0\" (UID: \"dfa654c8-1f3b-43f6-8a1f-709a05020b4e\") " pod="openstack/glance-default-external-api-0" Dec 11 02:21:43 crc kubenswrapper[4824]: I1211 02:21:43.114633 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dfa654c8-1f3b-43f6-8a1f-709a05020b4e-logs\") pod \"glance-default-external-api-0\" (UID: \"dfa654c8-1f3b-43f6-8a1f-709a05020b4e\") " pod="openstack/glance-default-external-api-0" Dec 11 02:21:43 crc kubenswrapper[4824]: I1211 02:21:43.115333 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"dfa654c8-1f3b-43f6-8a1f-709a05020b4e\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/glance-default-external-api-0" Dec 11 02:21:43 crc kubenswrapper[4824]: I1211 02:21:43.115588 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/dfa654c8-1f3b-43f6-8a1f-709a05020b4e-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"dfa654c8-1f3b-43f6-8a1f-709a05020b4e\") " pod="openstack/glance-default-external-api-0" Dec 11 02:21:43 crc kubenswrapper[4824]: I1211 02:21:43.120428 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dfa654c8-1f3b-43f6-8a1f-709a05020b4e-scripts\") pod \"glance-default-external-api-0\" (UID: \"dfa654c8-1f3b-43f6-8a1f-709a05020b4e\") " pod="openstack/glance-default-external-api-0" Dec 11 02:21:43 crc kubenswrapper[4824]: I1211 02:21:43.120473 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfa654c8-1f3b-43f6-8a1f-709a05020b4e-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"dfa654c8-1f3b-43f6-8a1f-709a05020b4e\") " pod="openstack/glance-default-external-api-0" Dec 11 02:21:43 crc kubenswrapper[4824]: I1211 02:21:43.123673 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dfa654c8-1f3b-43f6-8a1f-709a05020b4e-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"dfa654c8-1f3b-43f6-8a1f-709a05020b4e\") " pod="openstack/glance-default-external-api-0" Dec 11 02:21:43 crc kubenswrapper[4824]: I1211 02:21:43.140640 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q6vwj\" (UniqueName: \"kubernetes.io/projected/dfa654c8-1f3b-43f6-8a1f-709a05020b4e-kube-api-access-q6vwj\") pod \"glance-default-external-api-0\" (UID: \"dfa654c8-1f3b-43f6-8a1f-709a05020b4e\") " pod="openstack/glance-default-external-api-0" Dec 11 02:21:43 crc kubenswrapper[4824]: I1211 02:21:43.162071 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-787649dcdd-ts6qs"] Dec 11 02:21:43 crc kubenswrapper[4824]: I1211 02:21:43.164369 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-787649dcdd-ts6qs" Dec 11 02:21:43 crc kubenswrapper[4824]: I1211 02:21:43.166492 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Dec 11 02:21:43 crc kubenswrapper[4824]: I1211 02:21:43.172639 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-787649dcdd-ts6qs"] Dec 11 02:21:43 crc kubenswrapper[4824]: I1211 02:21:43.177512 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"dfa654c8-1f3b-43f6-8a1f-709a05020b4e\") " pod="openstack/glance-default-external-api-0" Dec 11 02:21:43 crc kubenswrapper[4824]: I1211 02:21:43.200617 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 11 02:21:43 crc kubenswrapper[4824]: I1211 02:21:43.208884 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5f68f577dc-nj2zp"] Dec 11 02:21:43 crc kubenswrapper[4824]: I1211 02:21:43.247981 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 11 02:21:43 crc kubenswrapper[4824]: I1211 02:21:43.256676 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-74cc6c88d8-zvlk7"] Dec 11 02:21:43 crc kubenswrapper[4824]: I1211 02:21:43.257971 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-74cc6c88d8-zvlk7" Dec 11 02:21:43 crc kubenswrapper[4824]: I1211 02:21:43.268366 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-74cc6c88d8-zvlk7"] Dec 11 02:21:43 crc kubenswrapper[4824]: I1211 02:21:43.311757 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f15f9ed0-98f2-42fe-8006-2142098cc2b6-horizon-secret-key\") pod \"horizon-787649dcdd-ts6qs\" (UID: \"f15f9ed0-98f2-42fe-8006-2142098cc2b6\") " pod="openstack/horizon-787649dcdd-ts6qs" Dec 11 02:21:43 crc kubenswrapper[4824]: I1211 02:21:43.311822 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f15f9ed0-98f2-42fe-8006-2142098cc2b6-logs\") pod \"horizon-787649dcdd-ts6qs\" (UID: \"f15f9ed0-98f2-42fe-8006-2142098cc2b6\") " pod="openstack/horizon-787649dcdd-ts6qs" Dec 11 02:21:43 crc kubenswrapper[4824]: I1211 02:21:43.311846 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f15f9ed0-98f2-42fe-8006-2142098cc2b6-combined-ca-bundle\") pod \"horizon-787649dcdd-ts6qs\" (UID: \"f15f9ed0-98f2-42fe-8006-2142098cc2b6\") " pod="openstack/horizon-787649dcdd-ts6qs" Dec 11 02:21:43 crc kubenswrapper[4824]: I1211 02:21:43.311889 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t6pdv\" (UniqueName: \"kubernetes.io/projected/f15f9ed0-98f2-42fe-8006-2142098cc2b6-kube-api-access-t6pdv\") pod \"horizon-787649dcdd-ts6qs\" (UID: \"f15f9ed0-98f2-42fe-8006-2142098cc2b6\") " pod="openstack/horizon-787649dcdd-ts6qs" Dec 11 02:21:43 crc kubenswrapper[4824]: I1211 02:21:43.311940 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f15f9ed0-98f2-42fe-8006-2142098cc2b6-config-data\") pod \"horizon-787649dcdd-ts6qs\" (UID: \"f15f9ed0-98f2-42fe-8006-2142098cc2b6\") " pod="openstack/horizon-787649dcdd-ts6qs" Dec 11 02:21:43 crc kubenswrapper[4824]: I1211 02:21:43.311982 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f15f9ed0-98f2-42fe-8006-2142098cc2b6-scripts\") pod \"horizon-787649dcdd-ts6qs\" (UID: \"f15f9ed0-98f2-42fe-8006-2142098cc2b6\") " pod="openstack/horizon-787649dcdd-ts6qs" Dec 11 02:21:43 crc kubenswrapper[4824]: I1211 02:21:43.312029 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/f15f9ed0-98f2-42fe-8006-2142098cc2b6-horizon-tls-certs\") pod \"horizon-787649dcdd-ts6qs\" (UID: \"f15f9ed0-98f2-42fe-8006-2142098cc2b6\") " pod="openstack/horizon-787649dcdd-ts6qs" Dec 11 02:21:43 crc kubenswrapper[4824]: I1211 02:21:43.314127 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 11 02:21:43 crc kubenswrapper[4824]: I1211 02:21:43.413545 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f15f9ed0-98f2-42fe-8006-2142098cc2b6-scripts\") pod \"horizon-787649dcdd-ts6qs\" (UID: \"f15f9ed0-98f2-42fe-8006-2142098cc2b6\") " pod="openstack/horizon-787649dcdd-ts6qs" Dec 11 02:21:43 crc kubenswrapper[4824]: I1211 02:21:43.413641 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1da5eb01-d709-42ad-b5f1-e7ea6310cc9e-scripts\") pod \"horizon-74cc6c88d8-zvlk7\" (UID: \"1da5eb01-d709-42ad-b5f1-e7ea6310cc9e\") " pod="openstack/horizon-74cc6c88d8-zvlk7" Dec 11 02:21:43 crc kubenswrapper[4824]: I1211 02:21:43.413699 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2wjkb\" (UniqueName: \"kubernetes.io/projected/1da5eb01-d709-42ad-b5f1-e7ea6310cc9e-kube-api-access-2wjkb\") pod \"horizon-74cc6c88d8-zvlk7\" (UID: \"1da5eb01-d709-42ad-b5f1-e7ea6310cc9e\") " pod="openstack/horizon-74cc6c88d8-zvlk7" Dec 11 02:21:43 crc kubenswrapper[4824]: I1211 02:21:43.413734 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/f15f9ed0-98f2-42fe-8006-2142098cc2b6-horizon-tls-certs\") pod \"horizon-787649dcdd-ts6qs\" (UID: \"f15f9ed0-98f2-42fe-8006-2142098cc2b6\") " pod="openstack/horizon-787649dcdd-ts6qs" Dec 11 02:21:43 crc kubenswrapper[4824]: I1211 02:21:43.413759 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f15f9ed0-98f2-42fe-8006-2142098cc2b6-horizon-secret-key\") pod \"horizon-787649dcdd-ts6qs\" (UID: \"f15f9ed0-98f2-42fe-8006-2142098cc2b6\") " pod="openstack/horizon-787649dcdd-ts6qs" Dec 11 02:21:43 crc kubenswrapper[4824]: I1211 02:21:43.413792 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f15f9ed0-98f2-42fe-8006-2142098cc2b6-logs\") pod \"horizon-787649dcdd-ts6qs\" (UID: \"f15f9ed0-98f2-42fe-8006-2142098cc2b6\") " pod="openstack/horizon-787649dcdd-ts6qs" Dec 11 02:21:43 crc kubenswrapper[4824]: I1211 02:21:43.413814 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1da5eb01-d709-42ad-b5f1-e7ea6310cc9e-horizon-secret-key\") pod \"horizon-74cc6c88d8-zvlk7\" (UID: \"1da5eb01-d709-42ad-b5f1-e7ea6310cc9e\") " pod="openstack/horizon-74cc6c88d8-zvlk7" Dec 11 02:21:43 crc kubenswrapper[4824]: I1211 02:21:43.413833 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f15f9ed0-98f2-42fe-8006-2142098cc2b6-combined-ca-bundle\") pod \"horizon-787649dcdd-ts6qs\" (UID: \"f15f9ed0-98f2-42fe-8006-2142098cc2b6\") " pod="openstack/horizon-787649dcdd-ts6qs" Dec 11 02:21:43 crc kubenswrapper[4824]: I1211 02:21:43.413849 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1da5eb01-d709-42ad-b5f1-e7ea6310cc9e-config-data\") pod \"horizon-74cc6c88d8-zvlk7\" (UID: \"1da5eb01-d709-42ad-b5f1-e7ea6310cc9e\") " pod="openstack/horizon-74cc6c88d8-zvlk7" Dec 11 02:21:43 crc kubenswrapper[4824]: I1211 02:21:43.413882 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t6pdv\" (UniqueName: \"kubernetes.io/projected/f15f9ed0-98f2-42fe-8006-2142098cc2b6-kube-api-access-t6pdv\") pod \"horizon-787649dcdd-ts6qs\" (UID: \"f15f9ed0-98f2-42fe-8006-2142098cc2b6\") " pod="openstack/horizon-787649dcdd-ts6qs" Dec 11 02:21:43 crc kubenswrapper[4824]: I1211 02:21:43.413914 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/1da5eb01-d709-42ad-b5f1-e7ea6310cc9e-horizon-tls-certs\") pod \"horizon-74cc6c88d8-zvlk7\" (UID: \"1da5eb01-d709-42ad-b5f1-e7ea6310cc9e\") " pod="openstack/horizon-74cc6c88d8-zvlk7" Dec 11 02:21:43 crc kubenswrapper[4824]: I1211 02:21:43.413938 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1da5eb01-d709-42ad-b5f1-e7ea6310cc9e-logs\") pod \"horizon-74cc6c88d8-zvlk7\" (UID: \"1da5eb01-d709-42ad-b5f1-e7ea6310cc9e\") " pod="openstack/horizon-74cc6c88d8-zvlk7" Dec 11 02:21:43 crc kubenswrapper[4824]: I1211 02:21:43.413959 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f15f9ed0-98f2-42fe-8006-2142098cc2b6-config-data\") pod \"horizon-787649dcdd-ts6qs\" (UID: \"f15f9ed0-98f2-42fe-8006-2142098cc2b6\") " pod="openstack/horizon-787649dcdd-ts6qs" Dec 11 02:21:43 crc kubenswrapper[4824]: I1211 02:21:43.413984 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1da5eb01-d709-42ad-b5f1-e7ea6310cc9e-combined-ca-bundle\") pod \"horizon-74cc6c88d8-zvlk7\" (UID: \"1da5eb01-d709-42ad-b5f1-e7ea6310cc9e\") " pod="openstack/horizon-74cc6c88d8-zvlk7" Dec 11 02:21:43 crc kubenswrapper[4824]: I1211 02:21:43.414355 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f15f9ed0-98f2-42fe-8006-2142098cc2b6-scripts\") pod \"horizon-787649dcdd-ts6qs\" (UID: \"f15f9ed0-98f2-42fe-8006-2142098cc2b6\") " pod="openstack/horizon-787649dcdd-ts6qs" Dec 11 02:21:43 crc kubenswrapper[4824]: I1211 02:21:43.415006 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f15f9ed0-98f2-42fe-8006-2142098cc2b6-logs\") pod \"horizon-787649dcdd-ts6qs\" (UID: \"f15f9ed0-98f2-42fe-8006-2142098cc2b6\") " pod="openstack/horizon-787649dcdd-ts6qs" Dec 11 02:21:43 crc kubenswrapper[4824]: I1211 02:21:43.417161 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f15f9ed0-98f2-42fe-8006-2142098cc2b6-config-data\") pod \"horizon-787649dcdd-ts6qs\" (UID: \"f15f9ed0-98f2-42fe-8006-2142098cc2b6\") " pod="openstack/horizon-787649dcdd-ts6qs" Dec 11 02:21:43 crc kubenswrapper[4824]: I1211 02:21:43.417710 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/f15f9ed0-98f2-42fe-8006-2142098cc2b6-horizon-tls-certs\") pod \"horizon-787649dcdd-ts6qs\" (UID: \"f15f9ed0-98f2-42fe-8006-2142098cc2b6\") " pod="openstack/horizon-787649dcdd-ts6qs" Dec 11 02:21:43 crc kubenswrapper[4824]: I1211 02:21:43.417859 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f15f9ed0-98f2-42fe-8006-2142098cc2b6-horizon-secret-key\") pod \"horizon-787649dcdd-ts6qs\" (UID: \"f15f9ed0-98f2-42fe-8006-2142098cc2b6\") " pod="openstack/horizon-787649dcdd-ts6qs" Dec 11 02:21:43 crc kubenswrapper[4824]: I1211 02:21:43.419463 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f15f9ed0-98f2-42fe-8006-2142098cc2b6-combined-ca-bundle\") pod \"horizon-787649dcdd-ts6qs\" (UID: \"f15f9ed0-98f2-42fe-8006-2142098cc2b6\") " pod="openstack/horizon-787649dcdd-ts6qs" Dec 11 02:21:43 crc kubenswrapper[4824]: I1211 02:21:43.456873 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t6pdv\" (UniqueName: \"kubernetes.io/projected/f15f9ed0-98f2-42fe-8006-2142098cc2b6-kube-api-access-t6pdv\") pod \"horizon-787649dcdd-ts6qs\" (UID: \"f15f9ed0-98f2-42fe-8006-2142098cc2b6\") " pod="openstack/horizon-787649dcdd-ts6qs" Dec 11 02:21:43 crc kubenswrapper[4824]: I1211 02:21:43.515461 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1da5eb01-d709-42ad-b5f1-e7ea6310cc9e-horizon-secret-key\") pod \"horizon-74cc6c88d8-zvlk7\" (UID: \"1da5eb01-d709-42ad-b5f1-e7ea6310cc9e\") " pod="openstack/horizon-74cc6c88d8-zvlk7" Dec 11 02:21:43 crc kubenswrapper[4824]: I1211 02:21:43.515499 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1da5eb01-d709-42ad-b5f1-e7ea6310cc9e-config-data\") pod \"horizon-74cc6c88d8-zvlk7\" (UID: \"1da5eb01-d709-42ad-b5f1-e7ea6310cc9e\") " pod="openstack/horizon-74cc6c88d8-zvlk7" Dec 11 02:21:43 crc kubenswrapper[4824]: I1211 02:21:43.515551 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/1da5eb01-d709-42ad-b5f1-e7ea6310cc9e-horizon-tls-certs\") pod \"horizon-74cc6c88d8-zvlk7\" (UID: \"1da5eb01-d709-42ad-b5f1-e7ea6310cc9e\") " pod="openstack/horizon-74cc6c88d8-zvlk7" Dec 11 02:21:43 crc kubenswrapper[4824]: I1211 02:21:43.515572 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1da5eb01-d709-42ad-b5f1-e7ea6310cc9e-logs\") pod \"horizon-74cc6c88d8-zvlk7\" (UID: \"1da5eb01-d709-42ad-b5f1-e7ea6310cc9e\") " pod="openstack/horizon-74cc6c88d8-zvlk7" Dec 11 02:21:43 crc kubenswrapper[4824]: I1211 02:21:43.515604 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1da5eb01-d709-42ad-b5f1-e7ea6310cc9e-combined-ca-bundle\") pod \"horizon-74cc6c88d8-zvlk7\" (UID: \"1da5eb01-d709-42ad-b5f1-e7ea6310cc9e\") " pod="openstack/horizon-74cc6c88d8-zvlk7" Dec 11 02:21:43 crc kubenswrapper[4824]: I1211 02:21:43.515631 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1da5eb01-d709-42ad-b5f1-e7ea6310cc9e-scripts\") pod \"horizon-74cc6c88d8-zvlk7\" (UID: \"1da5eb01-d709-42ad-b5f1-e7ea6310cc9e\") " pod="openstack/horizon-74cc6c88d8-zvlk7" Dec 11 02:21:43 crc kubenswrapper[4824]: I1211 02:21:43.515662 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2wjkb\" (UniqueName: \"kubernetes.io/projected/1da5eb01-d709-42ad-b5f1-e7ea6310cc9e-kube-api-access-2wjkb\") pod \"horizon-74cc6c88d8-zvlk7\" (UID: \"1da5eb01-d709-42ad-b5f1-e7ea6310cc9e\") " pod="openstack/horizon-74cc6c88d8-zvlk7" Dec 11 02:21:43 crc kubenswrapper[4824]: I1211 02:21:43.516543 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1da5eb01-d709-42ad-b5f1-e7ea6310cc9e-logs\") pod \"horizon-74cc6c88d8-zvlk7\" (UID: \"1da5eb01-d709-42ad-b5f1-e7ea6310cc9e\") " pod="openstack/horizon-74cc6c88d8-zvlk7" Dec 11 02:21:43 crc kubenswrapper[4824]: I1211 02:21:43.517483 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1da5eb01-d709-42ad-b5f1-e7ea6310cc9e-config-data\") pod \"horizon-74cc6c88d8-zvlk7\" (UID: \"1da5eb01-d709-42ad-b5f1-e7ea6310cc9e\") " pod="openstack/horizon-74cc6c88d8-zvlk7" Dec 11 02:21:43 crc kubenswrapper[4824]: I1211 02:21:43.520437 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/1da5eb01-d709-42ad-b5f1-e7ea6310cc9e-horizon-tls-certs\") pod \"horizon-74cc6c88d8-zvlk7\" (UID: \"1da5eb01-d709-42ad-b5f1-e7ea6310cc9e\") " pod="openstack/horizon-74cc6c88d8-zvlk7" Dec 11 02:21:43 crc kubenswrapper[4824]: I1211 02:21:43.520663 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1da5eb01-d709-42ad-b5f1-e7ea6310cc9e-horizon-secret-key\") pod \"horizon-74cc6c88d8-zvlk7\" (UID: \"1da5eb01-d709-42ad-b5f1-e7ea6310cc9e\") " pod="openstack/horizon-74cc6c88d8-zvlk7" Dec 11 02:21:43 crc kubenswrapper[4824]: I1211 02:21:43.520823 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1da5eb01-d709-42ad-b5f1-e7ea6310cc9e-scripts\") pod \"horizon-74cc6c88d8-zvlk7\" (UID: \"1da5eb01-d709-42ad-b5f1-e7ea6310cc9e\") " pod="openstack/horizon-74cc6c88d8-zvlk7" Dec 11 02:21:43 crc kubenswrapper[4824]: I1211 02:21:43.523351 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1da5eb01-d709-42ad-b5f1-e7ea6310cc9e-combined-ca-bundle\") pod \"horizon-74cc6c88d8-zvlk7\" (UID: \"1da5eb01-d709-42ad-b5f1-e7ea6310cc9e\") " pod="openstack/horizon-74cc6c88d8-zvlk7" Dec 11 02:21:43 crc kubenswrapper[4824]: I1211 02:21:43.540837 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2wjkb\" (UniqueName: \"kubernetes.io/projected/1da5eb01-d709-42ad-b5f1-e7ea6310cc9e-kube-api-access-2wjkb\") pod \"horizon-74cc6c88d8-zvlk7\" (UID: \"1da5eb01-d709-42ad-b5f1-e7ea6310cc9e\") " pod="openstack/horizon-74cc6c88d8-zvlk7" Dec 11 02:21:43 crc kubenswrapper[4824]: I1211 02:21:43.558518 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-787649dcdd-ts6qs" Dec 11 02:21:43 crc kubenswrapper[4824]: I1211 02:21:43.574560 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-74cc6c88d8-zvlk7" Dec 11 02:21:44 crc kubenswrapper[4824]: I1211 02:21:44.644507 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d658bd2-e722-433e-825c-b5f4068c7ae2" path="/var/lib/kubelet/pods/6d658bd2-e722-433e-825c-b5f4068c7ae2/volumes" Dec 11 02:21:44 crc kubenswrapper[4824]: I1211 02:21:44.803100 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-785d8bcb8c-mxg84" Dec 11 02:21:44 crc kubenswrapper[4824]: I1211 02:21:44.859739 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-4tss6"] Dec 11 02:21:44 crc kubenswrapper[4824]: I1211 02:21:44.859960 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-74f6bcbc87-4tss6" podUID="5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00" containerName="dnsmasq-dns" containerID="cri-o://ae65a91d118ac4d63800cb6c6bcfa72370d4dec5921ffa4c291a7fe609becdf5" gracePeriod=10 Dec 11 02:21:45 crc kubenswrapper[4824]: I1211 02:21:45.861502 4824 generic.go:334] "Generic (PLEG): container finished" podID="5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00" containerID="ae65a91d118ac4d63800cb6c6bcfa72370d4dec5921ffa4c291a7fe609becdf5" exitCode=0 Dec 11 02:21:45 crc kubenswrapper[4824]: I1211 02:21:45.861661 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-4tss6" event={"ID":"5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00","Type":"ContainerDied","Data":"ae65a91d118ac4d63800cb6c6bcfa72370d4dec5921ffa4c291a7fe609becdf5"} Dec 11 02:21:46 crc kubenswrapper[4824]: I1211 02:21:46.956876 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-74f6bcbc87-4tss6" podUID="5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.130:5353: connect: connection refused" Dec 11 02:21:51 crc kubenswrapper[4824]: E1211 02:21:51.710005 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Dec 11 02:21:51 crc kubenswrapper[4824]: E1211 02:21:51.710631 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68bh666h6h586h4h694h579h5c6h599hdbh669hfdhc5h65bh695hc9h695h55hf4h644h5c5h56h7h666h56bhb8h55h8dh5hc4h67ch548q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fcst2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-7656dcd5fc-pvdtd_openstack(97c61c1d-8fc0-46f3-9bb0-583a8c4aba1f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 11 02:21:51 crc kubenswrapper[4824]: E1211 02:21:51.713679 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-7656dcd5fc-pvdtd" podUID="97c61c1d-8fc0-46f3-9bb0-583a8c4aba1f" Dec 11 02:21:51 crc kubenswrapper[4824]: E1211 02:21:51.732995 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Dec 11 02:21:51 crc kubenswrapper[4824]: E1211 02:21:51.733167 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n554h646h56h5dbh677h685h6dh54dh574h55ch548h56h556h5c8hf4h5c7h54bh595h677h575hc9h588h5b5h8dhc5h544h599h56bh57fhcbh54dh645q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-x7b6b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-55cc49576f-sct4f_openstack(539eb117-6cac-4c4c-abe6-7840021f5233): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 11 02:21:51 crc kubenswrapper[4824]: E1211 02:21:51.735837 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-55cc49576f-sct4f" podUID="539eb117-6cac-4c4c-abe6-7840021f5233" Dec 11 02:21:51 crc kubenswrapper[4824]: I1211 02:21:51.794937 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-l45dd" Dec 11 02:21:51 crc kubenswrapper[4824]: I1211 02:21:51.870971 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f-combined-ca-bundle\") pod \"daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f\" (UID: \"daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f\") " Dec 11 02:21:51 crc kubenswrapper[4824]: I1211 02:21:51.871067 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f-fernet-keys\") pod \"daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f\" (UID: \"daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f\") " Dec 11 02:21:51 crc kubenswrapper[4824]: I1211 02:21:51.871224 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f-scripts\") pod \"daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f\" (UID: \"daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f\") " Dec 11 02:21:51 crc kubenswrapper[4824]: I1211 02:21:51.871272 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mp5wm\" (UniqueName: \"kubernetes.io/projected/daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f-kube-api-access-mp5wm\") pod \"daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f\" (UID: \"daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f\") " Dec 11 02:21:51 crc kubenswrapper[4824]: I1211 02:21:51.871327 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f-credential-keys\") pod \"daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f\" (UID: \"daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f\") " Dec 11 02:21:51 crc kubenswrapper[4824]: I1211 02:21:51.871352 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f-config-data\") pod \"daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f\" (UID: \"daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f\") " Dec 11 02:21:51 crc kubenswrapper[4824]: I1211 02:21:51.877802 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f-kube-api-access-mp5wm" (OuterVolumeSpecName: "kube-api-access-mp5wm") pod "daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f" (UID: "daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f"). InnerVolumeSpecName "kube-api-access-mp5wm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:21:51 crc kubenswrapper[4824]: I1211 02:21:51.878313 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f" (UID: "daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:21:51 crc kubenswrapper[4824]: I1211 02:21:51.878487 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f-scripts" (OuterVolumeSpecName: "scripts") pod "daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f" (UID: "daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:21:51 crc kubenswrapper[4824]: I1211 02:21:51.879210 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f" (UID: "daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:21:51 crc kubenswrapper[4824]: I1211 02:21:51.906728 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f" (UID: "daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:21:51 crc kubenswrapper[4824]: I1211 02:21:51.912194 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-l45dd" Dec 11 02:21:51 crc kubenswrapper[4824]: I1211 02:21:51.912237 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-l45dd" event={"ID":"daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f","Type":"ContainerDied","Data":"2d3935a901d32fee9fab9c6f749b8abde6996842fb034062a8a60344334460df"} Dec 11 02:21:51 crc kubenswrapper[4824]: I1211 02:21:51.912363 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2d3935a901d32fee9fab9c6f749b8abde6996842fb034062a8a60344334460df" Dec 11 02:21:51 crc kubenswrapper[4824]: I1211 02:21:51.936501 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f-config-data" (OuterVolumeSpecName: "config-data") pod "daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f" (UID: "daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:21:51 crc kubenswrapper[4824]: I1211 02:21:51.973965 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 02:21:51 crc kubenswrapper[4824]: I1211 02:21:51.974237 4824 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 11 02:21:51 crc kubenswrapper[4824]: I1211 02:21:51.974249 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 02:21:51 crc kubenswrapper[4824]: I1211 02:21:51.974288 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mp5wm\" (UniqueName: \"kubernetes.io/projected/daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f-kube-api-access-mp5wm\") on node \"crc\" DevicePath \"\"" Dec 11 02:21:51 crc kubenswrapper[4824]: I1211 02:21:51.974300 4824 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 11 02:21:51 crc kubenswrapper[4824]: I1211 02:21:51.974307 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 02:21:52 crc kubenswrapper[4824]: I1211 02:21:52.882851 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-l45dd"] Dec 11 02:21:52 crc kubenswrapper[4824]: I1211 02:21:52.892698 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-l45dd"] Dec 11 02:21:52 crc kubenswrapper[4824]: I1211 02:21:52.980854 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-752rf"] Dec 11 02:21:52 crc kubenswrapper[4824]: E1211 02:21:52.981655 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f" containerName="keystone-bootstrap" Dec 11 02:21:52 crc kubenswrapper[4824]: I1211 02:21:52.981726 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f" containerName="keystone-bootstrap" Dec 11 02:21:52 crc kubenswrapper[4824]: I1211 02:21:52.982008 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f" containerName="keystone-bootstrap" Dec 11 02:21:52 crc kubenswrapper[4824]: I1211 02:21:52.982648 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-752rf" Dec 11 02:21:52 crc kubenswrapper[4824]: I1211 02:21:52.985129 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-8rv79" Dec 11 02:21:52 crc kubenswrapper[4824]: I1211 02:21:52.985379 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 11 02:21:52 crc kubenswrapper[4824]: I1211 02:21:52.985554 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 11 02:21:52 crc kubenswrapper[4824]: I1211 02:21:52.985749 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 11 02:21:52 crc kubenswrapper[4824]: I1211 02:21:52.985902 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 11 02:21:52 crc kubenswrapper[4824]: I1211 02:21:52.989046 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-752rf"] Dec 11 02:21:53 crc kubenswrapper[4824]: I1211 02:21:53.094944 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68990bd4-eecf-4181-a76c-bf0d24154774-config-data\") pod \"keystone-bootstrap-752rf\" (UID: \"68990bd4-eecf-4181-a76c-bf0d24154774\") " pod="openstack/keystone-bootstrap-752rf" Dec 11 02:21:53 crc kubenswrapper[4824]: I1211 02:21:53.095163 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68990bd4-eecf-4181-a76c-bf0d24154774-combined-ca-bundle\") pod \"keystone-bootstrap-752rf\" (UID: \"68990bd4-eecf-4181-a76c-bf0d24154774\") " pod="openstack/keystone-bootstrap-752rf" Dec 11 02:21:53 crc kubenswrapper[4824]: I1211 02:21:53.095206 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c7skk\" (UniqueName: \"kubernetes.io/projected/68990bd4-eecf-4181-a76c-bf0d24154774-kube-api-access-c7skk\") pod \"keystone-bootstrap-752rf\" (UID: \"68990bd4-eecf-4181-a76c-bf0d24154774\") " pod="openstack/keystone-bootstrap-752rf" Dec 11 02:21:53 crc kubenswrapper[4824]: I1211 02:21:53.095341 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/68990bd4-eecf-4181-a76c-bf0d24154774-scripts\") pod \"keystone-bootstrap-752rf\" (UID: \"68990bd4-eecf-4181-a76c-bf0d24154774\") " pod="openstack/keystone-bootstrap-752rf" Dec 11 02:21:53 crc kubenswrapper[4824]: I1211 02:21:53.095402 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/68990bd4-eecf-4181-a76c-bf0d24154774-credential-keys\") pod \"keystone-bootstrap-752rf\" (UID: \"68990bd4-eecf-4181-a76c-bf0d24154774\") " pod="openstack/keystone-bootstrap-752rf" Dec 11 02:21:53 crc kubenswrapper[4824]: I1211 02:21:53.095660 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/68990bd4-eecf-4181-a76c-bf0d24154774-fernet-keys\") pod \"keystone-bootstrap-752rf\" (UID: \"68990bd4-eecf-4181-a76c-bf0d24154774\") " pod="openstack/keystone-bootstrap-752rf" Dec 11 02:21:53 crc kubenswrapper[4824]: I1211 02:21:53.197212 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/68990bd4-eecf-4181-a76c-bf0d24154774-scripts\") pod \"keystone-bootstrap-752rf\" (UID: \"68990bd4-eecf-4181-a76c-bf0d24154774\") " pod="openstack/keystone-bootstrap-752rf" Dec 11 02:21:53 crc kubenswrapper[4824]: I1211 02:21:53.197287 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/68990bd4-eecf-4181-a76c-bf0d24154774-credential-keys\") pod \"keystone-bootstrap-752rf\" (UID: \"68990bd4-eecf-4181-a76c-bf0d24154774\") " pod="openstack/keystone-bootstrap-752rf" Dec 11 02:21:53 crc kubenswrapper[4824]: I1211 02:21:53.197407 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/68990bd4-eecf-4181-a76c-bf0d24154774-fernet-keys\") pod \"keystone-bootstrap-752rf\" (UID: \"68990bd4-eecf-4181-a76c-bf0d24154774\") " pod="openstack/keystone-bootstrap-752rf" Dec 11 02:21:53 crc kubenswrapper[4824]: I1211 02:21:53.197496 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68990bd4-eecf-4181-a76c-bf0d24154774-config-data\") pod \"keystone-bootstrap-752rf\" (UID: \"68990bd4-eecf-4181-a76c-bf0d24154774\") " pod="openstack/keystone-bootstrap-752rf" Dec 11 02:21:53 crc kubenswrapper[4824]: I1211 02:21:53.197624 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c7skk\" (UniqueName: \"kubernetes.io/projected/68990bd4-eecf-4181-a76c-bf0d24154774-kube-api-access-c7skk\") pod \"keystone-bootstrap-752rf\" (UID: \"68990bd4-eecf-4181-a76c-bf0d24154774\") " pod="openstack/keystone-bootstrap-752rf" Dec 11 02:21:53 crc kubenswrapper[4824]: I1211 02:21:53.197664 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68990bd4-eecf-4181-a76c-bf0d24154774-combined-ca-bundle\") pod \"keystone-bootstrap-752rf\" (UID: \"68990bd4-eecf-4181-a76c-bf0d24154774\") " pod="openstack/keystone-bootstrap-752rf" Dec 11 02:21:53 crc kubenswrapper[4824]: I1211 02:21:53.203130 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/68990bd4-eecf-4181-a76c-bf0d24154774-credential-keys\") pod \"keystone-bootstrap-752rf\" (UID: \"68990bd4-eecf-4181-a76c-bf0d24154774\") " pod="openstack/keystone-bootstrap-752rf" Dec 11 02:21:53 crc kubenswrapper[4824]: I1211 02:21:53.203812 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68990bd4-eecf-4181-a76c-bf0d24154774-config-data\") pod \"keystone-bootstrap-752rf\" (UID: \"68990bd4-eecf-4181-a76c-bf0d24154774\") " pod="openstack/keystone-bootstrap-752rf" Dec 11 02:21:53 crc kubenswrapper[4824]: I1211 02:21:53.204459 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/68990bd4-eecf-4181-a76c-bf0d24154774-fernet-keys\") pod \"keystone-bootstrap-752rf\" (UID: \"68990bd4-eecf-4181-a76c-bf0d24154774\") " pod="openstack/keystone-bootstrap-752rf" Dec 11 02:21:53 crc kubenswrapper[4824]: I1211 02:21:53.205341 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/68990bd4-eecf-4181-a76c-bf0d24154774-scripts\") pod \"keystone-bootstrap-752rf\" (UID: \"68990bd4-eecf-4181-a76c-bf0d24154774\") " pod="openstack/keystone-bootstrap-752rf" Dec 11 02:21:53 crc kubenswrapper[4824]: I1211 02:21:53.205399 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68990bd4-eecf-4181-a76c-bf0d24154774-combined-ca-bundle\") pod \"keystone-bootstrap-752rf\" (UID: \"68990bd4-eecf-4181-a76c-bf0d24154774\") " pod="openstack/keystone-bootstrap-752rf" Dec 11 02:21:53 crc kubenswrapper[4824]: I1211 02:21:53.218008 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c7skk\" (UniqueName: \"kubernetes.io/projected/68990bd4-eecf-4181-a76c-bf0d24154774-kube-api-access-c7skk\") pod \"keystone-bootstrap-752rf\" (UID: \"68990bd4-eecf-4181-a76c-bf0d24154774\") " pod="openstack/keystone-bootstrap-752rf" Dec 11 02:21:53 crc kubenswrapper[4824]: I1211 02:21:53.304839 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-752rf" Dec 11 02:21:54 crc kubenswrapper[4824]: I1211 02:21:54.646288 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f" path="/var/lib/kubelet/pods/daf9ed2c-3c3a-4802-9a06-bc6e4ed2cc5f/volumes" Dec 11 02:21:56 crc kubenswrapper[4824]: I1211 02:21:56.956316 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-74f6bcbc87-4tss6" podUID="5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.130:5353: i/o timeout" Dec 11 02:21:58 crc kubenswrapper[4824]: I1211 02:21:58.865198 4824 scope.go:117] "RemoveContainer" containerID="50ee438d3d0bff2e98d14fa97e8da5909280aa6fe3eaf9c27b9e6c59e5c12125" Dec 11 02:21:59 crc kubenswrapper[4824]: E1211 02:21:59.239247 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified" Dec 11 02:21:59 crc kubenswrapper[4824]: E1211 02:21:59.239479 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n54h648h94h77h96h5bdh558h5f4h664h697h674h559h5bfh5c4h5d4h666h94h8bh96h5fchfbh6ch598h5c7h697h9h548h7chfdh676h64h644q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cqww2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(efa62db6-8875-4ae6-99ea-007e08be02bd): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 11 02:21:59 crc kubenswrapper[4824]: E1211 02:21:59.841240 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Dec 11 02:21:59 crc kubenswrapper[4824]: E1211 02:21:59.841690 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jh9z2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-pbsrf_openstack(d536a1b1-baa2-497a-ba0c-ad03599168cc): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 11 02:21:59 crc kubenswrapper[4824]: E1211 02:21:59.842867 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-pbsrf" podUID="d536a1b1-baa2-497a-ba0c-ad03599168cc" Dec 11 02:21:59 crc kubenswrapper[4824]: I1211 02:21:59.929811 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-55cc49576f-sct4f" Dec 11 02:21:59 crc kubenswrapper[4824]: I1211 02:21:59.937251 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7656dcd5fc-pvdtd" Dec 11 02:21:59 crc kubenswrapper[4824]: I1211 02:21:59.993039 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7656dcd5fc-pvdtd" Dec 11 02:21:59 crc kubenswrapper[4824]: I1211 02:21:59.993039 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7656dcd5fc-pvdtd" event={"ID":"97c61c1d-8fc0-46f3-9bb0-583a8c4aba1f","Type":"ContainerDied","Data":"3e28db81b0a324868b2b39a3b900d1a684da2db2d4786ac105c94fc9b19062f7"} Dec 11 02:21:59 crc kubenswrapper[4824]: I1211 02:21:59.996237 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-55cc49576f-sct4f" event={"ID":"539eb117-6cac-4c4c-abe6-7840021f5233","Type":"ContainerDied","Data":"346db3dfa2a85b752217e9c56567fb1b73235c430431b12b2bba32d0558c6dd6"} Dec 11 02:21:59 crc kubenswrapper[4824]: I1211 02:21:59.996257 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-55cc49576f-sct4f" Dec 11 02:21:59 crc kubenswrapper[4824]: E1211 02:21:59.997523 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-pbsrf" podUID="d536a1b1-baa2-497a-ba0c-ad03599168cc" Dec 11 02:22:00 crc kubenswrapper[4824]: I1211 02:22:00.027565 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/539eb117-6cac-4c4c-abe6-7840021f5233-config-data\") pod \"539eb117-6cac-4c4c-abe6-7840021f5233\" (UID: \"539eb117-6cac-4c4c-abe6-7840021f5233\") " Dec 11 02:22:00 crc kubenswrapper[4824]: I1211 02:22:00.027614 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/97c61c1d-8fc0-46f3-9bb0-583a8c4aba1f-config-data\") pod \"97c61c1d-8fc0-46f3-9bb0-583a8c4aba1f\" (UID: \"97c61c1d-8fc0-46f3-9bb0-583a8c4aba1f\") " Dec 11 02:22:00 crc kubenswrapper[4824]: I1211 02:22:00.027657 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/97c61c1d-8fc0-46f3-9bb0-583a8c4aba1f-horizon-secret-key\") pod \"97c61c1d-8fc0-46f3-9bb0-583a8c4aba1f\" (UID: \"97c61c1d-8fc0-46f3-9bb0-583a8c4aba1f\") " Dec 11 02:22:00 crc kubenswrapper[4824]: I1211 02:22:00.027675 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/539eb117-6cac-4c4c-abe6-7840021f5233-horizon-secret-key\") pod \"539eb117-6cac-4c4c-abe6-7840021f5233\" (UID: \"539eb117-6cac-4c4c-abe6-7840021f5233\") " Dec 11 02:22:00 crc kubenswrapper[4824]: I1211 02:22:00.027702 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/539eb117-6cac-4c4c-abe6-7840021f5233-logs\") pod \"539eb117-6cac-4c4c-abe6-7840021f5233\" (UID: \"539eb117-6cac-4c4c-abe6-7840021f5233\") " Dec 11 02:22:00 crc kubenswrapper[4824]: I1211 02:22:00.027722 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/97c61c1d-8fc0-46f3-9bb0-583a8c4aba1f-logs\") pod \"97c61c1d-8fc0-46f3-9bb0-583a8c4aba1f\" (UID: \"97c61c1d-8fc0-46f3-9bb0-583a8c4aba1f\") " Dec 11 02:22:00 crc kubenswrapper[4824]: I1211 02:22:00.027749 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7b6b\" (UniqueName: \"kubernetes.io/projected/539eb117-6cac-4c4c-abe6-7840021f5233-kube-api-access-x7b6b\") pod \"539eb117-6cac-4c4c-abe6-7840021f5233\" (UID: \"539eb117-6cac-4c4c-abe6-7840021f5233\") " Dec 11 02:22:00 crc kubenswrapper[4824]: I1211 02:22:00.027770 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcst2\" (UniqueName: \"kubernetes.io/projected/97c61c1d-8fc0-46f3-9bb0-583a8c4aba1f-kube-api-access-fcst2\") pod \"97c61c1d-8fc0-46f3-9bb0-583a8c4aba1f\" (UID: \"97c61c1d-8fc0-46f3-9bb0-583a8c4aba1f\") " Dec 11 02:22:00 crc kubenswrapper[4824]: I1211 02:22:00.027789 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/539eb117-6cac-4c4c-abe6-7840021f5233-scripts\") pod \"539eb117-6cac-4c4c-abe6-7840021f5233\" (UID: \"539eb117-6cac-4c4c-abe6-7840021f5233\") " Dec 11 02:22:00 crc kubenswrapper[4824]: I1211 02:22:00.027816 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/97c61c1d-8fc0-46f3-9bb0-583a8c4aba1f-scripts\") pod \"97c61c1d-8fc0-46f3-9bb0-583a8c4aba1f\" (UID: \"97c61c1d-8fc0-46f3-9bb0-583a8c4aba1f\") " Dec 11 02:22:00 crc kubenswrapper[4824]: I1211 02:22:00.028529 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/97c61c1d-8fc0-46f3-9bb0-583a8c4aba1f-logs" (OuterVolumeSpecName: "logs") pod "97c61c1d-8fc0-46f3-9bb0-583a8c4aba1f" (UID: "97c61c1d-8fc0-46f3-9bb0-583a8c4aba1f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:22:00 crc kubenswrapper[4824]: I1211 02:22:00.029017 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/539eb117-6cac-4c4c-abe6-7840021f5233-scripts" (OuterVolumeSpecName: "scripts") pod "539eb117-6cac-4c4c-abe6-7840021f5233" (UID: "539eb117-6cac-4c4c-abe6-7840021f5233"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:22:00 crc kubenswrapper[4824]: I1211 02:22:00.029087 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/539eb117-6cac-4c4c-abe6-7840021f5233-config-data" (OuterVolumeSpecName: "config-data") pod "539eb117-6cac-4c4c-abe6-7840021f5233" (UID: "539eb117-6cac-4c4c-abe6-7840021f5233"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:22:00 crc kubenswrapper[4824]: I1211 02:22:00.029179 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/539eb117-6cac-4c4c-abe6-7840021f5233-logs" (OuterVolumeSpecName: "logs") pod "539eb117-6cac-4c4c-abe6-7840021f5233" (UID: "539eb117-6cac-4c4c-abe6-7840021f5233"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:22:00 crc kubenswrapper[4824]: I1211 02:22:00.029522 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/97c61c1d-8fc0-46f3-9bb0-583a8c4aba1f-scripts" (OuterVolumeSpecName: "scripts") pod "97c61c1d-8fc0-46f3-9bb0-583a8c4aba1f" (UID: "97c61c1d-8fc0-46f3-9bb0-583a8c4aba1f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:22:00 crc kubenswrapper[4824]: I1211 02:22:00.029594 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/97c61c1d-8fc0-46f3-9bb0-583a8c4aba1f-config-data" (OuterVolumeSpecName: "config-data") pod "97c61c1d-8fc0-46f3-9bb0-583a8c4aba1f" (UID: "97c61c1d-8fc0-46f3-9bb0-583a8c4aba1f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:22:00 crc kubenswrapper[4824]: I1211 02:22:00.032743 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/539eb117-6cac-4c4c-abe6-7840021f5233-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "539eb117-6cac-4c4c-abe6-7840021f5233" (UID: "539eb117-6cac-4c4c-abe6-7840021f5233"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:22:00 crc kubenswrapper[4824]: I1211 02:22:00.032834 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97c61c1d-8fc0-46f3-9bb0-583a8c4aba1f-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "97c61c1d-8fc0-46f3-9bb0-583a8c4aba1f" (UID: "97c61c1d-8fc0-46f3-9bb0-583a8c4aba1f"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:22:00 crc kubenswrapper[4824]: I1211 02:22:00.033957 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97c61c1d-8fc0-46f3-9bb0-583a8c4aba1f-kube-api-access-fcst2" (OuterVolumeSpecName: "kube-api-access-fcst2") pod "97c61c1d-8fc0-46f3-9bb0-583a8c4aba1f" (UID: "97c61c1d-8fc0-46f3-9bb0-583a8c4aba1f"). InnerVolumeSpecName "kube-api-access-fcst2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:22:00 crc kubenswrapper[4824]: I1211 02:22:00.040278 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/539eb117-6cac-4c4c-abe6-7840021f5233-kube-api-access-x7b6b" (OuterVolumeSpecName: "kube-api-access-x7b6b") pod "539eb117-6cac-4c4c-abe6-7840021f5233" (UID: "539eb117-6cac-4c4c-abe6-7840021f5233"). InnerVolumeSpecName "kube-api-access-x7b6b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:22:00 crc kubenswrapper[4824]: I1211 02:22:00.129479 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/97c61c1d-8fc0-46f3-9bb0-583a8c4aba1f-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:00 crc kubenswrapper[4824]: I1211 02:22:00.129533 4824 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/97c61c1d-8fc0-46f3-9bb0-583a8c4aba1f-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:00 crc kubenswrapper[4824]: I1211 02:22:00.129555 4824 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/539eb117-6cac-4c4c-abe6-7840021f5233-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:00 crc kubenswrapper[4824]: I1211 02:22:00.129572 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/539eb117-6cac-4c4c-abe6-7840021f5233-logs\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:00 crc kubenswrapper[4824]: I1211 02:22:00.129588 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/97c61c1d-8fc0-46f3-9bb0-583a8c4aba1f-logs\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:00 crc kubenswrapper[4824]: I1211 02:22:00.129604 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7b6b\" (UniqueName: \"kubernetes.io/projected/539eb117-6cac-4c4c-abe6-7840021f5233-kube-api-access-x7b6b\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:00 crc kubenswrapper[4824]: I1211 02:22:00.129622 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcst2\" (UniqueName: \"kubernetes.io/projected/97c61c1d-8fc0-46f3-9bb0-583a8c4aba1f-kube-api-access-fcst2\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:00 crc kubenswrapper[4824]: I1211 02:22:00.129640 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/539eb117-6cac-4c4c-abe6-7840021f5233-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:00 crc kubenswrapper[4824]: I1211 02:22:00.129657 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/97c61c1d-8fc0-46f3-9bb0-583a8c4aba1f-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:00 crc kubenswrapper[4824]: I1211 02:22:00.129674 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/539eb117-6cac-4c4c-abe6-7840021f5233-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:00 crc kubenswrapper[4824]: I1211 02:22:00.365552 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7656dcd5fc-pvdtd"] Dec 11 02:22:00 crc kubenswrapper[4824]: I1211 02:22:00.375601 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7656dcd5fc-pvdtd"] Dec 11 02:22:00 crc kubenswrapper[4824]: I1211 02:22:00.397838 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-55cc49576f-sct4f"] Dec 11 02:22:00 crc kubenswrapper[4824]: I1211 02:22:00.404872 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-55cc49576f-sct4f"] Dec 11 02:22:00 crc kubenswrapper[4824]: I1211 02:22:00.642960 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="539eb117-6cac-4c4c-abe6-7840021f5233" path="/var/lib/kubelet/pods/539eb117-6cac-4c4c-abe6-7840021f5233/volumes" Dec 11 02:22:00 crc kubenswrapper[4824]: I1211 02:22:00.643455 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="97c61c1d-8fc0-46f3-9bb0-583a8c4aba1f" path="/var/lib/kubelet/pods/97c61c1d-8fc0-46f3-9bb0-583a8c4aba1f/volumes" Dec 11 02:22:00 crc kubenswrapper[4824]: E1211 02:22:00.877853 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Dec 11 02:22:00 crc kubenswrapper[4824]: E1211 02:22:00.878018 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vgswv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-tn658_openstack(8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 11 02:22:00 crc kubenswrapper[4824]: I1211 02:22:00.878334 4824 scope.go:117] "RemoveContainer" containerID="7bc894eb0df73abeaa4452cb22e48e8207ff47741a99a1f8b733b49245f935d6" Dec 11 02:22:00 crc kubenswrapper[4824]: E1211 02:22:00.878983 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7bc894eb0df73abeaa4452cb22e48e8207ff47741a99a1f8b733b49245f935d6\": container with ID starting with 7bc894eb0df73abeaa4452cb22e48e8207ff47741a99a1f8b733b49245f935d6 not found: ID does not exist" containerID="7bc894eb0df73abeaa4452cb22e48e8207ff47741a99a1f8b733b49245f935d6" Dec 11 02:22:00 crc kubenswrapper[4824]: I1211 02:22:00.879014 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7bc894eb0df73abeaa4452cb22e48e8207ff47741a99a1f8b733b49245f935d6"} err="failed to get container status \"7bc894eb0df73abeaa4452cb22e48e8207ff47741a99a1f8b733b49245f935d6\": rpc error: code = NotFound desc = could not find container \"7bc894eb0df73abeaa4452cb22e48e8207ff47741a99a1f8b733b49245f935d6\": container with ID starting with 7bc894eb0df73abeaa4452cb22e48e8207ff47741a99a1f8b733b49245f935d6 not found: ID does not exist" Dec 11 02:22:00 crc kubenswrapper[4824]: I1211 02:22:00.879036 4824 scope.go:117] "RemoveContainer" containerID="50ee438d3d0bff2e98d14fa97e8da5909280aa6fe3eaf9c27b9e6c59e5c12125" Dec 11 02:22:00 crc kubenswrapper[4824]: E1211 02:22:00.879535 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"50ee438d3d0bff2e98d14fa97e8da5909280aa6fe3eaf9c27b9e6c59e5c12125\": container with ID starting with 50ee438d3d0bff2e98d14fa97e8da5909280aa6fe3eaf9c27b9e6c59e5c12125 not found: ID does not exist" containerID="50ee438d3d0bff2e98d14fa97e8da5909280aa6fe3eaf9c27b9e6c59e5c12125" Dec 11 02:22:00 crc kubenswrapper[4824]: I1211 02:22:00.879550 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50ee438d3d0bff2e98d14fa97e8da5909280aa6fe3eaf9c27b9e6c59e5c12125"} err="failed to get container status \"50ee438d3d0bff2e98d14fa97e8da5909280aa6fe3eaf9c27b9e6c59e5c12125\": rpc error: code = NotFound desc = could not find container \"50ee438d3d0bff2e98d14fa97e8da5909280aa6fe3eaf9c27b9e6c59e5c12125\": container with ID starting with 50ee438d3d0bff2e98d14fa97e8da5909280aa6fe3eaf9c27b9e6c59e5c12125 not found: ID does not exist" Dec 11 02:22:00 crc kubenswrapper[4824]: I1211 02:22:00.879562 4824 scope.go:117] "RemoveContainer" containerID="7bc894eb0df73abeaa4452cb22e48e8207ff47741a99a1f8b733b49245f935d6" Dec 11 02:22:00 crc kubenswrapper[4824]: I1211 02:22:00.879837 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7bc894eb0df73abeaa4452cb22e48e8207ff47741a99a1f8b733b49245f935d6"} err="failed to get container status \"7bc894eb0df73abeaa4452cb22e48e8207ff47741a99a1f8b733b49245f935d6\": rpc error: code = NotFound desc = could not find container \"7bc894eb0df73abeaa4452cb22e48e8207ff47741a99a1f8b733b49245f935d6\": container with ID starting with 7bc894eb0df73abeaa4452cb22e48e8207ff47741a99a1f8b733b49245f935d6 not found: ID does not exist" Dec 11 02:22:00 crc kubenswrapper[4824]: I1211 02:22:00.879874 4824 scope.go:117] "RemoveContainer" containerID="50ee438d3d0bff2e98d14fa97e8da5909280aa6fe3eaf9c27b9e6c59e5c12125" Dec 11 02:22:00 crc kubenswrapper[4824]: I1211 02:22:00.880144 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50ee438d3d0bff2e98d14fa97e8da5909280aa6fe3eaf9c27b9e6c59e5c12125"} err="failed to get container status \"50ee438d3d0bff2e98d14fa97e8da5909280aa6fe3eaf9c27b9e6c59e5c12125\": rpc error: code = NotFound desc = could not find container \"50ee438d3d0bff2e98d14fa97e8da5909280aa6fe3eaf9c27b9e6c59e5c12125\": container with ID starting with 50ee438d3d0bff2e98d14fa97e8da5909280aa6fe3eaf9c27b9e6c59e5c12125 not found: ID does not exist" Dec 11 02:22:00 crc kubenswrapper[4824]: I1211 02:22:00.880160 4824 scope.go:117] "RemoveContainer" containerID="cf5110cb19c13deb1b1e0dcf9c2f0ba6163d78866f03c52187c7b51fa1b4cff5" Dec 11 02:22:00 crc kubenswrapper[4824]: E1211 02:22:00.885911 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-tn658" podUID="8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7" Dec 11 02:22:01 crc kubenswrapper[4824]: I1211 02:22:01.019496 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-4tss6" Dec 11 02:22:01 crc kubenswrapper[4824]: I1211 02:22:01.019710 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-4tss6" event={"ID":"5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00","Type":"ContainerDied","Data":"fedd41080434400a86cab37fd05ed9bf68b01384fa678fb9e8019221752f22cd"} Dec 11 02:22:01 crc kubenswrapper[4824]: I1211 02:22:01.044785 4824 scope.go:117] "RemoveContainer" containerID="0a8851eabb0da1bc53638228e1741f6f6ce45fdf946a6207eaf2c47b8e46f6dc" Dec 11 02:22:01 crc kubenswrapper[4824]: E1211 02:22:01.049163 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-tn658" podUID="8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7" Dec 11 02:22:01 crc kubenswrapper[4824]: I1211 02:22:01.049431 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00-dns-svc\") pod \"5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00\" (UID: \"5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00\") " Dec 11 02:22:01 crc kubenswrapper[4824]: I1211 02:22:01.124835 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00" (UID: "5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:22:01 crc kubenswrapper[4824]: I1211 02:22:01.150923 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00-ovsdbserver-nb\") pod \"5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00\" (UID: \"5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00\") " Dec 11 02:22:01 crc kubenswrapper[4824]: I1211 02:22:01.150981 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00-ovsdbserver-sb\") pod \"5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00\" (UID: \"5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00\") " Dec 11 02:22:01 crc kubenswrapper[4824]: I1211 02:22:01.151182 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00-config\") pod \"5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00\" (UID: \"5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00\") " Dec 11 02:22:01 crc kubenswrapper[4824]: I1211 02:22:01.151229 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mmltc\" (UniqueName: \"kubernetes.io/projected/5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00-kube-api-access-mmltc\") pod \"5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00\" (UID: \"5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00\") " Dec 11 02:22:01 crc kubenswrapper[4824]: I1211 02:22:01.151254 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00-dns-swift-storage-0\") pod \"5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00\" (UID: \"5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00\") " Dec 11 02:22:01 crc kubenswrapper[4824]: I1211 02:22:01.151699 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:01 crc kubenswrapper[4824]: I1211 02:22:01.155582 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00-kube-api-access-mmltc" (OuterVolumeSpecName: "kube-api-access-mmltc") pod "5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00" (UID: "5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00"). InnerVolumeSpecName "kube-api-access-mmltc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:22:01 crc kubenswrapper[4824]: I1211 02:22:01.177505 4824 scope.go:117] "RemoveContainer" containerID="ae65a91d118ac4d63800cb6c6bcfa72370d4dec5921ffa4c291a7fe609becdf5" Dec 11 02:22:01 crc kubenswrapper[4824]: I1211 02:22:01.200243 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00" (UID: "5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:22:01 crc kubenswrapper[4824]: I1211 02:22:01.221266 4824 scope.go:117] "RemoveContainer" containerID="63d0fc2a3cdbd5be8649bacd15b9dadb1f6545b630b7a004fdef91e677c73065" Dec 11 02:22:01 crc kubenswrapper[4824]: I1211 02:22:01.240663 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00-config" (OuterVolumeSpecName: "config") pod "5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00" (UID: "5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:22:01 crc kubenswrapper[4824]: I1211 02:22:01.248918 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00" (UID: "5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:22:01 crc kubenswrapper[4824]: I1211 02:22:01.249710 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00" (UID: "5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:22:01 crc kubenswrapper[4824]: I1211 02:22:01.256895 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:01 crc kubenswrapper[4824]: I1211 02:22:01.256923 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:01 crc kubenswrapper[4824]: I1211 02:22:01.257528 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00-config\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:01 crc kubenswrapper[4824]: I1211 02:22:01.257550 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mmltc\" (UniqueName: \"kubernetes.io/projected/5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00-kube-api-access-mmltc\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:01 crc kubenswrapper[4824]: I1211 02:22:01.257562 4824 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:01 crc kubenswrapper[4824]: I1211 02:22:01.451893 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 11 02:22:01 crc kubenswrapper[4824]: I1211 02:22:01.511402 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-74cc6c88d8-zvlk7"] Dec 11 02:22:01 crc kubenswrapper[4824]: I1211 02:22:01.521854 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-787649dcdd-ts6qs"] Dec 11 02:22:01 crc kubenswrapper[4824]: I1211 02:22:01.532135 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-752rf"] Dec 11 02:22:01 crc kubenswrapper[4824]: I1211 02:22:01.573103 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 11 02:22:01 crc kubenswrapper[4824]: W1211 02:22:01.703250 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod68990bd4_eecf_4181_a76c_bf0d24154774.slice/crio-dea64b06dcc752831f405e43ed8d5ded3e916e78358991cf31bcf8cfe2ce51de WatchSource:0}: Error finding container dea64b06dcc752831f405e43ed8d5ded3e916e78358991cf31bcf8cfe2ce51de: Status 404 returned error can't find the container with id dea64b06dcc752831f405e43ed8d5ded3e916e78358991cf31bcf8cfe2ce51de Dec 11 02:22:01 crc kubenswrapper[4824]: W1211 02:22:01.705676 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod189f7197_9da3_402f_9c9b_042e34ca6556.slice/crio-c0fc742b51065576f878140173b3938b377a748bc19f9230d9807d760bfdf2b1 WatchSource:0}: Error finding container c0fc742b51065576f878140173b3938b377a748bc19f9230d9807d760bfdf2b1: Status 404 returned error can't find the container with id c0fc742b51065576f878140173b3938b377a748bc19f9230d9807d760bfdf2b1 Dec 11 02:22:01 crc kubenswrapper[4824]: W1211 02:22:01.712727 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf15f9ed0_98f2_42fe_8006_2142098cc2b6.slice/crio-1be3ac260841695de97d022c86cb5be498b8cbff79af2a73fc724347cc80c50a WatchSource:0}: Error finding container 1be3ac260841695de97d022c86cb5be498b8cbff79af2a73fc724347cc80c50a: Status 404 returned error can't find the container with id 1be3ac260841695de97d022c86cb5be498b8cbff79af2a73fc724347cc80c50a Dec 11 02:22:01 crc kubenswrapper[4824]: W1211 02:22:01.716024 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddfa654c8_1f3b_43f6_8a1f_709a05020b4e.slice/crio-af7573a4ac2063e4700b5375272ebf15aa63938d09730714f024d29f7f1da940 WatchSource:0}: Error finding container af7573a4ac2063e4700b5375272ebf15aa63938d09730714f024d29f7f1da940: Status 404 returned error can't find the container with id af7573a4ac2063e4700b5375272ebf15aa63938d09730714f024d29f7f1da940 Dec 11 02:22:01 crc kubenswrapper[4824]: I1211 02:22:01.720465 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 11 02:22:01 crc kubenswrapper[4824]: I1211 02:22:01.956906 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-74f6bcbc87-4tss6" podUID="5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.130:5353: i/o timeout" Dec 11 02:22:01 crc kubenswrapper[4824]: I1211 02:22:01.956983 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-74f6bcbc87-4tss6" Dec 11 02:22:02 crc kubenswrapper[4824]: I1211 02:22:02.057853 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"189f7197-9da3-402f-9c9b-042e34ca6556","Type":"ContainerStarted","Data":"c0fc742b51065576f878140173b3938b377a748bc19f9230d9807d760bfdf2b1"} Dec 11 02:22:02 crc kubenswrapper[4824]: I1211 02:22:02.061171 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-787649dcdd-ts6qs" event={"ID":"f15f9ed0-98f2-42fe-8006-2142098cc2b6","Type":"ContainerStarted","Data":"1be3ac260841695de97d022c86cb5be498b8cbff79af2a73fc724347cc80c50a"} Dec 11 02:22:02 crc kubenswrapper[4824]: I1211 02:22:02.062856 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-74cc6c88d8-zvlk7" event={"ID":"1da5eb01-d709-42ad-b5f1-e7ea6310cc9e","Type":"ContainerStarted","Data":"2c15058b075c5cf8b0352a10c4459cdb75614a2ff53931bbb9cb0350e66c067f"} Dec 11 02:22:02 crc kubenswrapper[4824]: I1211 02:22:02.062907 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-74cc6c88d8-zvlk7" event={"ID":"1da5eb01-d709-42ad-b5f1-e7ea6310cc9e","Type":"ContainerStarted","Data":"87b5ac301bed80b7b9684aadb661a193d07428e35f4c9210af502bd030ca76cf"} Dec 11 02:22:02 crc kubenswrapper[4824]: I1211 02:22:02.064191 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-4tss6" Dec 11 02:22:02 crc kubenswrapper[4824]: I1211 02:22:02.086180 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-752rf" event={"ID":"68990bd4-eecf-4181-a76c-bf0d24154774","Type":"ContainerStarted","Data":"dea64b06dcc752831f405e43ed8d5ded3e916e78358991cf31bcf8cfe2ce51de"} Dec 11 02:22:02 crc kubenswrapper[4824]: I1211 02:22:02.088183 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"dfa654c8-1f3b-43f6-8a1f-709a05020b4e","Type":"ContainerStarted","Data":"af7573a4ac2063e4700b5375272ebf15aa63938d09730714f024d29f7f1da940"} Dec 11 02:22:02 crc kubenswrapper[4824]: I1211 02:22:02.090140 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-lpwl8" event={"ID":"68cc858a-9296-4a34-8322-b2dabc8ac1a2","Type":"ContainerStarted","Data":"a01fcd21e235ca3330d07c56038ef155168ad9722f0844751b85828e6c035962"} Dec 11 02:22:02 crc kubenswrapper[4824]: I1211 02:22:02.103913 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5f68f577dc-nj2zp" event={"ID":"d7637730-463d-4742-a5fb-4ed6029edd61","Type":"ContainerStarted","Data":"f5723003639a4381041359444637575c08e0809c0fa337753ae1b52e305e1592"} Dec 11 02:22:02 crc kubenswrapper[4824]: I1211 02:22:02.103944 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5f68f577dc-nj2zp" event={"ID":"d7637730-463d-4742-a5fb-4ed6029edd61","Type":"ContainerStarted","Data":"61677087639bc33f5f228d3d104a74c1b37ee383d8903ad663d4f217b067440a"} Dec 11 02:22:02 crc kubenswrapper[4824]: I1211 02:22:02.104049 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-5f68f577dc-nj2zp" podUID="d7637730-463d-4742-a5fb-4ed6029edd61" containerName="horizon-log" containerID="cri-o://61677087639bc33f5f228d3d104a74c1b37ee383d8903ad663d4f217b067440a" gracePeriod=30 Dec 11 02:22:02 crc kubenswrapper[4824]: I1211 02:22:02.104276 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-5f68f577dc-nj2zp" podUID="d7637730-463d-4742-a5fb-4ed6029edd61" containerName="horizon" containerID="cri-o://f5723003639a4381041359444637575c08e0809c0fa337753ae1b52e305e1592" gracePeriod=30 Dec 11 02:22:02 crc kubenswrapper[4824]: I1211 02:22:02.106065 4824 generic.go:334] "Generic (PLEG): container finished" podID="07b76bc0-c213-45d8-bebd-06b3e70322ab" containerID="d38997cd7f66152cf930c9deb794b8e30589e769a2b2d85afbc34bbe959d1c6b" exitCode=0 Dec 11 02:22:02 crc kubenswrapper[4824]: I1211 02:22:02.106106 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-nm5bg" event={"ID":"07b76bc0-c213-45d8-bebd-06b3e70322ab","Type":"ContainerDied","Data":"d38997cd7f66152cf930c9deb794b8e30589e769a2b2d85afbc34bbe959d1c6b"} Dec 11 02:22:02 crc kubenswrapper[4824]: I1211 02:22:02.118894 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-lpwl8" podStartSLOduration=3.659667652 podStartE2EDuration="28.11887295s" podCreationTimestamp="2025-12-11 02:21:34 +0000 UTC" firstStartedPulling="2025-12-11 02:21:35.38728697 +0000 UTC m=+1237.076324349" lastFinishedPulling="2025-12-11 02:21:59.846492258 +0000 UTC m=+1261.535529647" observedRunningTime="2025-12-11 02:22:02.110793818 +0000 UTC m=+1263.799831197" watchObservedRunningTime="2025-12-11 02:22:02.11887295 +0000 UTC m=+1263.807910329" Dec 11 02:22:02 crc kubenswrapper[4824]: I1211 02:22:02.144571 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-5f68f577dc-nj2zp" podStartSLOduration=3.7172614619999997 podStartE2EDuration="26.144552494s" podCreationTimestamp="2025-12-11 02:21:36 +0000 UTC" firstStartedPulling="2025-12-11 02:21:37.416394399 +0000 UTC m=+1239.105431788" lastFinishedPulling="2025-12-11 02:21:59.843685441 +0000 UTC m=+1261.532722820" observedRunningTime="2025-12-11 02:22:02.130634662 +0000 UTC m=+1263.819672041" watchObservedRunningTime="2025-12-11 02:22:02.144552494 +0000 UTC m=+1263.833589873" Dec 11 02:22:02 crc kubenswrapper[4824]: I1211 02:22:02.162585 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-4tss6"] Dec 11 02:22:02 crc kubenswrapper[4824]: I1211 02:22:02.170453 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-4tss6"] Dec 11 02:22:02 crc kubenswrapper[4824]: I1211 02:22:02.649731 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00" path="/var/lib/kubelet/pods/5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00/volumes" Dec 11 02:22:03 crc kubenswrapper[4824]: I1211 02:22:03.125863 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"efa62db6-8875-4ae6-99ea-007e08be02bd","Type":"ContainerStarted","Data":"765b4a0946857ba769223a1fb59d99286c294efe426c05d0a350f8fc43d2044e"} Dec 11 02:22:03 crc kubenswrapper[4824]: I1211 02:22:03.130738 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"dfa654c8-1f3b-43f6-8a1f-709a05020b4e","Type":"ContainerStarted","Data":"62aff7415eaffd64df1b48803198a5006854ea9c0d791bee37ad995fe18a6352"} Dec 11 02:22:03 crc kubenswrapper[4824]: I1211 02:22:03.133121 4824 generic.go:334] "Generic (PLEG): container finished" podID="68cc858a-9296-4a34-8322-b2dabc8ac1a2" containerID="a01fcd21e235ca3330d07c56038ef155168ad9722f0844751b85828e6c035962" exitCode=0 Dec 11 02:22:03 crc kubenswrapper[4824]: I1211 02:22:03.133173 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-lpwl8" event={"ID":"68cc858a-9296-4a34-8322-b2dabc8ac1a2","Type":"ContainerDied","Data":"a01fcd21e235ca3330d07c56038ef155168ad9722f0844751b85828e6c035962"} Dec 11 02:22:03 crc kubenswrapper[4824]: I1211 02:22:03.160703 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-74cc6c88d8-zvlk7" event={"ID":"1da5eb01-d709-42ad-b5f1-e7ea6310cc9e","Type":"ContainerStarted","Data":"fd2b2665bd7c0421188e0973c817cf28b77e8627bcddad1ebd3b57379b433fb3"} Dec 11 02:22:03 crc kubenswrapper[4824]: I1211 02:22:03.174299 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-752rf" event={"ID":"68990bd4-eecf-4181-a76c-bf0d24154774","Type":"ContainerStarted","Data":"61c05165b5969b15e231ae24f60dcbd4ca978cb8d81f0055fe9a2ab1ba1519d5"} Dec 11 02:22:03 crc kubenswrapper[4824]: I1211 02:22:03.190895 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"189f7197-9da3-402f-9c9b-042e34ca6556","Type":"ContainerStarted","Data":"e2fe9aca09a3e9a571c71da03de60b9dd22f41ed17659243623d520c5e2090de"} Dec 11 02:22:03 crc kubenswrapper[4824]: I1211 02:22:03.191028 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="189f7197-9da3-402f-9c9b-042e34ca6556" containerName="glance-log" containerID="cri-o://e2fe9aca09a3e9a571c71da03de60b9dd22f41ed17659243623d520c5e2090de" gracePeriod=30 Dec 11 02:22:03 crc kubenswrapper[4824]: I1211 02:22:03.191254 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="189f7197-9da3-402f-9c9b-042e34ca6556" containerName="glance-httpd" containerID="cri-o://38b57fc2e66e0271d14b42c29f82423a12edb93639980b09631c27764378933f" gracePeriod=30 Dec 11 02:22:03 crc kubenswrapper[4824]: I1211 02:22:03.194811 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-74cc6c88d8-zvlk7" podStartSLOduration=20.194786587 podStartE2EDuration="20.194786587s" podCreationTimestamp="2025-12-11 02:21:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:22:03.181887013 +0000 UTC m=+1264.870924402" watchObservedRunningTime="2025-12-11 02:22:03.194786587 +0000 UTC m=+1264.883823966" Dec 11 02:22:03 crc kubenswrapper[4824]: I1211 02:22:03.202161 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-787649dcdd-ts6qs" event={"ID":"f15f9ed0-98f2-42fe-8006-2142098cc2b6","Type":"ContainerStarted","Data":"588925740a1286914b8db0f80b0f9426464ecc031e105965e01b1abcbf40a9ba"} Dec 11 02:22:03 crc kubenswrapper[4824]: I1211 02:22:03.202298 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-787649dcdd-ts6qs" event={"ID":"f15f9ed0-98f2-42fe-8006-2142098cc2b6","Type":"ContainerStarted","Data":"e4de9df2d8c6dc730220bfb406d58ac47fefcc07528a72cb76ac4ea026179a39"} Dec 11 02:22:03 crc kubenswrapper[4824]: I1211 02:22:03.204745 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-752rf" podStartSLOduration=11.204726149 podStartE2EDuration="11.204726149s" podCreationTimestamp="2025-12-11 02:21:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:22:03.198005495 +0000 UTC m=+1264.887042894" watchObservedRunningTime="2025-12-11 02:22:03.204726149 +0000 UTC m=+1264.893763528" Dec 11 02:22:03 crc kubenswrapper[4824]: I1211 02:22:03.249148 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=24.249134417 podStartE2EDuration="24.249134417s" podCreationTimestamp="2025-12-11 02:21:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:22:03.224528462 +0000 UTC m=+1264.913565831" watchObservedRunningTime="2025-12-11 02:22:03.249134417 +0000 UTC m=+1264.938171786" Dec 11 02:22:03 crc kubenswrapper[4824]: I1211 02:22:03.257723 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-787649dcdd-ts6qs" podStartSLOduration=20.257706782 podStartE2EDuration="20.257706782s" podCreationTimestamp="2025-12-11 02:21:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:22:03.247546343 +0000 UTC m=+1264.936583722" watchObservedRunningTime="2025-12-11 02:22:03.257706782 +0000 UTC m=+1264.946744161" Dec 11 02:22:03 crc kubenswrapper[4824]: I1211 02:22:03.560391 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-787649dcdd-ts6qs" Dec 11 02:22:03 crc kubenswrapper[4824]: I1211 02:22:03.560944 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-787649dcdd-ts6qs" Dec 11 02:22:03 crc kubenswrapper[4824]: I1211 02:22:03.575141 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-74cc6c88d8-zvlk7" Dec 11 02:22:03 crc kubenswrapper[4824]: I1211 02:22:03.575184 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-74cc6c88d8-zvlk7" Dec 11 02:22:03 crc kubenswrapper[4824]: I1211 02:22:03.654879 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-nm5bg" Dec 11 02:22:03 crc kubenswrapper[4824]: I1211 02:22:03.837764 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4jdvp\" (UniqueName: \"kubernetes.io/projected/07b76bc0-c213-45d8-bebd-06b3e70322ab-kube-api-access-4jdvp\") pod \"07b76bc0-c213-45d8-bebd-06b3e70322ab\" (UID: \"07b76bc0-c213-45d8-bebd-06b3e70322ab\") " Dec 11 02:22:03 crc kubenswrapper[4824]: I1211 02:22:03.837835 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/07b76bc0-c213-45d8-bebd-06b3e70322ab-config\") pod \"07b76bc0-c213-45d8-bebd-06b3e70322ab\" (UID: \"07b76bc0-c213-45d8-bebd-06b3e70322ab\") " Dec 11 02:22:03 crc kubenswrapper[4824]: I1211 02:22:03.837945 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07b76bc0-c213-45d8-bebd-06b3e70322ab-combined-ca-bundle\") pod \"07b76bc0-c213-45d8-bebd-06b3e70322ab\" (UID: \"07b76bc0-c213-45d8-bebd-06b3e70322ab\") " Dec 11 02:22:03 crc kubenswrapper[4824]: I1211 02:22:03.844403 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/07b76bc0-c213-45d8-bebd-06b3e70322ab-kube-api-access-4jdvp" (OuterVolumeSpecName: "kube-api-access-4jdvp") pod "07b76bc0-c213-45d8-bebd-06b3e70322ab" (UID: "07b76bc0-c213-45d8-bebd-06b3e70322ab"). InnerVolumeSpecName "kube-api-access-4jdvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:22:03 crc kubenswrapper[4824]: I1211 02:22:03.862840 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07b76bc0-c213-45d8-bebd-06b3e70322ab-config" (OuterVolumeSpecName: "config") pod "07b76bc0-c213-45d8-bebd-06b3e70322ab" (UID: "07b76bc0-c213-45d8-bebd-06b3e70322ab"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:22:03 crc kubenswrapper[4824]: I1211 02:22:03.863329 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07b76bc0-c213-45d8-bebd-06b3e70322ab-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "07b76bc0-c213-45d8-bebd-06b3e70322ab" (UID: "07b76bc0-c213-45d8-bebd-06b3e70322ab"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:22:03 crc kubenswrapper[4824]: I1211 02:22:03.940882 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07b76bc0-c213-45d8-bebd-06b3e70322ab-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:03 crc kubenswrapper[4824]: I1211 02:22:03.940910 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4jdvp\" (UniqueName: \"kubernetes.io/projected/07b76bc0-c213-45d8-bebd-06b3e70322ab-kube-api-access-4jdvp\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:03 crc kubenswrapper[4824]: I1211 02:22:03.940919 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/07b76bc0-c213-45d8-bebd-06b3e70322ab-config\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:04 crc kubenswrapper[4824]: I1211 02:22:04.218957 4824 generic.go:334] "Generic (PLEG): container finished" podID="189f7197-9da3-402f-9c9b-042e34ca6556" containerID="38b57fc2e66e0271d14b42c29f82423a12edb93639980b09631c27764378933f" exitCode=143 Dec 11 02:22:04 crc kubenswrapper[4824]: I1211 02:22:04.219295 4824 generic.go:334] "Generic (PLEG): container finished" podID="189f7197-9da3-402f-9c9b-042e34ca6556" containerID="e2fe9aca09a3e9a571c71da03de60b9dd22f41ed17659243623d520c5e2090de" exitCode=143 Dec 11 02:22:04 crc kubenswrapper[4824]: I1211 02:22:04.219025 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"189f7197-9da3-402f-9c9b-042e34ca6556","Type":"ContainerDied","Data":"38b57fc2e66e0271d14b42c29f82423a12edb93639980b09631c27764378933f"} Dec 11 02:22:04 crc kubenswrapper[4824]: I1211 02:22:04.219375 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"189f7197-9da3-402f-9c9b-042e34ca6556","Type":"ContainerDied","Data":"e2fe9aca09a3e9a571c71da03de60b9dd22f41ed17659243623d520c5e2090de"} Dec 11 02:22:04 crc kubenswrapper[4824]: I1211 02:22:04.221267 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-nm5bg" event={"ID":"07b76bc0-c213-45d8-bebd-06b3e70322ab","Type":"ContainerDied","Data":"abbb752d7aa8e3149ae4ec2b431e8cdf4d21d3617f5f01e0405c39ecb5e0a48d"} Dec 11 02:22:04 crc kubenswrapper[4824]: I1211 02:22:04.221297 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="abbb752d7aa8e3149ae4ec2b431e8cdf4d21d3617f5f01e0405c39ecb5e0a48d" Dec 11 02:22:04 crc kubenswrapper[4824]: I1211 02:22:04.221303 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-nm5bg" Dec 11 02:22:04 crc kubenswrapper[4824]: I1211 02:22:04.226201 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"dfa654c8-1f3b-43f6-8a1f-709a05020b4e","Type":"ContainerStarted","Data":"ed47e0f4bf032969946a71cb7a3c67c26e5ce39ba4f3607e54fa38b0c5b34968"} Dec 11 02:22:04 crc kubenswrapper[4824]: I1211 02:22:04.227303 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="dfa654c8-1f3b-43f6-8a1f-709a05020b4e" containerName="glance-log" containerID="cri-o://62aff7415eaffd64df1b48803198a5006854ea9c0d791bee37ad995fe18a6352" gracePeriod=30 Dec 11 02:22:04 crc kubenswrapper[4824]: I1211 02:22:04.227360 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="dfa654c8-1f3b-43f6-8a1f-709a05020b4e" containerName="glance-httpd" containerID="cri-o://ed47e0f4bf032969946a71cb7a3c67c26e5ce39ba4f3607e54fa38b0c5b34968" gracePeriod=30 Dec 11 02:22:04 crc kubenswrapper[4824]: I1211 02:22:04.255785 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=22.255766424 podStartE2EDuration="22.255766424s" podCreationTimestamp="2025-12-11 02:21:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:22:04.249352008 +0000 UTC m=+1265.938389387" watchObservedRunningTime="2025-12-11 02:22:04.255766424 +0000 UTC m=+1265.944803793" Dec 11 02:22:04 crc kubenswrapper[4824]: I1211 02:22:04.368865 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-6wc4l"] Dec 11 02:22:04 crc kubenswrapper[4824]: E1211 02:22:04.369346 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00" containerName="init" Dec 11 02:22:04 crc kubenswrapper[4824]: I1211 02:22:04.369360 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00" containerName="init" Dec 11 02:22:04 crc kubenswrapper[4824]: E1211 02:22:04.369372 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00" containerName="dnsmasq-dns" Dec 11 02:22:04 crc kubenswrapper[4824]: I1211 02:22:04.369379 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00" containerName="dnsmasq-dns" Dec 11 02:22:04 crc kubenswrapper[4824]: E1211 02:22:04.369411 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07b76bc0-c213-45d8-bebd-06b3e70322ab" containerName="neutron-db-sync" Dec 11 02:22:04 crc kubenswrapper[4824]: I1211 02:22:04.369419 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="07b76bc0-c213-45d8-bebd-06b3e70322ab" containerName="neutron-db-sync" Dec 11 02:22:04 crc kubenswrapper[4824]: I1211 02:22:04.369603 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a1e4ee5-c52c-4c35-9978-2d03c2ba0d00" containerName="dnsmasq-dns" Dec 11 02:22:04 crc kubenswrapper[4824]: I1211 02:22:04.369623 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="07b76bc0-c213-45d8-bebd-06b3e70322ab" containerName="neutron-db-sync" Dec 11 02:22:04 crc kubenswrapper[4824]: I1211 02:22:04.370981 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-6wc4l" Dec 11 02:22:04 crc kubenswrapper[4824]: I1211 02:22:04.393735 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-6wc4l"] Dec 11 02:22:04 crc kubenswrapper[4824]: I1211 02:22:04.465237 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/89133b02-fcf7-4fe9-b511-c406d1b0d08b-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-6wc4l\" (UID: \"89133b02-fcf7-4fe9-b511-c406d1b0d08b\") " pod="openstack/dnsmasq-dns-55f844cf75-6wc4l" Dec 11 02:22:04 crc kubenswrapper[4824]: I1211 02:22:04.465314 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mj9hw\" (UniqueName: \"kubernetes.io/projected/89133b02-fcf7-4fe9-b511-c406d1b0d08b-kube-api-access-mj9hw\") pod \"dnsmasq-dns-55f844cf75-6wc4l\" (UID: \"89133b02-fcf7-4fe9-b511-c406d1b0d08b\") " pod="openstack/dnsmasq-dns-55f844cf75-6wc4l" Dec 11 02:22:04 crc kubenswrapper[4824]: I1211 02:22:04.465335 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89133b02-fcf7-4fe9-b511-c406d1b0d08b-config\") pod \"dnsmasq-dns-55f844cf75-6wc4l\" (UID: \"89133b02-fcf7-4fe9-b511-c406d1b0d08b\") " pod="openstack/dnsmasq-dns-55f844cf75-6wc4l" Dec 11 02:22:04 crc kubenswrapper[4824]: I1211 02:22:04.465354 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/89133b02-fcf7-4fe9-b511-c406d1b0d08b-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-6wc4l\" (UID: \"89133b02-fcf7-4fe9-b511-c406d1b0d08b\") " pod="openstack/dnsmasq-dns-55f844cf75-6wc4l" Dec 11 02:22:04 crc kubenswrapper[4824]: I1211 02:22:04.465387 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/89133b02-fcf7-4fe9-b511-c406d1b0d08b-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-6wc4l\" (UID: \"89133b02-fcf7-4fe9-b511-c406d1b0d08b\") " pod="openstack/dnsmasq-dns-55f844cf75-6wc4l" Dec 11 02:22:04 crc kubenswrapper[4824]: I1211 02:22:04.465408 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/89133b02-fcf7-4fe9-b511-c406d1b0d08b-dns-svc\") pod \"dnsmasq-dns-55f844cf75-6wc4l\" (UID: \"89133b02-fcf7-4fe9-b511-c406d1b0d08b\") " pod="openstack/dnsmasq-dns-55f844cf75-6wc4l" Dec 11 02:22:04 crc kubenswrapper[4824]: I1211 02:22:04.491442 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-5fbc99977b-tggml"] Dec 11 02:22:04 crc kubenswrapper[4824]: I1211 02:22:04.492827 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5fbc99977b-tggml" Dec 11 02:22:04 crc kubenswrapper[4824]: I1211 02:22:04.500020 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 11 02:22:04 crc kubenswrapper[4824]: I1211 02:22:04.502134 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-b7xg9" Dec 11 02:22:04 crc kubenswrapper[4824]: I1211 02:22:04.502226 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Dec 11 02:22:04 crc kubenswrapper[4824]: I1211 02:22:04.502268 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 11 02:22:04 crc kubenswrapper[4824]: I1211 02:22:04.525742 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5fbc99977b-tggml"] Dec 11 02:22:04 crc kubenswrapper[4824]: I1211 02:22:04.567139 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/89133b02-fcf7-4fe9-b511-c406d1b0d08b-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-6wc4l\" (UID: \"89133b02-fcf7-4fe9-b511-c406d1b0d08b\") " pod="openstack/dnsmasq-dns-55f844cf75-6wc4l" Dec 11 02:22:04 crc kubenswrapper[4824]: I1211 02:22:04.567503 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mj9hw\" (UniqueName: \"kubernetes.io/projected/89133b02-fcf7-4fe9-b511-c406d1b0d08b-kube-api-access-mj9hw\") pod \"dnsmasq-dns-55f844cf75-6wc4l\" (UID: \"89133b02-fcf7-4fe9-b511-c406d1b0d08b\") " pod="openstack/dnsmasq-dns-55f844cf75-6wc4l" Dec 11 02:22:04 crc kubenswrapper[4824]: I1211 02:22:04.567531 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89133b02-fcf7-4fe9-b511-c406d1b0d08b-config\") pod \"dnsmasq-dns-55f844cf75-6wc4l\" (UID: \"89133b02-fcf7-4fe9-b511-c406d1b0d08b\") " pod="openstack/dnsmasq-dns-55f844cf75-6wc4l" Dec 11 02:22:04 crc kubenswrapper[4824]: I1211 02:22:04.567548 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/89133b02-fcf7-4fe9-b511-c406d1b0d08b-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-6wc4l\" (UID: \"89133b02-fcf7-4fe9-b511-c406d1b0d08b\") " pod="openstack/dnsmasq-dns-55f844cf75-6wc4l" Dec 11 02:22:04 crc kubenswrapper[4824]: I1211 02:22:04.567578 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/89133b02-fcf7-4fe9-b511-c406d1b0d08b-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-6wc4l\" (UID: \"89133b02-fcf7-4fe9-b511-c406d1b0d08b\") " pod="openstack/dnsmasq-dns-55f844cf75-6wc4l" Dec 11 02:22:04 crc kubenswrapper[4824]: I1211 02:22:04.567601 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/89133b02-fcf7-4fe9-b511-c406d1b0d08b-dns-svc\") pod \"dnsmasq-dns-55f844cf75-6wc4l\" (UID: \"89133b02-fcf7-4fe9-b511-c406d1b0d08b\") " pod="openstack/dnsmasq-dns-55f844cf75-6wc4l" Dec 11 02:22:04 crc kubenswrapper[4824]: I1211 02:22:04.568152 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/89133b02-fcf7-4fe9-b511-c406d1b0d08b-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-6wc4l\" (UID: \"89133b02-fcf7-4fe9-b511-c406d1b0d08b\") " pod="openstack/dnsmasq-dns-55f844cf75-6wc4l" Dec 11 02:22:04 crc kubenswrapper[4824]: I1211 02:22:04.568345 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/89133b02-fcf7-4fe9-b511-c406d1b0d08b-dns-svc\") pod \"dnsmasq-dns-55f844cf75-6wc4l\" (UID: \"89133b02-fcf7-4fe9-b511-c406d1b0d08b\") " pod="openstack/dnsmasq-dns-55f844cf75-6wc4l" Dec 11 02:22:04 crc kubenswrapper[4824]: I1211 02:22:04.568927 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/89133b02-fcf7-4fe9-b511-c406d1b0d08b-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-6wc4l\" (UID: \"89133b02-fcf7-4fe9-b511-c406d1b0d08b\") " pod="openstack/dnsmasq-dns-55f844cf75-6wc4l" Dec 11 02:22:04 crc kubenswrapper[4824]: I1211 02:22:04.568993 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89133b02-fcf7-4fe9-b511-c406d1b0d08b-config\") pod \"dnsmasq-dns-55f844cf75-6wc4l\" (UID: \"89133b02-fcf7-4fe9-b511-c406d1b0d08b\") " pod="openstack/dnsmasq-dns-55f844cf75-6wc4l" Dec 11 02:22:04 crc kubenswrapper[4824]: I1211 02:22:04.569047 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/89133b02-fcf7-4fe9-b511-c406d1b0d08b-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-6wc4l\" (UID: \"89133b02-fcf7-4fe9-b511-c406d1b0d08b\") " pod="openstack/dnsmasq-dns-55f844cf75-6wc4l" Dec 11 02:22:04 crc kubenswrapper[4824]: I1211 02:22:04.591739 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mj9hw\" (UniqueName: \"kubernetes.io/projected/89133b02-fcf7-4fe9-b511-c406d1b0d08b-kube-api-access-mj9hw\") pod \"dnsmasq-dns-55f844cf75-6wc4l\" (UID: \"89133b02-fcf7-4fe9-b511-c406d1b0d08b\") " pod="openstack/dnsmasq-dns-55f844cf75-6wc4l" Dec 11 02:22:04 crc kubenswrapper[4824]: I1211 02:22:04.669019 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/0a39253e-8c29-41ae-a242-8c9fccfe2613-httpd-config\") pod \"neutron-5fbc99977b-tggml\" (UID: \"0a39253e-8c29-41ae-a242-8c9fccfe2613\") " pod="openstack/neutron-5fbc99977b-tggml" Dec 11 02:22:04 crc kubenswrapper[4824]: I1211 02:22:04.669061 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g9bsk\" (UniqueName: \"kubernetes.io/projected/0a39253e-8c29-41ae-a242-8c9fccfe2613-kube-api-access-g9bsk\") pod \"neutron-5fbc99977b-tggml\" (UID: \"0a39253e-8c29-41ae-a242-8c9fccfe2613\") " pod="openstack/neutron-5fbc99977b-tggml" Dec 11 02:22:04 crc kubenswrapper[4824]: I1211 02:22:04.669101 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/0a39253e-8c29-41ae-a242-8c9fccfe2613-config\") pod \"neutron-5fbc99977b-tggml\" (UID: \"0a39253e-8c29-41ae-a242-8c9fccfe2613\") " pod="openstack/neutron-5fbc99977b-tggml" Dec 11 02:22:04 crc kubenswrapper[4824]: I1211 02:22:04.669169 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a39253e-8c29-41ae-a242-8c9fccfe2613-ovndb-tls-certs\") pod \"neutron-5fbc99977b-tggml\" (UID: \"0a39253e-8c29-41ae-a242-8c9fccfe2613\") " pod="openstack/neutron-5fbc99977b-tggml" Dec 11 02:22:04 crc kubenswrapper[4824]: I1211 02:22:04.669192 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a39253e-8c29-41ae-a242-8c9fccfe2613-combined-ca-bundle\") pod \"neutron-5fbc99977b-tggml\" (UID: \"0a39253e-8c29-41ae-a242-8c9fccfe2613\") " pod="openstack/neutron-5fbc99977b-tggml" Dec 11 02:22:04 crc kubenswrapper[4824]: I1211 02:22:04.732609 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-6wc4l" Dec 11 02:22:04 crc kubenswrapper[4824]: I1211 02:22:04.770133 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a39253e-8c29-41ae-a242-8c9fccfe2613-ovndb-tls-certs\") pod \"neutron-5fbc99977b-tggml\" (UID: \"0a39253e-8c29-41ae-a242-8c9fccfe2613\") " pod="openstack/neutron-5fbc99977b-tggml" Dec 11 02:22:04 crc kubenswrapper[4824]: I1211 02:22:04.770178 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a39253e-8c29-41ae-a242-8c9fccfe2613-combined-ca-bundle\") pod \"neutron-5fbc99977b-tggml\" (UID: \"0a39253e-8c29-41ae-a242-8c9fccfe2613\") " pod="openstack/neutron-5fbc99977b-tggml" Dec 11 02:22:04 crc kubenswrapper[4824]: I1211 02:22:04.770319 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/0a39253e-8c29-41ae-a242-8c9fccfe2613-httpd-config\") pod \"neutron-5fbc99977b-tggml\" (UID: \"0a39253e-8c29-41ae-a242-8c9fccfe2613\") " pod="openstack/neutron-5fbc99977b-tggml" Dec 11 02:22:04 crc kubenswrapper[4824]: I1211 02:22:04.770338 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g9bsk\" (UniqueName: \"kubernetes.io/projected/0a39253e-8c29-41ae-a242-8c9fccfe2613-kube-api-access-g9bsk\") pod \"neutron-5fbc99977b-tggml\" (UID: \"0a39253e-8c29-41ae-a242-8c9fccfe2613\") " pod="openstack/neutron-5fbc99977b-tggml" Dec 11 02:22:04 crc kubenswrapper[4824]: I1211 02:22:04.770360 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/0a39253e-8c29-41ae-a242-8c9fccfe2613-config\") pod \"neutron-5fbc99977b-tggml\" (UID: \"0a39253e-8c29-41ae-a242-8c9fccfe2613\") " pod="openstack/neutron-5fbc99977b-tggml" Dec 11 02:22:04 crc kubenswrapper[4824]: I1211 02:22:04.773835 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-lpwl8" Dec 11 02:22:04 crc kubenswrapper[4824]: I1211 02:22:04.776100 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/0a39253e-8c29-41ae-a242-8c9fccfe2613-httpd-config\") pod \"neutron-5fbc99977b-tggml\" (UID: \"0a39253e-8c29-41ae-a242-8c9fccfe2613\") " pod="openstack/neutron-5fbc99977b-tggml" Dec 11 02:22:04 crc kubenswrapper[4824]: I1211 02:22:04.778845 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a39253e-8c29-41ae-a242-8c9fccfe2613-ovndb-tls-certs\") pod \"neutron-5fbc99977b-tggml\" (UID: \"0a39253e-8c29-41ae-a242-8c9fccfe2613\") " pod="openstack/neutron-5fbc99977b-tggml" Dec 11 02:22:04 crc kubenswrapper[4824]: I1211 02:22:04.783063 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/0a39253e-8c29-41ae-a242-8c9fccfe2613-config\") pod \"neutron-5fbc99977b-tggml\" (UID: \"0a39253e-8c29-41ae-a242-8c9fccfe2613\") " pod="openstack/neutron-5fbc99977b-tggml" Dec 11 02:22:04 crc kubenswrapper[4824]: I1211 02:22:04.785515 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a39253e-8c29-41ae-a242-8c9fccfe2613-combined-ca-bundle\") pod \"neutron-5fbc99977b-tggml\" (UID: \"0a39253e-8c29-41ae-a242-8c9fccfe2613\") " pod="openstack/neutron-5fbc99977b-tggml" Dec 11 02:22:04 crc kubenswrapper[4824]: I1211 02:22:04.796794 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g9bsk\" (UniqueName: \"kubernetes.io/projected/0a39253e-8c29-41ae-a242-8c9fccfe2613-kube-api-access-g9bsk\") pod \"neutron-5fbc99977b-tggml\" (UID: \"0a39253e-8c29-41ae-a242-8c9fccfe2613\") " pod="openstack/neutron-5fbc99977b-tggml" Dec 11 02:22:04 crc kubenswrapper[4824]: I1211 02:22:04.813632 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5fbc99977b-tggml" Dec 11 02:22:04 crc kubenswrapper[4824]: I1211 02:22:04.872037 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68cc858a-9296-4a34-8322-b2dabc8ac1a2-combined-ca-bundle\") pod \"68cc858a-9296-4a34-8322-b2dabc8ac1a2\" (UID: \"68cc858a-9296-4a34-8322-b2dabc8ac1a2\") " Dec 11 02:22:04 crc kubenswrapper[4824]: I1211 02:22:04.872100 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/68cc858a-9296-4a34-8322-b2dabc8ac1a2-scripts\") pod \"68cc858a-9296-4a34-8322-b2dabc8ac1a2\" (UID: \"68cc858a-9296-4a34-8322-b2dabc8ac1a2\") " Dec 11 02:22:04 crc kubenswrapper[4824]: I1211 02:22:04.872267 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68cc858a-9296-4a34-8322-b2dabc8ac1a2-config-data\") pod \"68cc858a-9296-4a34-8322-b2dabc8ac1a2\" (UID: \"68cc858a-9296-4a34-8322-b2dabc8ac1a2\") " Dec 11 02:22:04 crc kubenswrapper[4824]: I1211 02:22:04.872302 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n22sp\" (UniqueName: \"kubernetes.io/projected/68cc858a-9296-4a34-8322-b2dabc8ac1a2-kube-api-access-n22sp\") pod \"68cc858a-9296-4a34-8322-b2dabc8ac1a2\" (UID: \"68cc858a-9296-4a34-8322-b2dabc8ac1a2\") " Dec 11 02:22:04 crc kubenswrapper[4824]: I1211 02:22:04.872546 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/68cc858a-9296-4a34-8322-b2dabc8ac1a2-logs\") pod \"68cc858a-9296-4a34-8322-b2dabc8ac1a2\" (UID: \"68cc858a-9296-4a34-8322-b2dabc8ac1a2\") " Dec 11 02:22:04 crc kubenswrapper[4824]: I1211 02:22:04.876493 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68cc858a-9296-4a34-8322-b2dabc8ac1a2-scripts" (OuterVolumeSpecName: "scripts") pod "68cc858a-9296-4a34-8322-b2dabc8ac1a2" (UID: "68cc858a-9296-4a34-8322-b2dabc8ac1a2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:22:04 crc kubenswrapper[4824]: I1211 02:22:04.876604 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68cc858a-9296-4a34-8322-b2dabc8ac1a2-kube-api-access-n22sp" (OuterVolumeSpecName: "kube-api-access-n22sp") pod "68cc858a-9296-4a34-8322-b2dabc8ac1a2" (UID: "68cc858a-9296-4a34-8322-b2dabc8ac1a2"). InnerVolumeSpecName "kube-api-access-n22sp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:22:04 crc kubenswrapper[4824]: I1211 02:22:04.876614 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/68cc858a-9296-4a34-8322-b2dabc8ac1a2-logs" (OuterVolumeSpecName: "logs") pod "68cc858a-9296-4a34-8322-b2dabc8ac1a2" (UID: "68cc858a-9296-4a34-8322-b2dabc8ac1a2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:22:04 crc kubenswrapper[4824]: I1211 02:22:04.921370 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68cc858a-9296-4a34-8322-b2dabc8ac1a2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "68cc858a-9296-4a34-8322-b2dabc8ac1a2" (UID: "68cc858a-9296-4a34-8322-b2dabc8ac1a2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:22:04 crc kubenswrapper[4824]: I1211 02:22:04.957433 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68cc858a-9296-4a34-8322-b2dabc8ac1a2-config-data" (OuterVolumeSpecName: "config-data") pod "68cc858a-9296-4a34-8322-b2dabc8ac1a2" (UID: "68cc858a-9296-4a34-8322-b2dabc8ac1a2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:22:04 crc kubenswrapper[4824]: I1211 02:22:04.975153 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68cc858a-9296-4a34-8322-b2dabc8ac1a2-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:04 crc kubenswrapper[4824]: I1211 02:22:04.975184 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n22sp\" (UniqueName: \"kubernetes.io/projected/68cc858a-9296-4a34-8322-b2dabc8ac1a2-kube-api-access-n22sp\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:04 crc kubenswrapper[4824]: I1211 02:22:04.975194 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/68cc858a-9296-4a34-8322-b2dabc8ac1a2-logs\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:04 crc kubenswrapper[4824]: I1211 02:22:04.975203 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68cc858a-9296-4a34-8322-b2dabc8ac1a2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:04 crc kubenswrapper[4824]: I1211 02:22:04.975214 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/68cc858a-9296-4a34-8322-b2dabc8ac1a2-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.020030 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.179206 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/189f7197-9da3-402f-9c9b-042e34ca6556-logs\") pod \"189f7197-9da3-402f-9c9b-042e34ca6556\" (UID: \"189f7197-9da3-402f-9c9b-042e34ca6556\") " Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.179571 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/189f7197-9da3-402f-9c9b-042e34ca6556-scripts\") pod \"189f7197-9da3-402f-9c9b-042e34ca6556\" (UID: \"189f7197-9da3-402f-9c9b-042e34ca6556\") " Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.179591 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/189f7197-9da3-402f-9c9b-042e34ca6556-config-data\") pod \"189f7197-9da3-402f-9c9b-042e34ca6556\" (UID: \"189f7197-9da3-402f-9c9b-042e34ca6556\") " Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.179630 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/189f7197-9da3-402f-9c9b-042e34ca6556-httpd-run\") pod \"189f7197-9da3-402f-9c9b-042e34ca6556\" (UID: \"189f7197-9da3-402f-9c9b-042e34ca6556\") " Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.179697 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/189f7197-9da3-402f-9c9b-042e34ca6556-combined-ca-bundle\") pod \"189f7197-9da3-402f-9c9b-042e34ca6556\" (UID: \"189f7197-9da3-402f-9c9b-042e34ca6556\") " Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.179718 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"189f7197-9da3-402f-9c9b-042e34ca6556\" (UID: \"189f7197-9da3-402f-9c9b-042e34ca6556\") " Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.179803 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/189f7197-9da3-402f-9c9b-042e34ca6556-internal-tls-certs\") pod \"189f7197-9da3-402f-9c9b-042e34ca6556\" (UID: \"189f7197-9da3-402f-9c9b-042e34ca6556\") " Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.179903 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tnnmb\" (UniqueName: \"kubernetes.io/projected/189f7197-9da3-402f-9c9b-042e34ca6556-kube-api-access-tnnmb\") pod \"189f7197-9da3-402f-9c9b-042e34ca6556\" (UID: \"189f7197-9da3-402f-9c9b-042e34ca6556\") " Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.180604 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/189f7197-9da3-402f-9c9b-042e34ca6556-logs" (OuterVolumeSpecName: "logs") pod "189f7197-9da3-402f-9c9b-042e34ca6556" (UID: "189f7197-9da3-402f-9c9b-042e34ca6556"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.181345 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/189f7197-9da3-402f-9c9b-042e34ca6556-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "189f7197-9da3-402f-9c9b-042e34ca6556" (UID: "189f7197-9da3-402f-9c9b-042e34ca6556"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.192975 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/189f7197-9da3-402f-9c9b-042e34ca6556-scripts" (OuterVolumeSpecName: "scripts") pod "189f7197-9da3-402f-9c9b-042e34ca6556" (UID: "189f7197-9da3-402f-9c9b-042e34ca6556"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.193048 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/189f7197-9da3-402f-9c9b-042e34ca6556-kube-api-access-tnnmb" (OuterVolumeSpecName: "kube-api-access-tnnmb") pod "189f7197-9da3-402f-9c9b-042e34ca6556" (UID: "189f7197-9da3-402f-9c9b-042e34ca6556"). InnerVolumeSpecName "kube-api-access-tnnmb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.195744 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "189f7197-9da3-402f-9c9b-042e34ca6556" (UID: "189f7197-9da3-402f-9c9b-042e34ca6556"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.243137 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-78b68944cb-4p8xq"] Dec 11 02:22:05 crc kubenswrapper[4824]: E1211 02:22:05.243476 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="189f7197-9da3-402f-9c9b-042e34ca6556" containerName="glance-log" Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.243487 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="189f7197-9da3-402f-9c9b-042e34ca6556" containerName="glance-log" Dec 11 02:22:05 crc kubenswrapper[4824]: E1211 02:22:05.243498 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="189f7197-9da3-402f-9c9b-042e34ca6556" containerName="glance-httpd" Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.243504 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="189f7197-9da3-402f-9c9b-042e34ca6556" containerName="glance-httpd" Dec 11 02:22:05 crc kubenswrapper[4824]: E1211 02:22:05.243518 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68cc858a-9296-4a34-8322-b2dabc8ac1a2" containerName="placement-db-sync" Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.243524 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="68cc858a-9296-4a34-8322-b2dabc8ac1a2" containerName="placement-db-sync" Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.243686 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="189f7197-9da3-402f-9c9b-042e34ca6556" containerName="glance-log" Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.243699 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="189f7197-9da3-402f-9c9b-042e34ca6556" containerName="glance-httpd" Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.243713 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="68cc858a-9296-4a34-8322-b2dabc8ac1a2" containerName="placement-db-sync" Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.244528 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-78b68944cb-4p8xq" Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.245251 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/189f7197-9da3-402f-9c9b-042e34ca6556-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "189f7197-9da3-402f-9c9b-042e34ca6556" (UID: "189f7197-9da3-402f-9c9b-042e34ca6556"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.249810 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.250381 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.276916 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-78b68944cb-4p8xq"] Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.278961 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-lpwl8" event={"ID":"68cc858a-9296-4a34-8322-b2dabc8ac1a2","Type":"ContainerDied","Data":"06c8ade0aaed7f6fe86d287e1a891c7a424aa4e4c2af950bc4f08db03eb1c99d"} Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.278992 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="06c8ade0aaed7f6fe86d287e1a891c7a424aa4e4c2af950bc4f08db03eb1c99d" Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.279045 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-lpwl8" Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.282086 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/189f7197-9da3-402f-9c9b-042e34ca6556-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.282151 4824 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.282163 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tnnmb\" (UniqueName: \"kubernetes.io/projected/189f7197-9da3-402f-9c9b-042e34ca6556-kube-api-access-tnnmb\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.282172 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/189f7197-9da3-402f-9c9b-042e34ca6556-logs\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.282181 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/189f7197-9da3-402f-9c9b-042e34ca6556-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.282240 4824 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/189f7197-9da3-402f-9c9b-042e34ca6556-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.301099 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"189f7197-9da3-402f-9c9b-042e34ca6556","Type":"ContainerDied","Data":"c0fc742b51065576f878140173b3938b377a748bc19f9230d9807d760bfdf2b1"} Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.301164 4824 scope.go:117] "RemoveContainer" containerID="38b57fc2e66e0271d14b42c29f82423a12edb93639980b09631c27764378933f" Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.301299 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.306350 4824 generic.go:334] "Generic (PLEG): container finished" podID="dfa654c8-1f3b-43f6-8a1f-709a05020b4e" containerID="ed47e0f4bf032969946a71cb7a3c67c26e5ce39ba4f3607e54fa38b0c5b34968" exitCode=0 Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.306394 4824 generic.go:334] "Generic (PLEG): container finished" podID="dfa654c8-1f3b-43f6-8a1f-709a05020b4e" containerID="62aff7415eaffd64df1b48803198a5006854ea9c0d791bee37ad995fe18a6352" exitCode=143 Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.307228 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"dfa654c8-1f3b-43f6-8a1f-709a05020b4e","Type":"ContainerDied","Data":"ed47e0f4bf032969946a71cb7a3c67c26e5ce39ba4f3607e54fa38b0c5b34968"} Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.307253 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"dfa654c8-1f3b-43f6-8a1f-709a05020b4e","Type":"ContainerDied","Data":"62aff7415eaffd64df1b48803198a5006854ea9c0d791bee37ad995fe18a6352"} Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.320014 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/189f7197-9da3-402f-9c9b-042e34ca6556-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "189f7197-9da3-402f-9c9b-042e34ca6556" (UID: "189f7197-9da3-402f-9c9b-042e34ca6556"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.328977 4824 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.336934 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/189f7197-9da3-402f-9c9b-042e34ca6556-config-data" (OuterVolumeSpecName: "config-data") pod "189f7197-9da3-402f-9c9b-042e34ca6556" (UID: "189f7197-9da3-402f-9c9b-042e34ca6556"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.354245 4824 scope.go:117] "RemoveContainer" containerID="e2fe9aca09a3e9a571c71da03de60b9dd22f41ed17659243623d520c5e2090de" Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.387696 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4663f55-6344-4154-9df6-96a17ca59bdb-combined-ca-bundle\") pod \"placement-78b68944cb-4p8xq\" (UID: \"d4663f55-6344-4154-9df6-96a17ca59bdb\") " pod="openstack/placement-78b68944cb-4p8xq" Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.387748 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d4663f55-6344-4154-9df6-96a17ca59bdb-logs\") pod \"placement-78b68944cb-4p8xq\" (UID: \"d4663f55-6344-4154-9df6-96a17ca59bdb\") " pod="openstack/placement-78b68944cb-4p8xq" Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.387768 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4663f55-6344-4154-9df6-96a17ca59bdb-internal-tls-certs\") pod \"placement-78b68944cb-4p8xq\" (UID: \"d4663f55-6344-4154-9df6-96a17ca59bdb\") " pod="openstack/placement-78b68944cb-4p8xq" Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.387818 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4663f55-6344-4154-9df6-96a17ca59bdb-config-data\") pod \"placement-78b68944cb-4p8xq\" (UID: \"d4663f55-6344-4154-9df6-96a17ca59bdb\") " pod="openstack/placement-78b68944cb-4p8xq" Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.387846 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4663f55-6344-4154-9df6-96a17ca59bdb-public-tls-certs\") pod \"placement-78b68944cb-4p8xq\" (UID: \"d4663f55-6344-4154-9df6-96a17ca59bdb\") " pod="openstack/placement-78b68944cb-4p8xq" Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.387893 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4663f55-6344-4154-9df6-96a17ca59bdb-scripts\") pod \"placement-78b68944cb-4p8xq\" (UID: \"d4663f55-6344-4154-9df6-96a17ca59bdb\") " pod="openstack/placement-78b68944cb-4p8xq" Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.387921 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vtdcr\" (UniqueName: \"kubernetes.io/projected/d4663f55-6344-4154-9df6-96a17ca59bdb-kube-api-access-vtdcr\") pod \"placement-78b68944cb-4p8xq\" (UID: \"d4663f55-6344-4154-9df6-96a17ca59bdb\") " pod="openstack/placement-78b68944cb-4p8xq" Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.387958 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/189f7197-9da3-402f-9c9b-042e34ca6556-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.387976 4824 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.387984 4824 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/189f7197-9da3-402f-9c9b-042e34ca6556-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.472901 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5fbc99977b-tggml"] Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.488992 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4663f55-6344-4154-9df6-96a17ca59bdb-config-data\") pod \"placement-78b68944cb-4p8xq\" (UID: \"d4663f55-6344-4154-9df6-96a17ca59bdb\") " pod="openstack/placement-78b68944cb-4p8xq" Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.489067 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4663f55-6344-4154-9df6-96a17ca59bdb-public-tls-certs\") pod \"placement-78b68944cb-4p8xq\" (UID: \"d4663f55-6344-4154-9df6-96a17ca59bdb\") " pod="openstack/placement-78b68944cb-4p8xq" Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.489142 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4663f55-6344-4154-9df6-96a17ca59bdb-scripts\") pod \"placement-78b68944cb-4p8xq\" (UID: \"d4663f55-6344-4154-9df6-96a17ca59bdb\") " pod="openstack/placement-78b68944cb-4p8xq" Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.489187 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vtdcr\" (UniqueName: \"kubernetes.io/projected/d4663f55-6344-4154-9df6-96a17ca59bdb-kube-api-access-vtdcr\") pod \"placement-78b68944cb-4p8xq\" (UID: \"d4663f55-6344-4154-9df6-96a17ca59bdb\") " pod="openstack/placement-78b68944cb-4p8xq" Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.489231 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4663f55-6344-4154-9df6-96a17ca59bdb-combined-ca-bundle\") pod \"placement-78b68944cb-4p8xq\" (UID: \"d4663f55-6344-4154-9df6-96a17ca59bdb\") " pod="openstack/placement-78b68944cb-4p8xq" Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.489269 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d4663f55-6344-4154-9df6-96a17ca59bdb-logs\") pod \"placement-78b68944cb-4p8xq\" (UID: \"d4663f55-6344-4154-9df6-96a17ca59bdb\") " pod="openstack/placement-78b68944cb-4p8xq" Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.489291 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4663f55-6344-4154-9df6-96a17ca59bdb-internal-tls-certs\") pod \"placement-78b68944cb-4p8xq\" (UID: \"d4663f55-6344-4154-9df6-96a17ca59bdb\") " pod="openstack/placement-78b68944cb-4p8xq" Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.506748 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4663f55-6344-4154-9df6-96a17ca59bdb-public-tls-certs\") pod \"placement-78b68944cb-4p8xq\" (UID: \"d4663f55-6344-4154-9df6-96a17ca59bdb\") " pod="openstack/placement-78b68944cb-4p8xq" Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.511228 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4663f55-6344-4154-9df6-96a17ca59bdb-scripts\") pod \"placement-78b68944cb-4p8xq\" (UID: \"d4663f55-6344-4154-9df6-96a17ca59bdb\") " pod="openstack/placement-78b68944cb-4p8xq" Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.512037 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d4663f55-6344-4154-9df6-96a17ca59bdb-logs\") pod \"placement-78b68944cb-4p8xq\" (UID: \"d4663f55-6344-4154-9df6-96a17ca59bdb\") " pod="openstack/placement-78b68944cb-4p8xq" Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.523870 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4663f55-6344-4154-9df6-96a17ca59bdb-config-data\") pod \"placement-78b68944cb-4p8xq\" (UID: \"d4663f55-6344-4154-9df6-96a17ca59bdb\") " pod="openstack/placement-78b68944cb-4p8xq" Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.533894 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4663f55-6344-4154-9df6-96a17ca59bdb-internal-tls-certs\") pod \"placement-78b68944cb-4p8xq\" (UID: \"d4663f55-6344-4154-9df6-96a17ca59bdb\") " pod="openstack/placement-78b68944cb-4p8xq" Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.537889 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4663f55-6344-4154-9df6-96a17ca59bdb-combined-ca-bundle\") pod \"placement-78b68944cb-4p8xq\" (UID: \"d4663f55-6344-4154-9df6-96a17ca59bdb\") " pod="openstack/placement-78b68944cb-4p8xq" Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.538911 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vtdcr\" (UniqueName: \"kubernetes.io/projected/d4663f55-6344-4154-9df6-96a17ca59bdb-kube-api-access-vtdcr\") pod \"placement-78b68944cb-4p8xq\" (UID: \"d4663f55-6344-4154-9df6-96a17ca59bdb\") " pod="openstack/placement-78b68944cb-4p8xq" Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.655936 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-6wc4l"] Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.678025 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-78b68944cb-4p8xq" Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.746125 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.765205 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.791202 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.792681 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.806087 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.806524 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.810628 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.925703 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e7794eea-3d90-4bd6-8d69-732d80385243-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"e7794eea-3d90-4bd6-8d69-732d80385243\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.926313 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7794eea-3d90-4bd6-8d69-732d80385243-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"e7794eea-3d90-4bd6-8d69-732d80385243\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.926353 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7794eea-3d90-4bd6-8d69-732d80385243-config-data\") pod \"glance-default-internal-api-0\" (UID: \"e7794eea-3d90-4bd6-8d69-732d80385243\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.926523 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"e7794eea-3d90-4bd6-8d69-732d80385243\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.926583 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-65qz9\" (UniqueName: \"kubernetes.io/projected/e7794eea-3d90-4bd6-8d69-732d80385243-kube-api-access-65qz9\") pod \"glance-default-internal-api-0\" (UID: \"e7794eea-3d90-4bd6-8d69-732d80385243\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.926612 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e7794eea-3d90-4bd6-8d69-732d80385243-scripts\") pod \"glance-default-internal-api-0\" (UID: \"e7794eea-3d90-4bd6-8d69-732d80385243\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.926642 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e7794eea-3d90-4bd6-8d69-732d80385243-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"e7794eea-3d90-4bd6-8d69-732d80385243\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:22:05 crc kubenswrapper[4824]: I1211 02:22:05.926671 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e7794eea-3d90-4bd6-8d69-732d80385243-logs\") pod \"glance-default-internal-api-0\" (UID: \"e7794eea-3d90-4bd6-8d69-732d80385243\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:22:06 crc kubenswrapper[4824]: I1211 02:22:06.029164 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-65qz9\" (UniqueName: \"kubernetes.io/projected/e7794eea-3d90-4bd6-8d69-732d80385243-kube-api-access-65qz9\") pod \"glance-default-internal-api-0\" (UID: \"e7794eea-3d90-4bd6-8d69-732d80385243\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:22:06 crc kubenswrapper[4824]: I1211 02:22:06.029216 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e7794eea-3d90-4bd6-8d69-732d80385243-scripts\") pod \"glance-default-internal-api-0\" (UID: \"e7794eea-3d90-4bd6-8d69-732d80385243\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:22:06 crc kubenswrapper[4824]: I1211 02:22:06.029245 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e7794eea-3d90-4bd6-8d69-732d80385243-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"e7794eea-3d90-4bd6-8d69-732d80385243\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:22:06 crc kubenswrapper[4824]: I1211 02:22:06.029266 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e7794eea-3d90-4bd6-8d69-732d80385243-logs\") pod \"glance-default-internal-api-0\" (UID: \"e7794eea-3d90-4bd6-8d69-732d80385243\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:22:06 crc kubenswrapper[4824]: I1211 02:22:06.029302 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e7794eea-3d90-4bd6-8d69-732d80385243-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"e7794eea-3d90-4bd6-8d69-732d80385243\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:22:06 crc kubenswrapper[4824]: I1211 02:22:06.029319 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7794eea-3d90-4bd6-8d69-732d80385243-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"e7794eea-3d90-4bd6-8d69-732d80385243\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:22:06 crc kubenswrapper[4824]: I1211 02:22:06.029341 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7794eea-3d90-4bd6-8d69-732d80385243-config-data\") pod \"glance-default-internal-api-0\" (UID: \"e7794eea-3d90-4bd6-8d69-732d80385243\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:22:06 crc kubenswrapper[4824]: I1211 02:22:06.029431 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"e7794eea-3d90-4bd6-8d69-732d80385243\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:22:06 crc kubenswrapper[4824]: I1211 02:22:06.029896 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e7794eea-3d90-4bd6-8d69-732d80385243-logs\") pod \"glance-default-internal-api-0\" (UID: \"e7794eea-3d90-4bd6-8d69-732d80385243\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:22:06 crc kubenswrapper[4824]: I1211 02:22:06.030613 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"e7794eea-3d90-4bd6-8d69-732d80385243\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-internal-api-0" Dec 11 02:22:06 crc kubenswrapper[4824]: I1211 02:22:06.032475 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e7794eea-3d90-4bd6-8d69-732d80385243-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"e7794eea-3d90-4bd6-8d69-732d80385243\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:22:06 crc kubenswrapper[4824]: I1211 02:22:06.040544 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7794eea-3d90-4bd6-8d69-732d80385243-config-data\") pod \"glance-default-internal-api-0\" (UID: \"e7794eea-3d90-4bd6-8d69-732d80385243\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:22:06 crc kubenswrapper[4824]: I1211 02:22:06.051966 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e7794eea-3d90-4bd6-8d69-732d80385243-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"e7794eea-3d90-4bd6-8d69-732d80385243\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:22:06 crc kubenswrapper[4824]: I1211 02:22:06.056563 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7794eea-3d90-4bd6-8d69-732d80385243-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"e7794eea-3d90-4bd6-8d69-732d80385243\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:22:06 crc kubenswrapper[4824]: I1211 02:22:06.069791 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-65qz9\" (UniqueName: \"kubernetes.io/projected/e7794eea-3d90-4bd6-8d69-732d80385243-kube-api-access-65qz9\") pod \"glance-default-internal-api-0\" (UID: \"e7794eea-3d90-4bd6-8d69-732d80385243\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:22:06 crc kubenswrapper[4824]: I1211 02:22:06.077942 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e7794eea-3d90-4bd6-8d69-732d80385243-scripts\") pod \"glance-default-internal-api-0\" (UID: \"e7794eea-3d90-4bd6-8d69-732d80385243\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:22:06 crc kubenswrapper[4824]: I1211 02:22:06.116373 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"e7794eea-3d90-4bd6-8d69-732d80385243\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:22:06 crc kubenswrapper[4824]: I1211 02:22:06.131442 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 11 02:22:06 crc kubenswrapper[4824]: I1211 02:22:06.241456 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-78b68944cb-4p8xq"] Dec 11 02:22:06 crc kubenswrapper[4824]: I1211 02:22:06.341357 4824 generic.go:334] "Generic (PLEG): container finished" podID="89133b02-fcf7-4fe9-b511-c406d1b0d08b" containerID="aaaa7aefcd04234f29f5d526421d68dd1abc945578a44da6db1f41f30c0fae32" exitCode=0 Dec 11 02:22:06 crc kubenswrapper[4824]: I1211 02:22:06.341424 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-6wc4l" event={"ID":"89133b02-fcf7-4fe9-b511-c406d1b0d08b","Type":"ContainerDied","Data":"aaaa7aefcd04234f29f5d526421d68dd1abc945578a44da6db1f41f30c0fae32"} Dec 11 02:22:06 crc kubenswrapper[4824]: I1211 02:22:06.341449 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-6wc4l" event={"ID":"89133b02-fcf7-4fe9-b511-c406d1b0d08b","Type":"ContainerStarted","Data":"b1196e415d722aae4e9d7d26febd55839579c84917b75a3911438ccc3b6ed0d2"} Dec 11 02:22:06 crc kubenswrapper[4824]: I1211 02:22:06.373186 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5fbc99977b-tggml" event={"ID":"0a39253e-8c29-41ae-a242-8c9fccfe2613","Type":"ContainerStarted","Data":"4b69e28249b7383bb629b55abdf4c31bcabee14ef7807ca36836547df2950751"} Dec 11 02:22:06 crc kubenswrapper[4824]: I1211 02:22:06.373226 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5fbc99977b-tggml" event={"ID":"0a39253e-8c29-41ae-a242-8c9fccfe2613","Type":"ContainerStarted","Data":"d9f58bff9914c49685dacc672e47422ad4e80328ce63f58355ad585e4c25788d"} Dec 11 02:22:06 crc kubenswrapper[4824]: I1211 02:22:06.650776 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="189f7197-9da3-402f-9c9b-042e34ca6556" path="/var/lib/kubelet/pods/189f7197-9da3-402f-9c9b-042e34ca6556/volumes" Dec 11 02:22:06 crc kubenswrapper[4824]: I1211 02:22:06.694015 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-5f68f577dc-nj2zp" Dec 11 02:22:07 crc kubenswrapper[4824]: I1211 02:22:07.388532 4824 generic.go:334] "Generic (PLEG): container finished" podID="68990bd4-eecf-4181-a76c-bf0d24154774" containerID="61c05165b5969b15e231ae24f60dcbd4ca978cb8d81f0055fe9a2ab1ba1519d5" exitCode=0 Dec 11 02:22:07 crc kubenswrapper[4824]: I1211 02:22:07.388574 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-752rf" event={"ID":"68990bd4-eecf-4181-a76c-bf0d24154774","Type":"ContainerDied","Data":"61c05165b5969b15e231ae24f60dcbd4ca978cb8d81f0055fe9a2ab1ba1519d5"} Dec 11 02:22:07 crc kubenswrapper[4824]: I1211 02:22:07.755987 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-5b766ddd67-lflrw"] Dec 11 02:22:07 crc kubenswrapper[4824]: I1211 02:22:07.757288 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5b766ddd67-lflrw" Dec 11 02:22:07 crc kubenswrapper[4824]: I1211 02:22:07.760959 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Dec 11 02:22:07 crc kubenswrapper[4824]: I1211 02:22:07.761095 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Dec 11 02:22:07 crc kubenswrapper[4824]: I1211 02:22:07.784911 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5b766ddd67-lflrw"] Dec 11 02:22:07 crc kubenswrapper[4824]: I1211 02:22:07.870295 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/cde933a7-1f3a-4567-b6fd-7644a02cb745-ovndb-tls-certs\") pod \"neutron-5b766ddd67-lflrw\" (UID: \"cde933a7-1f3a-4567-b6fd-7644a02cb745\") " pod="openstack/neutron-5b766ddd67-lflrw" Dec 11 02:22:07 crc kubenswrapper[4824]: I1211 02:22:07.870358 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4bcs8\" (UniqueName: \"kubernetes.io/projected/cde933a7-1f3a-4567-b6fd-7644a02cb745-kube-api-access-4bcs8\") pod \"neutron-5b766ddd67-lflrw\" (UID: \"cde933a7-1f3a-4567-b6fd-7644a02cb745\") " pod="openstack/neutron-5b766ddd67-lflrw" Dec 11 02:22:07 crc kubenswrapper[4824]: I1211 02:22:07.870385 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/cde933a7-1f3a-4567-b6fd-7644a02cb745-httpd-config\") pod \"neutron-5b766ddd67-lflrw\" (UID: \"cde933a7-1f3a-4567-b6fd-7644a02cb745\") " pod="openstack/neutron-5b766ddd67-lflrw" Dec 11 02:22:07 crc kubenswrapper[4824]: I1211 02:22:07.870586 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/cde933a7-1f3a-4567-b6fd-7644a02cb745-config\") pod \"neutron-5b766ddd67-lflrw\" (UID: \"cde933a7-1f3a-4567-b6fd-7644a02cb745\") " pod="openstack/neutron-5b766ddd67-lflrw" Dec 11 02:22:07 crc kubenswrapper[4824]: I1211 02:22:07.870739 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cde933a7-1f3a-4567-b6fd-7644a02cb745-internal-tls-certs\") pod \"neutron-5b766ddd67-lflrw\" (UID: \"cde933a7-1f3a-4567-b6fd-7644a02cb745\") " pod="openstack/neutron-5b766ddd67-lflrw" Dec 11 02:22:07 crc kubenswrapper[4824]: I1211 02:22:07.870978 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cde933a7-1f3a-4567-b6fd-7644a02cb745-public-tls-certs\") pod \"neutron-5b766ddd67-lflrw\" (UID: \"cde933a7-1f3a-4567-b6fd-7644a02cb745\") " pod="openstack/neutron-5b766ddd67-lflrw" Dec 11 02:22:07 crc kubenswrapper[4824]: I1211 02:22:07.871086 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cde933a7-1f3a-4567-b6fd-7644a02cb745-combined-ca-bundle\") pod \"neutron-5b766ddd67-lflrw\" (UID: \"cde933a7-1f3a-4567-b6fd-7644a02cb745\") " pod="openstack/neutron-5b766ddd67-lflrw" Dec 11 02:22:07 crc kubenswrapper[4824]: I1211 02:22:07.977160 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cde933a7-1f3a-4567-b6fd-7644a02cb745-public-tls-certs\") pod \"neutron-5b766ddd67-lflrw\" (UID: \"cde933a7-1f3a-4567-b6fd-7644a02cb745\") " pod="openstack/neutron-5b766ddd67-lflrw" Dec 11 02:22:07 crc kubenswrapper[4824]: I1211 02:22:07.977261 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cde933a7-1f3a-4567-b6fd-7644a02cb745-combined-ca-bundle\") pod \"neutron-5b766ddd67-lflrw\" (UID: \"cde933a7-1f3a-4567-b6fd-7644a02cb745\") " pod="openstack/neutron-5b766ddd67-lflrw" Dec 11 02:22:07 crc kubenswrapper[4824]: I1211 02:22:07.977420 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/cde933a7-1f3a-4567-b6fd-7644a02cb745-ovndb-tls-certs\") pod \"neutron-5b766ddd67-lflrw\" (UID: \"cde933a7-1f3a-4567-b6fd-7644a02cb745\") " pod="openstack/neutron-5b766ddd67-lflrw" Dec 11 02:22:07 crc kubenswrapper[4824]: I1211 02:22:07.977474 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4bcs8\" (UniqueName: \"kubernetes.io/projected/cde933a7-1f3a-4567-b6fd-7644a02cb745-kube-api-access-4bcs8\") pod \"neutron-5b766ddd67-lflrw\" (UID: \"cde933a7-1f3a-4567-b6fd-7644a02cb745\") " pod="openstack/neutron-5b766ddd67-lflrw" Dec 11 02:22:07 crc kubenswrapper[4824]: I1211 02:22:07.977499 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/cde933a7-1f3a-4567-b6fd-7644a02cb745-httpd-config\") pod \"neutron-5b766ddd67-lflrw\" (UID: \"cde933a7-1f3a-4567-b6fd-7644a02cb745\") " pod="openstack/neutron-5b766ddd67-lflrw" Dec 11 02:22:07 crc kubenswrapper[4824]: I1211 02:22:07.977561 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/cde933a7-1f3a-4567-b6fd-7644a02cb745-config\") pod \"neutron-5b766ddd67-lflrw\" (UID: \"cde933a7-1f3a-4567-b6fd-7644a02cb745\") " pod="openstack/neutron-5b766ddd67-lflrw" Dec 11 02:22:07 crc kubenswrapper[4824]: I1211 02:22:07.977594 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cde933a7-1f3a-4567-b6fd-7644a02cb745-internal-tls-certs\") pod \"neutron-5b766ddd67-lflrw\" (UID: \"cde933a7-1f3a-4567-b6fd-7644a02cb745\") " pod="openstack/neutron-5b766ddd67-lflrw" Dec 11 02:22:08 crc kubenswrapper[4824]: I1211 02:22:08.003729 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cde933a7-1f3a-4567-b6fd-7644a02cb745-combined-ca-bundle\") pod \"neutron-5b766ddd67-lflrw\" (UID: \"cde933a7-1f3a-4567-b6fd-7644a02cb745\") " pod="openstack/neutron-5b766ddd67-lflrw" Dec 11 02:22:08 crc kubenswrapper[4824]: I1211 02:22:08.004528 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/cde933a7-1f3a-4567-b6fd-7644a02cb745-config\") pod \"neutron-5b766ddd67-lflrw\" (UID: \"cde933a7-1f3a-4567-b6fd-7644a02cb745\") " pod="openstack/neutron-5b766ddd67-lflrw" Dec 11 02:22:08 crc kubenswrapper[4824]: I1211 02:22:08.010631 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cde933a7-1f3a-4567-b6fd-7644a02cb745-public-tls-certs\") pod \"neutron-5b766ddd67-lflrw\" (UID: \"cde933a7-1f3a-4567-b6fd-7644a02cb745\") " pod="openstack/neutron-5b766ddd67-lflrw" Dec 11 02:22:08 crc kubenswrapper[4824]: I1211 02:22:08.010778 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/cde933a7-1f3a-4567-b6fd-7644a02cb745-ovndb-tls-certs\") pod \"neutron-5b766ddd67-lflrw\" (UID: \"cde933a7-1f3a-4567-b6fd-7644a02cb745\") " pod="openstack/neutron-5b766ddd67-lflrw" Dec 11 02:22:08 crc kubenswrapper[4824]: I1211 02:22:08.018529 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cde933a7-1f3a-4567-b6fd-7644a02cb745-internal-tls-certs\") pod \"neutron-5b766ddd67-lflrw\" (UID: \"cde933a7-1f3a-4567-b6fd-7644a02cb745\") " pod="openstack/neutron-5b766ddd67-lflrw" Dec 11 02:22:08 crc kubenswrapper[4824]: I1211 02:22:08.031847 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4bcs8\" (UniqueName: \"kubernetes.io/projected/cde933a7-1f3a-4567-b6fd-7644a02cb745-kube-api-access-4bcs8\") pod \"neutron-5b766ddd67-lflrw\" (UID: \"cde933a7-1f3a-4567-b6fd-7644a02cb745\") " pod="openstack/neutron-5b766ddd67-lflrw" Dec 11 02:22:08 crc kubenswrapper[4824]: I1211 02:22:08.036791 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/cde933a7-1f3a-4567-b6fd-7644a02cb745-httpd-config\") pod \"neutron-5b766ddd67-lflrw\" (UID: \"cde933a7-1f3a-4567-b6fd-7644a02cb745\") " pod="openstack/neutron-5b766ddd67-lflrw" Dec 11 02:22:08 crc kubenswrapper[4824]: I1211 02:22:08.130730 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5b766ddd67-lflrw" Dec 11 02:22:10 crc kubenswrapper[4824]: W1211 02:22:10.464475 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd4663f55_6344_4154_9df6_96a17ca59bdb.slice/crio-c8439216c3f2bd1b9b9b71ad4eaa2d13b8a0dc05ddee38fb37ce982b88ab85ca WatchSource:0}: Error finding container c8439216c3f2bd1b9b9b71ad4eaa2d13b8a0dc05ddee38fb37ce982b88ab85ca: Status 404 returned error can't find the container with id c8439216c3f2bd1b9b9b71ad4eaa2d13b8a0dc05ddee38fb37ce982b88ab85ca Dec 11 02:22:10 crc kubenswrapper[4824]: I1211 02:22:10.688849 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-752rf" Dec 11 02:22:10 crc kubenswrapper[4824]: I1211 02:22:10.703527 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 11 02:22:10 crc kubenswrapper[4824]: I1211 02:22:10.840441 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfa654c8-1f3b-43f6-8a1f-709a05020b4e-combined-ca-bundle\") pod \"dfa654c8-1f3b-43f6-8a1f-709a05020b4e\" (UID: \"dfa654c8-1f3b-43f6-8a1f-709a05020b4e\") " Dec 11 02:22:10 crc kubenswrapper[4824]: I1211 02:22:10.840842 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/68990bd4-eecf-4181-a76c-bf0d24154774-credential-keys\") pod \"68990bd4-eecf-4181-a76c-bf0d24154774\" (UID: \"68990bd4-eecf-4181-a76c-bf0d24154774\") " Dec 11 02:22:10 crc kubenswrapper[4824]: I1211 02:22:10.841008 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68990bd4-eecf-4181-a76c-bf0d24154774-combined-ca-bundle\") pod \"68990bd4-eecf-4181-a76c-bf0d24154774\" (UID: \"68990bd4-eecf-4181-a76c-bf0d24154774\") " Dec 11 02:22:10 crc kubenswrapper[4824]: I1211 02:22:10.841055 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/68990bd4-eecf-4181-a76c-bf0d24154774-fernet-keys\") pod \"68990bd4-eecf-4181-a76c-bf0d24154774\" (UID: \"68990bd4-eecf-4181-a76c-bf0d24154774\") " Dec 11 02:22:10 crc kubenswrapper[4824]: I1211 02:22:10.841076 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c7skk\" (UniqueName: \"kubernetes.io/projected/68990bd4-eecf-4181-a76c-bf0d24154774-kube-api-access-c7skk\") pod \"68990bd4-eecf-4181-a76c-bf0d24154774\" (UID: \"68990bd4-eecf-4181-a76c-bf0d24154774\") " Dec 11 02:22:10 crc kubenswrapper[4824]: I1211 02:22:10.841137 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68990bd4-eecf-4181-a76c-bf0d24154774-config-data\") pod \"68990bd4-eecf-4181-a76c-bf0d24154774\" (UID: \"68990bd4-eecf-4181-a76c-bf0d24154774\") " Dec 11 02:22:10 crc kubenswrapper[4824]: I1211 02:22:10.841175 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"dfa654c8-1f3b-43f6-8a1f-709a05020b4e\" (UID: \"dfa654c8-1f3b-43f6-8a1f-709a05020b4e\") " Dec 11 02:22:10 crc kubenswrapper[4824]: I1211 02:22:10.841307 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/dfa654c8-1f3b-43f6-8a1f-709a05020b4e-httpd-run\") pod \"dfa654c8-1f3b-43f6-8a1f-709a05020b4e\" (UID: \"dfa654c8-1f3b-43f6-8a1f-709a05020b4e\") " Dec 11 02:22:10 crc kubenswrapper[4824]: I1211 02:22:10.841335 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dfa654c8-1f3b-43f6-8a1f-709a05020b4e-public-tls-certs\") pod \"dfa654c8-1f3b-43f6-8a1f-709a05020b4e\" (UID: \"dfa654c8-1f3b-43f6-8a1f-709a05020b4e\") " Dec 11 02:22:10 crc kubenswrapper[4824]: I1211 02:22:10.841362 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q6vwj\" (UniqueName: \"kubernetes.io/projected/dfa654c8-1f3b-43f6-8a1f-709a05020b4e-kube-api-access-q6vwj\") pod \"dfa654c8-1f3b-43f6-8a1f-709a05020b4e\" (UID: \"dfa654c8-1f3b-43f6-8a1f-709a05020b4e\") " Dec 11 02:22:10 crc kubenswrapper[4824]: I1211 02:22:10.841381 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dfa654c8-1f3b-43f6-8a1f-709a05020b4e-logs\") pod \"dfa654c8-1f3b-43f6-8a1f-709a05020b4e\" (UID: \"dfa654c8-1f3b-43f6-8a1f-709a05020b4e\") " Dec 11 02:22:10 crc kubenswrapper[4824]: I1211 02:22:10.841404 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dfa654c8-1f3b-43f6-8a1f-709a05020b4e-config-data\") pod \"dfa654c8-1f3b-43f6-8a1f-709a05020b4e\" (UID: \"dfa654c8-1f3b-43f6-8a1f-709a05020b4e\") " Dec 11 02:22:10 crc kubenswrapper[4824]: I1211 02:22:10.841426 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dfa654c8-1f3b-43f6-8a1f-709a05020b4e-scripts\") pod \"dfa654c8-1f3b-43f6-8a1f-709a05020b4e\" (UID: \"dfa654c8-1f3b-43f6-8a1f-709a05020b4e\") " Dec 11 02:22:10 crc kubenswrapper[4824]: I1211 02:22:10.841472 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/68990bd4-eecf-4181-a76c-bf0d24154774-scripts\") pod \"68990bd4-eecf-4181-a76c-bf0d24154774\" (UID: \"68990bd4-eecf-4181-a76c-bf0d24154774\") " Dec 11 02:22:10 crc kubenswrapper[4824]: I1211 02:22:10.842315 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dfa654c8-1f3b-43f6-8a1f-709a05020b4e-logs" (OuterVolumeSpecName: "logs") pod "dfa654c8-1f3b-43f6-8a1f-709a05020b4e" (UID: "dfa654c8-1f3b-43f6-8a1f-709a05020b4e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:22:10 crc kubenswrapper[4824]: I1211 02:22:10.843897 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dfa654c8-1f3b-43f6-8a1f-709a05020b4e-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "dfa654c8-1f3b-43f6-8a1f-709a05020b4e" (UID: "dfa654c8-1f3b-43f6-8a1f-709a05020b4e"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:22:10 crc kubenswrapper[4824]: I1211 02:22:10.864551 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68990bd4-eecf-4181-a76c-bf0d24154774-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "68990bd4-eecf-4181-a76c-bf0d24154774" (UID: "68990bd4-eecf-4181-a76c-bf0d24154774"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:22:10 crc kubenswrapper[4824]: I1211 02:22:10.866410 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dfa654c8-1f3b-43f6-8a1f-709a05020b4e-scripts" (OuterVolumeSpecName: "scripts") pod "dfa654c8-1f3b-43f6-8a1f-709a05020b4e" (UID: "dfa654c8-1f3b-43f6-8a1f-709a05020b4e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:22:10 crc kubenswrapper[4824]: I1211 02:22:10.867092 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dfa654c8-1f3b-43f6-8a1f-709a05020b4e-kube-api-access-q6vwj" (OuterVolumeSpecName: "kube-api-access-q6vwj") pod "dfa654c8-1f3b-43f6-8a1f-709a05020b4e" (UID: "dfa654c8-1f3b-43f6-8a1f-709a05020b4e"). InnerVolumeSpecName "kube-api-access-q6vwj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:22:10 crc kubenswrapper[4824]: I1211 02:22:10.867214 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68990bd4-eecf-4181-a76c-bf0d24154774-kube-api-access-c7skk" (OuterVolumeSpecName: "kube-api-access-c7skk") pod "68990bd4-eecf-4181-a76c-bf0d24154774" (UID: "68990bd4-eecf-4181-a76c-bf0d24154774"). InnerVolumeSpecName "kube-api-access-c7skk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:22:10 crc kubenswrapper[4824]: I1211 02:22:10.868953 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "glance") pod "dfa654c8-1f3b-43f6-8a1f-709a05020b4e" (UID: "dfa654c8-1f3b-43f6-8a1f-709a05020b4e"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 11 02:22:10 crc kubenswrapper[4824]: I1211 02:22:10.870224 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68990bd4-eecf-4181-a76c-bf0d24154774-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "68990bd4-eecf-4181-a76c-bf0d24154774" (UID: "68990bd4-eecf-4181-a76c-bf0d24154774"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:22:10 crc kubenswrapper[4824]: I1211 02:22:10.880506 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68990bd4-eecf-4181-a76c-bf0d24154774-scripts" (OuterVolumeSpecName: "scripts") pod "68990bd4-eecf-4181-a76c-bf0d24154774" (UID: "68990bd4-eecf-4181-a76c-bf0d24154774"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:22:10 crc kubenswrapper[4824]: I1211 02:22:10.944010 4824 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/68990bd4-eecf-4181-a76c-bf0d24154774-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:10 crc kubenswrapper[4824]: I1211 02:22:10.944061 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c7skk\" (UniqueName: \"kubernetes.io/projected/68990bd4-eecf-4181-a76c-bf0d24154774-kube-api-access-c7skk\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:10 crc kubenswrapper[4824]: I1211 02:22:10.944083 4824 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Dec 11 02:22:10 crc kubenswrapper[4824]: I1211 02:22:10.944092 4824 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/dfa654c8-1f3b-43f6-8a1f-709a05020b4e-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:10 crc kubenswrapper[4824]: I1211 02:22:10.944101 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q6vwj\" (UniqueName: \"kubernetes.io/projected/dfa654c8-1f3b-43f6-8a1f-709a05020b4e-kube-api-access-q6vwj\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:10 crc kubenswrapper[4824]: I1211 02:22:10.944122 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dfa654c8-1f3b-43f6-8a1f-709a05020b4e-logs\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:10 crc kubenswrapper[4824]: I1211 02:22:10.944130 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dfa654c8-1f3b-43f6-8a1f-709a05020b4e-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:10 crc kubenswrapper[4824]: I1211 02:22:10.944138 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/68990bd4-eecf-4181-a76c-bf0d24154774-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:10 crc kubenswrapper[4824]: I1211 02:22:10.944148 4824 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/68990bd4-eecf-4181-a76c-bf0d24154774-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.007699 4824 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.046194 4824 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.049745 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dfa654c8-1f3b-43f6-8a1f-709a05020b4e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dfa654c8-1f3b-43f6-8a1f-709a05020b4e" (UID: "dfa654c8-1f3b-43f6-8a1f-709a05020b4e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.062513 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68990bd4-eecf-4181-a76c-bf0d24154774-config-data" (OuterVolumeSpecName: "config-data") pod "68990bd4-eecf-4181-a76c-bf0d24154774" (UID: "68990bd4-eecf-4181-a76c-bf0d24154774"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.063868 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68990bd4-eecf-4181-a76c-bf0d24154774-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "68990bd4-eecf-4181-a76c-bf0d24154774" (UID: "68990bd4-eecf-4181-a76c-bf0d24154774"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.082400 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dfa654c8-1f3b-43f6-8a1f-709a05020b4e-config-data" (OuterVolumeSpecName: "config-data") pod "dfa654c8-1f3b-43f6-8a1f-709a05020b4e" (UID: "dfa654c8-1f3b-43f6-8a1f-709a05020b4e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.083517 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dfa654c8-1f3b-43f6-8a1f-709a05020b4e-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "dfa654c8-1f3b-43f6-8a1f-709a05020b4e" (UID: "dfa654c8-1f3b-43f6-8a1f-709a05020b4e"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.154142 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfa654c8-1f3b-43f6-8a1f-709a05020b4e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.154653 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68990bd4-eecf-4181-a76c-bf0d24154774-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.154713 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68990bd4-eecf-4181-a76c-bf0d24154774-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.154763 4824 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dfa654c8-1f3b-43f6-8a1f-709a05020b4e-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.154828 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dfa654c8-1f3b-43f6-8a1f-709a05020b4e-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:11 crc kubenswrapper[4824]: W1211 02:22:11.259419 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcde933a7_1f3a_4567_b6fd_7644a02cb745.slice/crio-9ca61a5fc88dcdae631799d090c27a13dbb35345bc2163cb5d40c51fe7ae43c4 WatchSource:0}: Error finding container 9ca61a5fc88dcdae631799d090c27a13dbb35345bc2163cb5d40c51fe7ae43c4: Status 404 returned error can't find the container with id 9ca61a5fc88dcdae631799d090c27a13dbb35345bc2163cb5d40c51fe7ae43c4 Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.261189 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5b766ddd67-lflrw"] Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.434542 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5fbc99977b-tggml" event={"ID":"0a39253e-8c29-41ae-a242-8c9fccfe2613","Type":"ContainerStarted","Data":"4f93a423b9aeeca2bc45f2bbf56b517529c486e408c84b916c2eb9ed430fa56f"} Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.435733 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-5fbc99977b-tggml" Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.437345 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-752rf" Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.438493 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-752rf" event={"ID":"68990bd4-eecf-4181-a76c-bf0d24154774","Type":"ContainerDied","Data":"dea64b06dcc752831f405e43ed8d5ded3e916e78358991cf31bcf8cfe2ce51de"} Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.438534 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dea64b06dcc752831f405e43ed8d5ded3e916e78358991cf31bcf8cfe2ce51de" Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.439903 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-78b68944cb-4p8xq" event={"ID":"d4663f55-6344-4154-9df6-96a17ca59bdb","Type":"ContainerStarted","Data":"ecc06e354d1582bc54539f434dabaecc67b5bcec1afd95a702e95879b2b8ffe2"} Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.439932 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-78b68944cb-4p8xq" event={"ID":"d4663f55-6344-4154-9df6-96a17ca59bdb","Type":"ContainerStarted","Data":"829ed638ea92c1f91e9cdda3ec17cb49346260079ab443a50586f82ce6c9e0f5"} Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.439943 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-78b68944cb-4p8xq" event={"ID":"d4663f55-6344-4154-9df6-96a17ca59bdb","Type":"ContainerStarted","Data":"c8439216c3f2bd1b9b9b71ad4eaa2d13b8a0dc05ddee38fb37ce982b88ab85ca"} Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.440102 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-78b68944cb-4p8xq" Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.440192 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-78b68944cb-4p8xq" Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.445162 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5b766ddd67-lflrw" event={"ID":"cde933a7-1f3a-4567-b6fd-7644a02cb745","Type":"ContainerStarted","Data":"f967ef0104ccb7d78d525a52e7eddf7476739210674c405491836342009099d7"} Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.445214 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5b766ddd67-lflrw" event={"ID":"cde933a7-1f3a-4567-b6fd-7644a02cb745","Type":"ContainerStarted","Data":"9ca61a5fc88dcdae631799d090c27a13dbb35345bc2163cb5d40c51fe7ae43c4"} Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.447074 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-6wc4l" event={"ID":"89133b02-fcf7-4fe9-b511-c406d1b0d08b","Type":"ContainerStarted","Data":"02c2354e3bfffc0ef0c43117b0a6607d70e56a2fc12c37f28004c3744f830cb3"} Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.447283 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-55f844cf75-6wc4l" Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.448737 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"efa62db6-8875-4ae6-99ea-007e08be02bd","Type":"ContainerStarted","Data":"fec321104810659464b218840e399c47290eb82d1c7ea38755cdcea70b40557b"} Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.462276 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"dfa654c8-1f3b-43f6-8a1f-709a05020b4e","Type":"ContainerDied","Data":"af7573a4ac2063e4700b5375272ebf15aa63938d09730714f024d29f7f1da940"} Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.462497 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.462586 4824 scope.go:117] "RemoveContainer" containerID="ed47e0f4bf032969946a71cb7a3c67c26e5ce39ba4f3607e54fa38b0c5b34968" Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.490968 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-5fbc99977b-tggml" podStartSLOduration=7.49094631 podStartE2EDuration="7.49094631s" podCreationTimestamp="2025-12-11 02:22:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:22:11.475564759 +0000 UTC m=+1273.164602148" watchObservedRunningTime="2025-12-11 02:22:11.49094631 +0000 UTC m=+1273.179983689" Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.500430 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-55f844cf75-6wc4l" podStartSLOduration=7.500411279 podStartE2EDuration="7.500411279s" podCreationTimestamp="2025-12-11 02:22:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:22:11.496442421 +0000 UTC m=+1273.185479800" watchObservedRunningTime="2025-12-11 02:22:11.500411279 +0000 UTC m=+1273.189448648" Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.529135 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-78b68944cb-4p8xq" podStartSLOduration=6.529102007 podStartE2EDuration="6.529102007s" podCreationTimestamp="2025-12-11 02:22:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:22:11.514393333 +0000 UTC m=+1273.203430712" watchObservedRunningTime="2025-12-11 02:22:11.529102007 +0000 UTC m=+1273.218139386" Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.652859 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.661353 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.674395 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 11 02:22:11 crc kubenswrapper[4824]: E1211 02:22:11.674735 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfa654c8-1f3b-43f6-8a1f-709a05020b4e" containerName="glance-httpd" Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.674770 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfa654c8-1f3b-43f6-8a1f-709a05020b4e" containerName="glance-httpd" Dec 11 02:22:11 crc kubenswrapper[4824]: E1211 02:22:11.674786 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68990bd4-eecf-4181-a76c-bf0d24154774" containerName="keystone-bootstrap" Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.674793 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="68990bd4-eecf-4181-a76c-bf0d24154774" containerName="keystone-bootstrap" Dec 11 02:22:11 crc kubenswrapper[4824]: E1211 02:22:11.674817 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfa654c8-1f3b-43f6-8a1f-709a05020b4e" containerName="glance-log" Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.674823 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfa654c8-1f3b-43f6-8a1f-709a05020b4e" containerName="glance-log" Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.674992 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="dfa654c8-1f3b-43f6-8a1f-709a05020b4e" containerName="glance-httpd" Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.675015 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="dfa654c8-1f3b-43f6-8a1f-709a05020b4e" containerName="glance-log" Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.675026 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="68990bd4-eecf-4181-a76c-bf0d24154774" containerName="keystone-bootstrap" Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.675939 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.680017 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.680131 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.710784 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.785059 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-6c99457c66-dcp4p"] Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.786372 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6c99457c66-dcp4p" Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.788609 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.789068 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.789298 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.789606 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.789769 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-8rv79" Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.792667 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.793188 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ee999ccf-334d-49ff-85bf-9c6601b97bf1-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"ee999ccf-334d-49ff-85bf-9c6601b97bf1\") " pod="openstack/glance-default-external-api-0" Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.793284 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee999ccf-334d-49ff-85bf-9c6601b97bf1-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"ee999ccf-334d-49ff-85bf-9c6601b97bf1\") " pod="openstack/glance-default-external-api-0" Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.793312 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee999ccf-334d-49ff-85bf-9c6601b97bf1-config-data\") pod \"glance-default-external-api-0\" (UID: \"ee999ccf-334d-49ff-85bf-9c6601b97bf1\") " pod="openstack/glance-default-external-api-0" Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.793348 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ee999ccf-334d-49ff-85bf-9c6601b97bf1-logs\") pod \"glance-default-external-api-0\" (UID: \"ee999ccf-334d-49ff-85bf-9c6601b97bf1\") " pod="openstack/glance-default-external-api-0" Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.793387 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"ee999ccf-334d-49ff-85bf-9c6601b97bf1\") " pod="openstack/glance-default-external-api-0" Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.793403 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee999ccf-334d-49ff-85bf-9c6601b97bf1-scripts\") pod \"glance-default-external-api-0\" (UID: \"ee999ccf-334d-49ff-85bf-9c6601b97bf1\") " pod="openstack/glance-default-external-api-0" Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.793440 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee999ccf-334d-49ff-85bf-9c6601b97bf1-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"ee999ccf-334d-49ff-85bf-9c6601b97bf1\") " pod="openstack/glance-default-external-api-0" Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.793541 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hzlq2\" (UniqueName: \"kubernetes.io/projected/ee999ccf-334d-49ff-85bf-9c6601b97bf1-kube-api-access-hzlq2\") pod \"glance-default-external-api-0\" (UID: \"ee999ccf-334d-49ff-85bf-9c6601b97bf1\") " pod="openstack/glance-default-external-api-0" Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.794443 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-6c99457c66-dcp4p"] Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.885718 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.894783 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/55bc7333-3e36-4d58-9451-fe018b085c0b-fernet-keys\") pod \"keystone-6c99457c66-dcp4p\" (UID: \"55bc7333-3e36-4d58-9451-fe018b085c0b\") " pod="openstack/keystone-6c99457c66-dcp4p" Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.894822 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hzlq2\" (UniqueName: \"kubernetes.io/projected/ee999ccf-334d-49ff-85bf-9c6601b97bf1-kube-api-access-hzlq2\") pod \"glance-default-external-api-0\" (UID: \"ee999ccf-334d-49ff-85bf-9c6601b97bf1\") " pod="openstack/glance-default-external-api-0" Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.894844 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55bc7333-3e36-4d58-9451-fe018b085c0b-config-data\") pod \"keystone-6c99457c66-dcp4p\" (UID: \"55bc7333-3e36-4d58-9451-fe018b085c0b\") " pod="openstack/keystone-6c99457c66-dcp4p" Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.894864 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/55bc7333-3e36-4d58-9451-fe018b085c0b-internal-tls-certs\") pod \"keystone-6c99457c66-dcp4p\" (UID: \"55bc7333-3e36-4d58-9451-fe018b085c0b\") " pod="openstack/keystone-6c99457c66-dcp4p" Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.894884 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55bc7333-3e36-4d58-9451-fe018b085c0b-combined-ca-bundle\") pod \"keystone-6c99457c66-dcp4p\" (UID: \"55bc7333-3e36-4d58-9451-fe018b085c0b\") " pod="openstack/keystone-6c99457c66-dcp4p" Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.894916 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ee999ccf-334d-49ff-85bf-9c6601b97bf1-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"ee999ccf-334d-49ff-85bf-9c6601b97bf1\") " pod="openstack/glance-default-external-api-0" Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.894880 4824 scope.go:117] "RemoveContainer" containerID="62aff7415eaffd64df1b48803198a5006854ea9c0d791bee37ad995fe18a6352" Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.894971 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d56mh\" (UniqueName: \"kubernetes.io/projected/55bc7333-3e36-4d58-9451-fe018b085c0b-kube-api-access-d56mh\") pod \"keystone-6c99457c66-dcp4p\" (UID: \"55bc7333-3e36-4d58-9451-fe018b085c0b\") " pod="openstack/keystone-6c99457c66-dcp4p" Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.895392 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/55bc7333-3e36-4d58-9451-fe018b085c0b-scripts\") pod \"keystone-6c99457c66-dcp4p\" (UID: \"55bc7333-3e36-4d58-9451-fe018b085c0b\") " pod="openstack/keystone-6c99457c66-dcp4p" Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.895480 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee999ccf-334d-49ff-85bf-9c6601b97bf1-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"ee999ccf-334d-49ff-85bf-9c6601b97bf1\") " pod="openstack/glance-default-external-api-0" Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.895510 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee999ccf-334d-49ff-85bf-9c6601b97bf1-config-data\") pod \"glance-default-external-api-0\" (UID: \"ee999ccf-334d-49ff-85bf-9c6601b97bf1\") " pod="openstack/glance-default-external-api-0" Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.895544 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ee999ccf-334d-49ff-85bf-9c6601b97bf1-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"ee999ccf-334d-49ff-85bf-9c6601b97bf1\") " pod="openstack/glance-default-external-api-0" Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.895568 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ee999ccf-334d-49ff-85bf-9c6601b97bf1-logs\") pod \"glance-default-external-api-0\" (UID: \"ee999ccf-334d-49ff-85bf-9c6601b97bf1\") " pod="openstack/glance-default-external-api-0" Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.895967 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ee999ccf-334d-49ff-85bf-9c6601b97bf1-logs\") pod \"glance-default-external-api-0\" (UID: \"ee999ccf-334d-49ff-85bf-9c6601b97bf1\") " pod="openstack/glance-default-external-api-0" Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.896031 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"ee999ccf-334d-49ff-85bf-9c6601b97bf1\") " pod="openstack/glance-default-external-api-0" Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.896049 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee999ccf-334d-49ff-85bf-9c6601b97bf1-scripts\") pod \"glance-default-external-api-0\" (UID: \"ee999ccf-334d-49ff-85bf-9c6601b97bf1\") " pod="openstack/glance-default-external-api-0" Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.896068 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/55bc7333-3e36-4d58-9451-fe018b085c0b-credential-keys\") pod \"keystone-6c99457c66-dcp4p\" (UID: \"55bc7333-3e36-4d58-9451-fe018b085c0b\") " pod="openstack/keystone-6c99457c66-dcp4p" Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.896084 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/55bc7333-3e36-4d58-9451-fe018b085c0b-public-tls-certs\") pod \"keystone-6c99457c66-dcp4p\" (UID: \"55bc7333-3e36-4d58-9451-fe018b085c0b\") " pod="openstack/keystone-6c99457c66-dcp4p" Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.896129 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee999ccf-334d-49ff-85bf-9c6601b97bf1-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"ee999ccf-334d-49ff-85bf-9c6601b97bf1\") " pod="openstack/glance-default-external-api-0" Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.897135 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"ee999ccf-334d-49ff-85bf-9c6601b97bf1\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/glance-default-external-api-0" Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.902378 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee999ccf-334d-49ff-85bf-9c6601b97bf1-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"ee999ccf-334d-49ff-85bf-9c6601b97bf1\") " pod="openstack/glance-default-external-api-0" Dec 11 02:22:11 crc kubenswrapper[4824]: W1211 02:22:11.903824 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode7794eea_3d90_4bd6_8d69_732d80385243.slice/crio-0168d06ac466cff6709aa1e943886d2ba1cfb35d6cf024f295c68b7819325d02 WatchSource:0}: Error finding container 0168d06ac466cff6709aa1e943886d2ba1cfb35d6cf024f295c68b7819325d02: Status 404 returned error can't find the container with id 0168d06ac466cff6709aa1e943886d2ba1cfb35d6cf024f295c68b7819325d02 Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.904393 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee999ccf-334d-49ff-85bf-9c6601b97bf1-scripts\") pod \"glance-default-external-api-0\" (UID: \"ee999ccf-334d-49ff-85bf-9c6601b97bf1\") " pod="openstack/glance-default-external-api-0" Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.906332 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee999ccf-334d-49ff-85bf-9c6601b97bf1-config-data\") pod \"glance-default-external-api-0\" (UID: \"ee999ccf-334d-49ff-85bf-9c6601b97bf1\") " pod="openstack/glance-default-external-api-0" Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.911481 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee999ccf-334d-49ff-85bf-9c6601b97bf1-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"ee999ccf-334d-49ff-85bf-9c6601b97bf1\") " pod="openstack/glance-default-external-api-0" Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.919253 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hzlq2\" (UniqueName: \"kubernetes.io/projected/ee999ccf-334d-49ff-85bf-9c6601b97bf1-kube-api-access-hzlq2\") pod \"glance-default-external-api-0\" (UID: \"ee999ccf-334d-49ff-85bf-9c6601b97bf1\") " pod="openstack/glance-default-external-api-0" Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.933478 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"ee999ccf-334d-49ff-85bf-9c6601b97bf1\") " pod="openstack/glance-default-external-api-0" Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.997432 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55bc7333-3e36-4d58-9451-fe018b085c0b-config-data\") pod \"keystone-6c99457c66-dcp4p\" (UID: \"55bc7333-3e36-4d58-9451-fe018b085c0b\") " pod="openstack/keystone-6c99457c66-dcp4p" Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.997489 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/55bc7333-3e36-4d58-9451-fe018b085c0b-internal-tls-certs\") pod \"keystone-6c99457c66-dcp4p\" (UID: \"55bc7333-3e36-4d58-9451-fe018b085c0b\") " pod="openstack/keystone-6c99457c66-dcp4p" Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.997517 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55bc7333-3e36-4d58-9451-fe018b085c0b-combined-ca-bundle\") pod \"keystone-6c99457c66-dcp4p\" (UID: \"55bc7333-3e36-4d58-9451-fe018b085c0b\") " pod="openstack/keystone-6c99457c66-dcp4p" Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.997558 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d56mh\" (UniqueName: \"kubernetes.io/projected/55bc7333-3e36-4d58-9451-fe018b085c0b-kube-api-access-d56mh\") pod \"keystone-6c99457c66-dcp4p\" (UID: \"55bc7333-3e36-4d58-9451-fe018b085c0b\") " pod="openstack/keystone-6c99457c66-dcp4p" Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.997593 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/55bc7333-3e36-4d58-9451-fe018b085c0b-scripts\") pod \"keystone-6c99457c66-dcp4p\" (UID: \"55bc7333-3e36-4d58-9451-fe018b085c0b\") " pod="openstack/keystone-6c99457c66-dcp4p" Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.997723 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/55bc7333-3e36-4d58-9451-fe018b085c0b-credential-keys\") pod \"keystone-6c99457c66-dcp4p\" (UID: \"55bc7333-3e36-4d58-9451-fe018b085c0b\") " pod="openstack/keystone-6c99457c66-dcp4p" Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.997757 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/55bc7333-3e36-4d58-9451-fe018b085c0b-public-tls-certs\") pod \"keystone-6c99457c66-dcp4p\" (UID: \"55bc7333-3e36-4d58-9451-fe018b085c0b\") " pod="openstack/keystone-6c99457c66-dcp4p" Dec 11 02:22:11 crc kubenswrapper[4824]: I1211 02:22:11.997858 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/55bc7333-3e36-4d58-9451-fe018b085c0b-fernet-keys\") pod \"keystone-6c99457c66-dcp4p\" (UID: \"55bc7333-3e36-4d58-9451-fe018b085c0b\") " pod="openstack/keystone-6c99457c66-dcp4p" Dec 11 02:22:12 crc kubenswrapper[4824]: I1211 02:22:12.004230 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/55bc7333-3e36-4d58-9451-fe018b085c0b-fernet-keys\") pod \"keystone-6c99457c66-dcp4p\" (UID: \"55bc7333-3e36-4d58-9451-fe018b085c0b\") " pod="openstack/keystone-6c99457c66-dcp4p" Dec 11 02:22:12 crc kubenswrapper[4824]: I1211 02:22:12.005236 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55bc7333-3e36-4d58-9451-fe018b085c0b-combined-ca-bundle\") pod \"keystone-6c99457c66-dcp4p\" (UID: \"55bc7333-3e36-4d58-9451-fe018b085c0b\") " pod="openstack/keystone-6c99457c66-dcp4p" Dec 11 02:22:12 crc kubenswrapper[4824]: I1211 02:22:12.006544 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/55bc7333-3e36-4d58-9451-fe018b085c0b-public-tls-certs\") pod \"keystone-6c99457c66-dcp4p\" (UID: \"55bc7333-3e36-4d58-9451-fe018b085c0b\") " pod="openstack/keystone-6c99457c66-dcp4p" Dec 11 02:22:12 crc kubenswrapper[4824]: I1211 02:22:12.006605 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/55bc7333-3e36-4d58-9451-fe018b085c0b-internal-tls-certs\") pod \"keystone-6c99457c66-dcp4p\" (UID: \"55bc7333-3e36-4d58-9451-fe018b085c0b\") " pod="openstack/keystone-6c99457c66-dcp4p" Dec 11 02:22:12 crc kubenswrapper[4824]: I1211 02:22:12.006758 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/55bc7333-3e36-4d58-9451-fe018b085c0b-scripts\") pod \"keystone-6c99457c66-dcp4p\" (UID: \"55bc7333-3e36-4d58-9451-fe018b085c0b\") " pod="openstack/keystone-6c99457c66-dcp4p" Dec 11 02:22:12 crc kubenswrapper[4824]: I1211 02:22:12.007091 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 11 02:22:12 crc kubenswrapper[4824]: I1211 02:22:12.007422 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55bc7333-3e36-4d58-9451-fe018b085c0b-config-data\") pod \"keystone-6c99457c66-dcp4p\" (UID: \"55bc7333-3e36-4d58-9451-fe018b085c0b\") " pod="openstack/keystone-6c99457c66-dcp4p" Dec 11 02:22:12 crc kubenswrapper[4824]: I1211 02:22:12.013585 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/55bc7333-3e36-4d58-9451-fe018b085c0b-credential-keys\") pod \"keystone-6c99457c66-dcp4p\" (UID: \"55bc7333-3e36-4d58-9451-fe018b085c0b\") " pod="openstack/keystone-6c99457c66-dcp4p" Dec 11 02:22:12 crc kubenswrapper[4824]: I1211 02:22:12.017674 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d56mh\" (UniqueName: \"kubernetes.io/projected/55bc7333-3e36-4d58-9451-fe018b085c0b-kube-api-access-d56mh\") pod \"keystone-6c99457c66-dcp4p\" (UID: \"55bc7333-3e36-4d58-9451-fe018b085c0b\") " pod="openstack/keystone-6c99457c66-dcp4p" Dec 11 02:22:12 crc kubenswrapper[4824]: I1211 02:22:12.105287 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6c99457c66-dcp4p" Dec 11 02:22:12 crc kubenswrapper[4824]: I1211 02:22:12.485036 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e7794eea-3d90-4bd6-8d69-732d80385243","Type":"ContainerStarted","Data":"0168d06ac466cff6709aa1e943886d2ba1cfb35d6cf024f295c68b7819325d02"} Dec 11 02:22:12 crc kubenswrapper[4824]: I1211 02:22:12.491819 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5b766ddd67-lflrw" event={"ID":"cde933a7-1f3a-4567-b6fd-7644a02cb745","Type":"ContainerStarted","Data":"712555cb6339b61766e4d9ad217f895f1e3698419f8ddbb30ccd9cf7e407a402"} Dec 11 02:22:12 crc kubenswrapper[4824]: I1211 02:22:12.492367 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-5b766ddd67-lflrw" Dec 11 02:22:12 crc kubenswrapper[4824]: I1211 02:22:12.510728 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-5b766ddd67-lflrw" podStartSLOduration=5.510712767 podStartE2EDuration="5.510712767s" podCreationTimestamp="2025-12-11 02:22:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:22:12.507984203 +0000 UTC m=+1274.197021582" watchObservedRunningTime="2025-12-11 02:22:12.510712767 +0000 UTC m=+1274.199750146" Dec 11 02:22:12 crc kubenswrapper[4824]: I1211 02:22:12.594255 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 11 02:22:12 crc kubenswrapper[4824]: I1211 02:22:12.678636 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dfa654c8-1f3b-43f6-8a1f-709a05020b4e" path="/var/lib/kubelet/pods/dfa654c8-1f3b-43f6-8a1f-709a05020b4e/volumes" Dec 11 02:22:12 crc kubenswrapper[4824]: I1211 02:22:12.679570 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-6c99457c66-dcp4p"] Dec 11 02:22:12 crc kubenswrapper[4824]: W1211 02:22:12.699966 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod55bc7333_3e36_4d58_9451_fe018b085c0b.slice/crio-76b1d15980e3887a26eec907fa7db90ffd092a562e991843df6fcd7405268af1 WatchSource:0}: Error finding container 76b1d15980e3887a26eec907fa7db90ffd092a562e991843df6fcd7405268af1: Status 404 returned error can't find the container with id 76b1d15980e3887a26eec907fa7db90ffd092a562e991843df6fcd7405268af1 Dec 11 02:22:13 crc kubenswrapper[4824]: I1211 02:22:13.530128 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6c99457c66-dcp4p" event={"ID":"55bc7333-3e36-4d58-9451-fe018b085c0b","Type":"ContainerStarted","Data":"e0af3fbedf3ddd1932e3f0ec79edefef90c7fda555f70a9688614c5deb3d47a6"} Dec 11 02:22:13 crc kubenswrapper[4824]: I1211 02:22:13.530442 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6c99457c66-dcp4p" event={"ID":"55bc7333-3e36-4d58-9451-fe018b085c0b","Type":"ContainerStarted","Data":"76b1d15980e3887a26eec907fa7db90ffd092a562e991843df6fcd7405268af1"} Dec 11 02:22:13 crc kubenswrapper[4824]: I1211 02:22:13.530495 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-6c99457c66-dcp4p" Dec 11 02:22:13 crc kubenswrapper[4824]: I1211 02:22:13.539201 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ee999ccf-334d-49ff-85bf-9c6601b97bf1","Type":"ContainerStarted","Data":"2fa8b7e2e0ed1d8d5b5c08d6ab83a548819dfb7841d471d788f368f2ca2743ae"} Dec 11 02:22:13 crc kubenswrapper[4824]: I1211 02:22:13.539226 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ee999ccf-334d-49ff-85bf-9c6601b97bf1","Type":"ContainerStarted","Data":"313644e840e2f94a47045b60e8e6a6b4395a3c55059af1cec115f6d6a77b43ce"} Dec 11 02:22:13 crc kubenswrapper[4824]: I1211 02:22:13.548610 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-6c99457c66-dcp4p" podStartSLOduration=2.548586632 podStartE2EDuration="2.548586632s" podCreationTimestamp="2025-12-11 02:22:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:22:13.548092917 +0000 UTC m=+1275.237130296" watchObservedRunningTime="2025-12-11 02:22:13.548586632 +0000 UTC m=+1275.237624021" Dec 11 02:22:13 crc kubenswrapper[4824]: I1211 02:22:13.559620 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e7794eea-3d90-4bd6-8d69-732d80385243","Type":"ContainerStarted","Data":"4364c167b32c39ebe179b655d9e1d469552dee7e3454b999439b9e634cc9cf53"} Dec 11 02:22:13 crc kubenswrapper[4824]: I1211 02:22:13.559661 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e7794eea-3d90-4bd6-8d69-732d80385243","Type":"ContainerStarted","Data":"8411d991de0195412273e5c06a0d59951db008bb73091da57111623429573c22"} Dec 11 02:22:13 crc kubenswrapper[4824]: I1211 02:22:13.565838 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-787649dcdd-ts6qs" podUID="f15f9ed0-98f2-42fe-8006-2142098cc2b6" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.147:8443: connect: connection refused" Dec 11 02:22:13 crc kubenswrapper[4824]: I1211 02:22:13.584306 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-74cc6c88d8-zvlk7" podUID="1da5eb01-d709-42ad-b5f1-e7ea6310cc9e" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.148:8443: connect: connection refused" Dec 11 02:22:13 crc kubenswrapper[4824]: I1211 02:22:13.588873 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=8.588853016 podStartE2EDuration="8.588853016s" podCreationTimestamp="2025-12-11 02:22:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:22:13.576214349 +0000 UTC m=+1275.265251738" watchObservedRunningTime="2025-12-11 02:22:13.588853016 +0000 UTC m=+1275.277890395" Dec 11 02:22:14 crc kubenswrapper[4824]: I1211 02:22:14.566470 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ee999ccf-334d-49ff-85bf-9c6601b97bf1","Type":"ContainerStarted","Data":"22b549fc7d47ca2a6cc9968758dbb602d55f41fdce561681cce9f4f80206b283"} Dec 11 02:22:14 crc kubenswrapper[4824]: I1211 02:22:14.641269 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.641255569 podStartE2EDuration="3.641255569s" podCreationTimestamp="2025-12-11 02:22:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:22:14.633189767 +0000 UTC m=+1276.322227166" watchObservedRunningTime="2025-12-11 02:22:14.641255569 +0000 UTC m=+1276.330292948" Dec 11 02:22:15 crc kubenswrapper[4824]: I1211 02:22:15.574449 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-pbsrf" event={"ID":"d536a1b1-baa2-497a-ba0c-ad03599168cc","Type":"ContainerStarted","Data":"a3b44f56c6628b1b240b0bad78da5c6cc9a39004a9cb7deab6fe0bd7586b95a4"} Dec 11 02:22:15 crc kubenswrapper[4824]: I1211 02:22:15.600481 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-pbsrf" podStartSLOduration=2.141150304 podStartE2EDuration="41.600450444s" podCreationTimestamp="2025-12-11 02:21:34 +0000 UTC" firstStartedPulling="2025-12-11 02:21:35.837405134 +0000 UTC m=+1237.526442513" lastFinishedPulling="2025-12-11 02:22:15.296705274 +0000 UTC m=+1276.985742653" observedRunningTime="2025-12-11 02:22:15.592528717 +0000 UTC m=+1277.281566096" watchObservedRunningTime="2025-12-11 02:22:15.600450444 +0000 UTC m=+1277.289487863" Dec 11 02:22:16 crc kubenswrapper[4824]: I1211 02:22:16.131757 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 11 02:22:16 crc kubenswrapper[4824]: I1211 02:22:16.135783 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 11 02:22:16 crc kubenswrapper[4824]: I1211 02:22:16.181194 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 11 02:22:16 crc kubenswrapper[4824]: I1211 02:22:16.185759 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 11 02:22:16 crc kubenswrapper[4824]: I1211 02:22:16.582563 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-tn658" event={"ID":"8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7","Type":"ContainerStarted","Data":"6b97e27f717fafb1660abc5454cbf78410b26ec165a238a422af3cc98a9cf41b"} Dec 11 02:22:16 crc kubenswrapper[4824]: I1211 02:22:16.582772 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 11 02:22:16 crc kubenswrapper[4824]: I1211 02:22:16.582814 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 11 02:22:16 crc kubenswrapper[4824]: I1211 02:22:16.611944 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-tn658" podStartSLOduration=2.452569526 podStartE2EDuration="42.611925484s" podCreationTimestamp="2025-12-11 02:21:34 +0000 UTC" firstStartedPulling="2025-12-11 02:21:35.623594901 +0000 UTC m=+1237.312632280" lastFinishedPulling="2025-12-11 02:22:15.782950819 +0000 UTC m=+1277.471988238" observedRunningTime="2025-12-11 02:22:16.604348436 +0000 UTC m=+1278.293385815" watchObservedRunningTime="2025-12-11 02:22:16.611925484 +0000 UTC m=+1278.300962863" Dec 11 02:22:18 crc kubenswrapper[4824]: I1211 02:22:18.602641 4824 generic.go:334] "Generic (PLEG): container finished" podID="d536a1b1-baa2-497a-ba0c-ad03599168cc" containerID="a3b44f56c6628b1b240b0bad78da5c6cc9a39004a9cb7deab6fe0bd7586b95a4" exitCode=0 Dec 11 02:22:18 crc kubenswrapper[4824]: I1211 02:22:18.602689 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-pbsrf" event={"ID":"d536a1b1-baa2-497a-ba0c-ad03599168cc","Type":"ContainerDied","Data":"a3b44f56c6628b1b240b0bad78da5c6cc9a39004a9cb7deab6fe0bd7586b95a4"} Dec 11 02:22:19 crc kubenswrapper[4824]: I1211 02:22:19.498862 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 11 02:22:19 crc kubenswrapper[4824]: I1211 02:22:19.735316 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-55f844cf75-6wc4l" Dec 11 02:22:19 crc kubenswrapper[4824]: I1211 02:22:19.802748 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-mxg84"] Dec 11 02:22:19 crc kubenswrapper[4824]: I1211 02:22:19.803018 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-785d8bcb8c-mxg84" podUID="a36c6797-0e71-4dd1-b2ae-fb540f63e794" containerName="dnsmasq-dns" containerID="cri-o://c3f9dde6135d2febf970553b4531a0212d07687850754b4751ff75ce07523c00" gracePeriod=10 Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.096529 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-pbsrf" Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.164792 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d536a1b1-baa2-497a-ba0c-ad03599168cc-combined-ca-bundle\") pod \"d536a1b1-baa2-497a-ba0c-ad03599168cc\" (UID: \"d536a1b1-baa2-497a-ba0c-ad03599168cc\") " Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.164866 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jh9z2\" (UniqueName: \"kubernetes.io/projected/d536a1b1-baa2-497a-ba0c-ad03599168cc-kube-api-access-jh9z2\") pod \"d536a1b1-baa2-497a-ba0c-ad03599168cc\" (UID: \"d536a1b1-baa2-497a-ba0c-ad03599168cc\") " Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.165007 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d536a1b1-baa2-497a-ba0c-ad03599168cc-db-sync-config-data\") pod \"d536a1b1-baa2-497a-ba0c-ad03599168cc\" (UID: \"d536a1b1-baa2-497a-ba0c-ad03599168cc\") " Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.174628 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d536a1b1-baa2-497a-ba0c-ad03599168cc-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "d536a1b1-baa2-497a-ba0c-ad03599168cc" (UID: "d536a1b1-baa2-497a-ba0c-ad03599168cc"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.174751 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d536a1b1-baa2-497a-ba0c-ad03599168cc-kube-api-access-jh9z2" (OuterVolumeSpecName: "kube-api-access-jh9z2") pod "d536a1b1-baa2-497a-ba0c-ad03599168cc" (UID: "d536a1b1-baa2-497a-ba0c-ad03599168cc"). InnerVolumeSpecName "kube-api-access-jh9z2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.201308 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d536a1b1-baa2-497a-ba0c-ad03599168cc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d536a1b1-baa2-497a-ba0c-ad03599168cc" (UID: "d536a1b1-baa2-497a-ba0c-ad03599168cc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.268501 4824 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d536a1b1-baa2-497a-ba0c-ad03599168cc-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.268535 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d536a1b1-baa2-497a-ba0c-ad03599168cc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.268544 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jh9z2\" (UniqueName: \"kubernetes.io/projected/d536a1b1-baa2-497a-ba0c-ad03599168cc-kube-api-access-jh9z2\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.296464 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-mxg84" Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.369702 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a36c6797-0e71-4dd1-b2ae-fb540f63e794-config\") pod \"a36c6797-0e71-4dd1-b2ae-fb540f63e794\" (UID: \"a36c6797-0e71-4dd1-b2ae-fb540f63e794\") " Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.369792 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a36c6797-0e71-4dd1-b2ae-fb540f63e794-dns-swift-storage-0\") pod \"a36c6797-0e71-4dd1-b2ae-fb540f63e794\" (UID: \"a36c6797-0e71-4dd1-b2ae-fb540f63e794\") " Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.369831 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a36c6797-0e71-4dd1-b2ae-fb540f63e794-ovsdbserver-sb\") pod \"a36c6797-0e71-4dd1-b2ae-fb540f63e794\" (UID: \"a36c6797-0e71-4dd1-b2ae-fb540f63e794\") " Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.369882 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a36c6797-0e71-4dd1-b2ae-fb540f63e794-ovsdbserver-nb\") pod \"a36c6797-0e71-4dd1-b2ae-fb540f63e794\" (UID: \"a36c6797-0e71-4dd1-b2ae-fb540f63e794\") " Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.369937 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xrq8w\" (UniqueName: \"kubernetes.io/projected/a36c6797-0e71-4dd1-b2ae-fb540f63e794-kube-api-access-xrq8w\") pod \"a36c6797-0e71-4dd1-b2ae-fb540f63e794\" (UID: \"a36c6797-0e71-4dd1-b2ae-fb540f63e794\") " Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.369974 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a36c6797-0e71-4dd1-b2ae-fb540f63e794-dns-svc\") pod \"a36c6797-0e71-4dd1-b2ae-fb540f63e794\" (UID: \"a36c6797-0e71-4dd1-b2ae-fb540f63e794\") " Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.383316 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a36c6797-0e71-4dd1-b2ae-fb540f63e794-kube-api-access-xrq8w" (OuterVolumeSpecName: "kube-api-access-xrq8w") pod "a36c6797-0e71-4dd1-b2ae-fb540f63e794" (UID: "a36c6797-0e71-4dd1-b2ae-fb540f63e794"). InnerVolumeSpecName "kube-api-access-xrq8w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.433817 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a36c6797-0e71-4dd1-b2ae-fb540f63e794-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "a36c6797-0e71-4dd1-b2ae-fb540f63e794" (UID: "a36c6797-0e71-4dd1-b2ae-fb540f63e794"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.436192 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a36c6797-0e71-4dd1-b2ae-fb540f63e794-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a36c6797-0e71-4dd1-b2ae-fb540f63e794" (UID: "a36c6797-0e71-4dd1-b2ae-fb540f63e794"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.438038 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a36c6797-0e71-4dd1-b2ae-fb540f63e794-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a36c6797-0e71-4dd1-b2ae-fb540f63e794" (UID: "a36c6797-0e71-4dd1-b2ae-fb540f63e794"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.441450 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a36c6797-0e71-4dd1-b2ae-fb540f63e794-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a36c6797-0e71-4dd1-b2ae-fb540f63e794" (UID: "a36c6797-0e71-4dd1-b2ae-fb540f63e794"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.442300 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a36c6797-0e71-4dd1-b2ae-fb540f63e794-config" (OuterVolumeSpecName: "config") pod "a36c6797-0e71-4dd1-b2ae-fb540f63e794" (UID: "a36c6797-0e71-4dd1-b2ae-fb540f63e794"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.471997 4824 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a36c6797-0e71-4dd1-b2ae-fb540f63e794-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.472023 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a36c6797-0e71-4dd1-b2ae-fb540f63e794-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.472033 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a36c6797-0e71-4dd1-b2ae-fb540f63e794-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.472042 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xrq8w\" (UniqueName: \"kubernetes.io/projected/a36c6797-0e71-4dd1-b2ae-fb540f63e794-kube-api-access-xrq8w\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.472054 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a36c6797-0e71-4dd1-b2ae-fb540f63e794-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.472062 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a36c6797-0e71-4dd1-b2ae-fb540f63e794-config\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.653501 4824 generic.go:334] "Generic (PLEG): container finished" podID="a36c6797-0e71-4dd1-b2ae-fb540f63e794" containerID="c3f9dde6135d2febf970553b4531a0212d07687850754b4751ff75ce07523c00" exitCode=0 Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.653559 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-mxg84" event={"ID":"a36c6797-0e71-4dd1-b2ae-fb540f63e794","Type":"ContainerDied","Data":"c3f9dde6135d2febf970553b4531a0212d07687850754b4751ff75ce07523c00"} Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.653618 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-mxg84" Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.653944 4824 scope.go:117] "RemoveContainer" containerID="c3f9dde6135d2febf970553b4531a0212d07687850754b4751ff75ce07523c00" Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.653874 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-mxg84" event={"ID":"a36c6797-0e71-4dd1-b2ae-fb540f63e794","Type":"ContainerDied","Data":"0acb45c451d715906cd7ce451913139c9c04e01e1d5c35766d7143e2e63cade4"} Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.656701 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-pbsrf" event={"ID":"d536a1b1-baa2-497a-ba0c-ad03599168cc","Type":"ContainerDied","Data":"74fb14189c9001a125a5b8284f126c27ce67a62e029fe6f80e77f8d47a0075b9"} Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.656773 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="74fb14189c9001a125a5b8284f126c27ce67a62e029fe6f80e77f8d47a0075b9" Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.656872 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-pbsrf" Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.687461 4824 scope.go:117] "RemoveContainer" containerID="634f409a880778b883a02c8949b0f0ac6c137e947d48e01593f6dc39b1311df4" Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.702254 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-mxg84"] Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.709401 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-mxg84"] Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.756136 4824 scope.go:117] "RemoveContainer" containerID="c3f9dde6135d2febf970553b4531a0212d07687850754b4751ff75ce07523c00" Dec 11 02:22:20 crc kubenswrapper[4824]: E1211 02:22:20.756706 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c3f9dde6135d2febf970553b4531a0212d07687850754b4751ff75ce07523c00\": container with ID starting with c3f9dde6135d2febf970553b4531a0212d07687850754b4751ff75ce07523c00 not found: ID does not exist" containerID="c3f9dde6135d2febf970553b4531a0212d07687850754b4751ff75ce07523c00" Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.756734 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c3f9dde6135d2febf970553b4531a0212d07687850754b4751ff75ce07523c00"} err="failed to get container status \"c3f9dde6135d2febf970553b4531a0212d07687850754b4751ff75ce07523c00\": rpc error: code = NotFound desc = could not find container \"c3f9dde6135d2febf970553b4531a0212d07687850754b4751ff75ce07523c00\": container with ID starting with c3f9dde6135d2febf970553b4531a0212d07687850754b4751ff75ce07523c00 not found: ID does not exist" Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.756751 4824 scope.go:117] "RemoveContainer" containerID="634f409a880778b883a02c8949b0f0ac6c137e947d48e01593f6dc39b1311df4" Dec 11 02:22:20 crc kubenswrapper[4824]: E1211 02:22:20.757087 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"634f409a880778b883a02c8949b0f0ac6c137e947d48e01593f6dc39b1311df4\": container with ID starting with 634f409a880778b883a02c8949b0f0ac6c137e947d48e01593f6dc39b1311df4 not found: ID does not exist" containerID="634f409a880778b883a02c8949b0f0ac6c137e947d48e01593f6dc39b1311df4" Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.757140 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"634f409a880778b883a02c8949b0f0ac6c137e947d48e01593f6dc39b1311df4"} err="failed to get container status \"634f409a880778b883a02c8949b0f0ac6c137e947d48e01593f6dc39b1311df4\": rpc error: code = NotFound desc = could not find container \"634f409a880778b883a02c8949b0f0ac6c137e947d48e01593f6dc39b1311df4\": container with ID starting with 634f409a880778b883a02c8949b0f0ac6c137e947d48e01593f6dc39b1311df4 not found: ID does not exist" Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.852277 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-6967b7c86c-47gl7"] Dec 11 02:22:20 crc kubenswrapper[4824]: E1211 02:22:20.852635 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a36c6797-0e71-4dd1-b2ae-fb540f63e794" containerName="init" Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.852652 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="a36c6797-0e71-4dd1-b2ae-fb540f63e794" containerName="init" Dec 11 02:22:20 crc kubenswrapper[4824]: E1211 02:22:20.852670 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d536a1b1-baa2-497a-ba0c-ad03599168cc" containerName="barbican-db-sync" Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.852677 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d536a1b1-baa2-497a-ba0c-ad03599168cc" containerName="barbican-db-sync" Dec 11 02:22:20 crc kubenswrapper[4824]: E1211 02:22:20.852698 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a36c6797-0e71-4dd1-b2ae-fb540f63e794" containerName="dnsmasq-dns" Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.852704 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="a36c6797-0e71-4dd1-b2ae-fb540f63e794" containerName="dnsmasq-dns" Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.852848 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="a36c6797-0e71-4dd1-b2ae-fb540f63e794" containerName="dnsmasq-dns" Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.852875 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="d536a1b1-baa2-497a-ba0c-ad03599168cc" containerName="barbican-db-sync" Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.853763 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-6967b7c86c-47gl7" Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.858483 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.858617 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.858738 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-bdr5s" Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.866993 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-6967b7c86c-47gl7"] Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.884232 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ggvxt\" (UniqueName: \"kubernetes.io/projected/33482b7a-6aaa-48c4-9a46-c3de05cacebe-kube-api-access-ggvxt\") pod \"barbican-worker-6967b7c86c-47gl7\" (UID: \"33482b7a-6aaa-48c4-9a46-c3de05cacebe\") " pod="openstack/barbican-worker-6967b7c86c-47gl7" Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.884277 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/33482b7a-6aaa-48c4-9a46-c3de05cacebe-logs\") pod \"barbican-worker-6967b7c86c-47gl7\" (UID: \"33482b7a-6aaa-48c4-9a46-c3de05cacebe\") " pod="openstack/barbican-worker-6967b7c86c-47gl7" Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.884301 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/33482b7a-6aaa-48c4-9a46-c3de05cacebe-config-data-custom\") pod \"barbican-worker-6967b7c86c-47gl7\" (UID: \"33482b7a-6aaa-48c4-9a46-c3de05cacebe\") " pod="openstack/barbican-worker-6967b7c86c-47gl7" Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.884350 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33482b7a-6aaa-48c4-9a46-c3de05cacebe-combined-ca-bundle\") pod \"barbican-worker-6967b7c86c-47gl7\" (UID: \"33482b7a-6aaa-48c4-9a46-c3de05cacebe\") " pod="openstack/barbican-worker-6967b7c86c-47gl7" Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.884388 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33482b7a-6aaa-48c4-9a46-c3de05cacebe-config-data\") pod \"barbican-worker-6967b7c86c-47gl7\" (UID: \"33482b7a-6aaa-48c4-9a46-c3de05cacebe\") " pod="openstack/barbican-worker-6967b7c86c-47gl7" Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.901797 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-664bd5db46-xt8tr"] Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.903223 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-664bd5db46-xt8tr" Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.918958 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-664bd5db46-xt8tr"] Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.939657 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.963469 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-p9vh6"] Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.964948 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-p9vh6" Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.983174 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-p9vh6"] Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.985362 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ggvxt\" (UniqueName: \"kubernetes.io/projected/33482b7a-6aaa-48c4-9a46-c3de05cacebe-kube-api-access-ggvxt\") pod \"barbican-worker-6967b7c86c-47gl7\" (UID: \"33482b7a-6aaa-48c4-9a46-c3de05cacebe\") " pod="openstack/barbican-worker-6967b7c86c-47gl7" Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.985397 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/33482b7a-6aaa-48c4-9a46-c3de05cacebe-logs\") pod \"barbican-worker-6967b7c86c-47gl7\" (UID: \"33482b7a-6aaa-48c4-9a46-c3de05cacebe\") " pod="openstack/barbican-worker-6967b7c86c-47gl7" Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.985424 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/33482b7a-6aaa-48c4-9a46-c3de05cacebe-config-data-custom\") pod \"barbican-worker-6967b7c86c-47gl7\" (UID: \"33482b7a-6aaa-48c4-9a46-c3de05cacebe\") " pod="openstack/barbican-worker-6967b7c86c-47gl7" Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.985453 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rqw5q\" (UniqueName: \"kubernetes.io/projected/cd747b97-a287-4df6-8b35-f30121c8c0b9-kube-api-access-rqw5q\") pod \"barbican-keystone-listener-664bd5db46-xt8tr\" (UID: \"cd747b97-a287-4df6-8b35-f30121c8c0b9\") " pod="openstack/barbican-keystone-listener-664bd5db46-xt8tr" Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.985475 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0ae0553d-4ec8-425b-bdf4-b7992b7cbda6-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-p9vh6\" (UID: \"0ae0553d-4ec8-425b-bdf4-b7992b7cbda6\") " pod="openstack/dnsmasq-dns-85ff748b95-p9vh6" Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.985507 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd747b97-a287-4df6-8b35-f30121c8c0b9-config-data\") pod \"barbican-keystone-listener-664bd5db46-xt8tr\" (UID: \"cd747b97-a287-4df6-8b35-f30121c8c0b9\") " pod="openstack/barbican-keystone-listener-664bd5db46-xt8tr" Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.985528 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0ae0553d-4ec8-425b-bdf4-b7992b7cbda6-dns-svc\") pod \"dnsmasq-dns-85ff748b95-p9vh6\" (UID: \"0ae0553d-4ec8-425b-bdf4-b7992b7cbda6\") " pod="openstack/dnsmasq-dns-85ff748b95-p9vh6" Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.985549 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33482b7a-6aaa-48c4-9a46-c3de05cacebe-combined-ca-bundle\") pod \"barbican-worker-6967b7c86c-47gl7\" (UID: \"33482b7a-6aaa-48c4-9a46-c3de05cacebe\") " pod="openstack/barbican-worker-6967b7c86c-47gl7" Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.985572 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd747b97-a287-4df6-8b35-f30121c8c0b9-combined-ca-bundle\") pod \"barbican-keystone-listener-664bd5db46-xt8tr\" (UID: \"cd747b97-a287-4df6-8b35-f30121c8c0b9\") " pod="openstack/barbican-keystone-listener-664bd5db46-xt8tr" Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.985596 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0ae0553d-4ec8-425b-bdf4-b7992b7cbda6-config\") pod \"dnsmasq-dns-85ff748b95-p9vh6\" (UID: \"0ae0553d-4ec8-425b-bdf4-b7992b7cbda6\") " pod="openstack/dnsmasq-dns-85ff748b95-p9vh6" Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.985617 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxr9p\" (UniqueName: \"kubernetes.io/projected/0ae0553d-4ec8-425b-bdf4-b7992b7cbda6-kube-api-access-gxr9p\") pod \"dnsmasq-dns-85ff748b95-p9vh6\" (UID: \"0ae0553d-4ec8-425b-bdf4-b7992b7cbda6\") " pod="openstack/dnsmasq-dns-85ff748b95-p9vh6" Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.985645 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33482b7a-6aaa-48c4-9a46-c3de05cacebe-config-data\") pod \"barbican-worker-6967b7c86c-47gl7\" (UID: \"33482b7a-6aaa-48c4-9a46-c3de05cacebe\") " pod="openstack/barbican-worker-6967b7c86c-47gl7" Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.985721 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd747b97-a287-4df6-8b35-f30121c8c0b9-logs\") pod \"barbican-keystone-listener-664bd5db46-xt8tr\" (UID: \"cd747b97-a287-4df6-8b35-f30121c8c0b9\") " pod="openstack/barbican-keystone-listener-664bd5db46-xt8tr" Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.985764 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0ae0553d-4ec8-425b-bdf4-b7992b7cbda6-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-p9vh6\" (UID: \"0ae0553d-4ec8-425b-bdf4-b7992b7cbda6\") " pod="openstack/dnsmasq-dns-85ff748b95-p9vh6" Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.985808 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cd747b97-a287-4df6-8b35-f30121c8c0b9-config-data-custom\") pod \"barbican-keystone-listener-664bd5db46-xt8tr\" (UID: \"cd747b97-a287-4df6-8b35-f30121c8c0b9\") " pod="openstack/barbican-keystone-listener-664bd5db46-xt8tr" Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.985888 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0ae0553d-4ec8-425b-bdf4-b7992b7cbda6-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-p9vh6\" (UID: \"0ae0553d-4ec8-425b-bdf4-b7992b7cbda6\") " pod="openstack/dnsmasq-dns-85ff748b95-p9vh6" Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.986501 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/33482b7a-6aaa-48c4-9a46-c3de05cacebe-logs\") pod \"barbican-worker-6967b7c86c-47gl7\" (UID: \"33482b7a-6aaa-48c4-9a46-c3de05cacebe\") " pod="openstack/barbican-worker-6967b7c86c-47gl7" Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.993156 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/33482b7a-6aaa-48c4-9a46-c3de05cacebe-config-data-custom\") pod \"barbican-worker-6967b7c86c-47gl7\" (UID: \"33482b7a-6aaa-48c4-9a46-c3de05cacebe\") " pod="openstack/barbican-worker-6967b7c86c-47gl7" Dec 11 02:22:20 crc kubenswrapper[4824]: I1211 02:22:20.998949 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33482b7a-6aaa-48c4-9a46-c3de05cacebe-config-data\") pod \"barbican-worker-6967b7c86c-47gl7\" (UID: \"33482b7a-6aaa-48c4-9a46-c3de05cacebe\") " pod="openstack/barbican-worker-6967b7c86c-47gl7" Dec 11 02:22:21 crc kubenswrapper[4824]: I1211 02:22:21.013388 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ggvxt\" (UniqueName: \"kubernetes.io/projected/33482b7a-6aaa-48c4-9a46-c3de05cacebe-kube-api-access-ggvxt\") pod \"barbican-worker-6967b7c86c-47gl7\" (UID: \"33482b7a-6aaa-48c4-9a46-c3de05cacebe\") " pod="openstack/barbican-worker-6967b7c86c-47gl7" Dec 11 02:22:21 crc kubenswrapper[4824]: I1211 02:22:21.014095 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33482b7a-6aaa-48c4-9a46-c3de05cacebe-combined-ca-bundle\") pod \"barbican-worker-6967b7c86c-47gl7\" (UID: \"33482b7a-6aaa-48c4-9a46-c3de05cacebe\") " pod="openstack/barbican-worker-6967b7c86c-47gl7" Dec 11 02:22:21 crc kubenswrapper[4824]: I1211 02:22:21.087282 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd747b97-a287-4df6-8b35-f30121c8c0b9-logs\") pod \"barbican-keystone-listener-664bd5db46-xt8tr\" (UID: \"cd747b97-a287-4df6-8b35-f30121c8c0b9\") " pod="openstack/barbican-keystone-listener-664bd5db46-xt8tr" Dec 11 02:22:21 crc kubenswrapper[4824]: I1211 02:22:21.087332 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0ae0553d-4ec8-425b-bdf4-b7992b7cbda6-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-p9vh6\" (UID: \"0ae0553d-4ec8-425b-bdf4-b7992b7cbda6\") " pod="openstack/dnsmasq-dns-85ff748b95-p9vh6" Dec 11 02:22:21 crc kubenswrapper[4824]: I1211 02:22:21.087357 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cd747b97-a287-4df6-8b35-f30121c8c0b9-config-data-custom\") pod \"barbican-keystone-listener-664bd5db46-xt8tr\" (UID: \"cd747b97-a287-4df6-8b35-f30121c8c0b9\") " pod="openstack/barbican-keystone-listener-664bd5db46-xt8tr" Dec 11 02:22:21 crc kubenswrapper[4824]: I1211 02:22:21.087394 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0ae0553d-4ec8-425b-bdf4-b7992b7cbda6-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-p9vh6\" (UID: \"0ae0553d-4ec8-425b-bdf4-b7992b7cbda6\") " pod="openstack/dnsmasq-dns-85ff748b95-p9vh6" Dec 11 02:22:21 crc kubenswrapper[4824]: I1211 02:22:21.087449 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rqw5q\" (UniqueName: \"kubernetes.io/projected/cd747b97-a287-4df6-8b35-f30121c8c0b9-kube-api-access-rqw5q\") pod \"barbican-keystone-listener-664bd5db46-xt8tr\" (UID: \"cd747b97-a287-4df6-8b35-f30121c8c0b9\") " pod="openstack/barbican-keystone-listener-664bd5db46-xt8tr" Dec 11 02:22:21 crc kubenswrapper[4824]: I1211 02:22:21.087470 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0ae0553d-4ec8-425b-bdf4-b7992b7cbda6-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-p9vh6\" (UID: \"0ae0553d-4ec8-425b-bdf4-b7992b7cbda6\") " pod="openstack/dnsmasq-dns-85ff748b95-p9vh6" Dec 11 02:22:21 crc kubenswrapper[4824]: I1211 02:22:21.087499 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd747b97-a287-4df6-8b35-f30121c8c0b9-config-data\") pod \"barbican-keystone-listener-664bd5db46-xt8tr\" (UID: \"cd747b97-a287-4df6-8b35-f30121c8c0b9\") " pod="openstack/barbican-keystone-listener-664bd5db46-xt8tr" Dec 11 02:22:21 crc kubenswrapper[4824]: I1211 02:22:21.087522 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0ae0553d-4ec8-425b-bdf4-b7992b7cbda6-dns-svc\") pod \"dnsmasq-dns-85ff748b95-p9vh6\" (UID: \"0ae0553d-4ec8-425b-bdf4-b7992b7cbda6\") " pod="openstack/dnsmasq-dns-85ff748b95-p9vh6" Dec 11 02:22:21 crc kubenswrapper[4824]: I1211 02:22:21.087546 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd747b97-a287-4df6-8b35-f30121c8c0b9-combined-ca-bundle\") pod \"barbican-keystone-listener-664bd5db46-xt8tr\" (UID: \"cd747b97-a287-4df6-8b35-f30121c8c0b9\") " pod="openstack/barbican-keystone-listener-664bd5db46-xt8tr" Dec 11 02:22:21 crc kubenswrapper[4824]: I1211 02:22:21.087572 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0ae0553d-4ec8-425b-bdf4-b7992b7cbda6-config\") pod \"dnsmasq-dns-85ff748b95-p9vh6\" (UID: \"0ae0553d-4ec8-425b-bdf4-b7992b7cbda6\") " pod="openstack/dnsmasq-dns-85ff748b95-p9vh6" Dec 11 02:22:21 crc kubenswrapper[4824]: I1211 02:22:21.087593 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxr9p\" (UniqueName: \"kubernetes.io/projected/0ae0553d-4ec8-425b-bdf4-b7992b7cbda6-kube-api-access-gxr9p\") pod \"dnsmasq-dns-85ff748b95-p9vh6\" (UID: \"0ae0553d-4ec8-425b-bdf4-b7992b7cbda6\") " pod="openstack/dnsmasq-dns-85ff748b95-p9vh6" Dec 11 02:22:21 crc kubenswrapper[4824]: I1211 02:22:21.088330 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd747b97-a287-4df6-8b35-f30121c8c0b9-logs\") pod \"barbican-keystone-listener-664bd5db46-xt8tr\" (UID: \"cd747b97-a287-4df6-8b35-f30121c8c0b9\") " pod="openstack/barbican-keystone-listener-664bd5db46-xt8tr" Dec 11 02:22:21 crc kubenswrapper[4824]: I1211 02:22:21.088651 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0ae0553d-4ec8-425b-bdf4-b7992b7cbda6-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-p9vh6\" (UID: \"0ae0553d-4ec8-425b-bdf4-b7992b7cbda6\") " pod="openstack/dnsmasq-dns-85ff748b95-p9vh6" Dec 11 02:22:21 crc kubenswrapper[4824]: I1211 02:22:21.089000 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0ae0553d-4ec8-425b-bdf4-b7992b7cbda6-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-p9vh6\" (UID: \"0ae0553d-4ec8-425b-bdf4-b7992b7cbda6\") " pod="openstack/dnsmasq-dns-85ff748b95-p9vh6" Dec 11 02:22:21 crc kubenswrapper[4824]: I1211 02:22:21.098673 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0ae0553d-4ec8-425b-bdf4-b7992b7cbda6-dns-svc\") pod \"dnsmasq-dns-85ff748b95-p9vh6\" (UID: \"0ae0553d-4ec8-425b-bdf4-b7992b7cbda6\") " pod="openstack/dnsmasq-dns-85ff748b95-p9vh6" Dec 11 02:22:21 crc kubenswrapper[4824]: I1211 02:22:21.101267 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cd747b97-a287-4df6-8b35-f30121c8c0b9-config-data-custom\") pod \"barbican-keystone-listener-664bd5db46-xt8tr\" (UID: \"cd747b97-a287-4df6-8b35-f30121c8c0b9\") " pod="openstack/barbican-keystone-listener-664bd5db46-xt8tr" Dec 11 02:22:21 crc kubenswrapper[4824]: I1211 02:22:21.109681 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd747b97-a287-4df6-8b35-f30121c8c0b9-combined-ca-bundle\") pod \"barbican-keystone-listener-664bd5db46-xt8tr\" (UID: \"cd747b97-a287-4df6-8b35-f30121c8c0b9\") " pod="openstack/barbican-keystone-listener-664bd5db46-xt8tr" Dec 11 02:22:21 crc kubenswrapper[4824]: I1211 02:22:21.111673 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0ae0553d-4ec8-425b-bdf4-b7992b7cbda6-config\") pod \"dnsmasq-dns-85ff748b95-p9vh6\" (UID: \"0ae0553d-4ec8-425b-bdf4-b7992b7cbda6\") " pod="openstack/dnsmasq-dns-85ff748b95-p9vh6" Dec 11 02:22:21 crc kubenswrapper[4824]: I1211 02:22:21.120180 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-7c49d4759d-9h88v"] Dec 11 02:22:21 crc kubenswrapper[4824]: I1211 02:22:21.122076 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7c49d4759d-9h88v" Dec 11 02:22:21 crc kubenswrapper[4824]: I1211 02:22:21.123771 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0ae0553d-4ec8-425b-bdf4-b7992b7cbda6-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-p9vh6\" (UID: \"0ae0553d-4ec8-425b-bdf4-b7992b7cbda6\") " pod="openstack/dnsmasq-dns-85ff748b95-p9vh6" Dec 11 02:22:21 crc kubenswrapper[4824]: I1211 02:22:21.124154 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd747b97-a287-4df6-8b35-f30121c8c0b9-config-data\") pod \"barbican-keystone-listener-664bd5db46-xt8tr\" (UID: \"cd747b97-a287-4df6-8b35-f30121c8c0b9\") " pod="openstack/barbican-keystone-listener-664bd5db46-xt8tr" Dec 11 02:22:21 crc kubenswrapper[4824]: I1211 02:22:21.124308 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7c49d4759d-9h88v"] Dec 11 02:22:21 crc kubenswrapper[4824]: I1211 02:22:21.125759 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rqw5q\" (UniqueName: \"kubernetes.io/projected/cd747b97-a287-4df6-8b35-f30121c8c0b9-kube-api-access-rqw5q\") pod \"barbican-keystone-listener-664bd5db46-xt8tr\" (UID: \"cd747b97-a287-4df6-8b35-f30121c8c0b9\") " pod="openstack/barbican-keystone-listener-664bd5db46-xt8tr" Dec 11 02:22:21 crc kubenswrapper[4824]: I1211 02:22:21.131060 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxr9p\" (UniqueName: \"kubernetes.io/projected/0ae0553d-4ec8-425b-bdf4-b7992b7cbda6-kube-api-access-gxr9p\") pod \"dnsmasq-dns-85ff748b95-p9vh6\" (UID: \"0ae0553d-4ec8-425b-bdf4-b7992b7cbda6\") " pod="openstack/dnsmasq-dns-85ff748b95-p9vh6" Dec 11 02:22:21 crc kubenswrapper[4824]: I1211 02:22:21.131186 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Dec 11 02:22:21 crc kubenswrapper[4824]: I1211 02:22:21.174782 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-6967b7c86c-47gl7" Dec 11 02:22:21 crc kubenswrapper[4824]: I1211 02:22:21.290458 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-664bd5db46-xt8tr" Dec 11 02:22:21 crc kubenswrapper[4824]: I1211 02:22:21.310590 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f07eb1ab-ee38-41b4-9079-1ce271bdd3e9-combined-ca-bundle\") pod \"barbican-api-7c49d4759d-9h88v\" (UID: \"f07eb1ab-ee38-41b4-9079-1ce271bdd3e9\") " pod="openstack/barbican-api-7c49d4759d-9h88v" Dec 11 02:22:21 crc kubenswrapper[4824]: I1211 02:22:21.310647 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vcprh\" (UniqueName: \"kubernetes.io/projected/f07eb1ab-ee38-41b4-9079-1ce271bdd3e9-kube-api-access-vcprh\") pod \"barbican-api-7c49d4759d-9h88v\" (UID: \"f07eb1ab-ee38-41b4-9079-1ce271bdd3e9\") " pod="openstack/barbican-api-7c49d4759d-9h88v" Dec 11 02:22:21 crc kubenswrapper[4824]: I1211 02:22:21.310698 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f07eb1ab-ee38-41b4-9079-1ce271bdd3e9-config-data-custom\") pod \"barbican-api-7c49d4759d-9h88v\" (UID: \"f07eb1ab-ee38-41b4-9079-1ce271bdd3e9\") " pod="openstack/barbican-api-7c49d4759d-9h88v" Dec 11 02:22:21 crc kubenswrapper[4824]: I1211 02:22:21.310715 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f07eb1ab-ee38-41b4-9079-1ce271bdd3e9-logs\") pod \"barbican-api-7c49d4759d-9h88v\" (UID: \"f07eb1ab-ee38-41b4-9079-1ce271bdd3e9\") " pod="openstack/barbican-api-7c49d4759d-9h88v" Dec 11 02:22:21 crc kubenswrapper[4824]: I1211 02:22:21.310758 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f07eb1ab-ee38-41b4-9079-1ce271bdd3e9-config-data\") pod \"barbican-api-7c49d4759d-9h88v\" (UID: \"f07eb1ab-ee38-41b4-9079-1ce271bdd3e9\") " pod="openstack/barbican-api-7c49d4759d-9h88v" Dec 11 02:22:21 crc kubenswrapper[4824]: I1211 02:22:21.380502 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-p9vh6" Dec 11 02:22:21 crc kubenswrapper[4824]: I1211 02:22:21.412646 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vcprh\" (UniqueName: \"kubernetes.io/projected/f07eb1ab-ee38-41b4-9079-1ce271bdd3e9-kube-api-access-vcprh\") pod \"barbican-api-7c49d4759d-9h88v\" (UID: \"f07eb1ab-ee38-41b4-9079-1ce271bdd3e9\") " pod="openstack/barbican-api-7c49d4759d-9h88v" Dec 11 02:22:21 crc kubenswrapper[4824]: I1211 02:22:21.412734 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f07eb1ab-ee38-41b4-9079-1ce271bdd3e9-config-data-custom\") pod \"barbican-api-7c49d4759d-9h88v\" (UID: \"f07eb1ab-ee38-41b4-9079-1ce271bdd3e9\") " pod="openstack/barbican-api-7c49d4759d-9h88v" Dec 11 02:22:21 crc kubenswrapper[4824]: I1211 02:22:21.412763 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f07eb1ab-ee38-41b4-9079-1ce271bdd3e9-logs\") pod \"barbican-api-7c49d4759d-9h88v\" (UID: \"f07eb1ab-ee38-41b4-9079-1ce271bdd3e9\") " pod="openstack/barbican-api-7c49d4759d-9h88v" Dec 11 02:22:21 crc kubenswrapper[4824]: I1211 02:22:21.412812 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f07eb1ab-ee38-41b4-9079-1ce271bdd3e9-config-data\") pod \"barbican-api-7c49d4759d-9h88v\" (UID: \"f07eb1ab-ee38-41b4-9079-1ce271bdd3e9\") " pod="openstack/barbican-api-7c49d4759d-9h88v" Dec 11 02:22:21 crc kubenswrapper[4824]: I1211 02:22:21.412874 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f07eb1ab-ee38-41b4-9079-1ce271bdd3e9-combined-ca-bundle\") pod \"barbican-api-7c49d4759d-9h88v\" (UID: \"f07eb1ab-ee38-41b4-9079-1ce271bdd3e9\") " pod="openstack/barbican-api-7c49d4759d-9h88v" Dec 11 02:22:21 crc kubenswrapper[4824]: I1211 02:22:21.413469 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f07eb1ab-ee38-41b4-9079-1ce271bdd3e9-logs\") pod \"barbican-api-7c49d4759d-9h88v\" (UID: \"f07eb1ab-ee38-41b4-9079-1ce271bdd3e9\") " pod="openstack/barbican-api-7c49d4759d-9h88v" Dec 11 02:22:21 crc kubenswrapper[4824]: I1211 02:22:21.420818 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f07eb1ab-ee38-41b4-9079-1ce271bdd3e9-combined-ca-bundle\") pod \"barbican-api-7c49d4759d-9h88v\" (UID: \"f07eb1ab-ee38-41b4-9079-1ce271bdd3e9\") " pod="openstack/barbican-api-7c49d4759d-9h88v" Dec 11 02:22:21 crc kubenswrapper[4824]: I1211 02:22:21.421587 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f07eb1ab-ee38-41b4-9079-1ce271bdd3e9-config-data\") pod \"barbican-api-7c49d4759d-9h88v\" (UID: \"f07eb1ab-ee38-41b4-9079-1ce271bdd3e9\") " pod="openstack/barbican-api-7c49d4759d-9h88v" Dec 11 02:22:21 crc kubenswrapper[4824]: I1211 02:22:21.429631 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f07eb1ab-ee38-41b4-9079-1ce271bdd3e9-config-data-custom\") pod \"barbican-api-7c49d4759d-9h88v\" (UID: \"f07eb1ab-ee38-41b4-9079-1ce271bdd3e9\") " pod="openstack/barbican-api-7c49d4759d-9h88v" Dec 11 02:22:21 crc kubenswrapper[4824]: I1211 02:22:21.431921 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vcprh\" (UniqueName: \"kubernetes.io/projected/f07eb1ab-ee38-41b4-9079-1ce271bdd3e9-kube-api-access-vcprh\") pod \"barbican-api-7c49d4759d-9h88v\" (UID: \"f07eb1ab-ee38-41b4-9079-1ce271bdd3e9\") " pod="openstack/barbican-api-7c49d4759d-9h88v" Dec 11 02:22:21 crc kubenswrapper[4824]: I1211 02:22:21.532049 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7c49d4759d-9h88v" Dec 11 02:22:21 crc kubenswrapper[4824]: I1211 02:22:21.641517 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 11 02:22:21 crc kubenswrapper[4824]: I1211 02:22:21.679767 4824 generic.go:334] "Generic (PLEG): container finished" podID="8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7" containerID="6b97e27f717fafb1660abc5454cbf78410b26ec165a238a422af3cc98a9cf41b" exitCode=0 Dec 11 02:22:21 crc kubenswrapper[4824]: I1211 02:22:21.680124 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-tn658" event={"ID":"8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7","Type":"ContainerDied","Data":"6b97e27f717fafb1660abc5454cbf78410b26ec165a238a422af3cc98a9cf41b"} Dec 11 02:22:21 crc kubenswrapper[4824]: I1211 02:22:21.685528 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-6967b7c86c-47gl7"] Dec 11 02:22:21 crc kubenswrapper[4824]: I1211 02:22:21.760993 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-664bd5db46-xt8tr"] Dec 11 02:22:21 crc kubenswrapper[4824]: W1211 02:22:21.767311 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcd747b97_a287_4df6_8b35_f30121c8c0b9.slice/crio-0d0fe29a89b0f2637570126908d4bb2626b58341f28df864218a287de730b2f3 WatchSource:0}: Error finding container 0d0fe29a89b0f2637570126908d4bb2626b58341f28df864218a287de730b2f3: Status 404 returned error can't find the container with id 0d0fe29a89b0f2637570126908d4bb2626b58341f28df864218a287de730b2f3 Dec 11 02:22:21 crc kubenswrapper[4824]: I1211 02:22:21.884632 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-p9vh6"] Dec 11 02:22:21 crc kubenswrapper[4824]: W1211 02:22:21.887527 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0ae0553d_4ec8_425b_bdf4_b7992b7cbda6.slice/crio-ddd7469743b9d8c5ea57db2c7c21542c66f174a6162db1d7ef78e056bc0b908b WatchSource:0}: Error finding container ddd7469743b9d8c5ea57db2c7c21542c66f174a6162db1d7ef78e056bc0b908b: Status 404 returned error can't find the container with id ddd7469743b9d8c5ea57db2c7c21542c66f174a6162db1d7ef78e056bc0b908b Dec 11 02:22:22 crc kubenswrapper[4824]: I1211 02:22:22.007991 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 11 02:22:22 crc kubenswrapper[4824]: I1211 02:22:22.008030 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 11 02:22:22 crc kubenswrapper[4824]: I1211 02:22:22.044039 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7c49d4759d-9h88v"] Dec 11 02:22:22 crc kubenswrapper[4824]: W1211 02:22:22.071734 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf07eb1ab_ee38_41b4_9079_1ce271bdd3e9.slice/crio-833ba186d8b42a0a51aaaa7a41831d8dd2640e3b5e0c4b7ba3b281197bf01212 WatchSource:0}: Error finding container 833ba186d8b42a0a51aaaa7a41831d8dd2640e3b5e0c4b7ba3b281197bf01212: Status 404 returned error can't find the container with id 833ba186d8b42a0a51aaaa7a41831d8dd2640e3b5e0c4b7ba3b281197bf01212 Dec 11 02:22:22 crc kubenswrapper[4824]: I1211 02:22:22.074218 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 11 02:22:22 crc kubenswrapper[4824]: I1211 02:22:22.086098 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 11 02:22:22 crc kubenswrapper[4824]: I1211 02:22:22.656310 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a36c6797-0e71-4dd1-b2ae-fb540f63e794" path="/var/lib/kubelet/pods/a36c6797-0e71-4dd1-b2ae-fb540f63e794/volumes" Dec 11 02:22:22 crc kubenswrapper[4824]: I1211 02:22:22.720633 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7c49d4759d-9h88v" event={"ID":"f07eb1ab-ee38-41b4-9079-1ce271bdd3e9","Type":"ContainerStarted","Data":"7f36efde65203331d341ec6eb5de62d50e5c1ddbae4723e622401cc845678a61"} Dec 11 02:22:22 crc kubenswrapper[4824]: I1211 02:22:22.720865 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7c49d4759d-9h88v" Dec 11 02:22:22 crc kubenswrapper[4824]: I1211 02:22:22.720882 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7c49d4759d-9h88v" Dec 11 02:22:22 crc kubenswrapper[4824]: I1211 02:22:22.720890 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7c49d4759d-9h88v" event={"ID":"f07eb1ab-ee38-41b4-9079-1ce271bdd3e9","Type":"ContainerStarted","Data":"814eac34539186694805a58af437492cdbd9fe8b4d535c481361454f1f4495c1"} Dec 11 02:22:22 crc kubenswrapper[4824]: I1211 02:22:22.720899 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7c49d4759d-9h88v" event={"ID":"f07eb1ab-ee38-41b4-9079-1ce271bdd3e9","Type":"ContainerStarted","Data":"833ba186d8b42a0a51aaaa7a41831d8dd2640e3b5e0c4b7ba3b281197bf01212"} Dec 11 02:22:22 crc kubenswrapper[4824]: I1211 02:22:22.722283 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-6967b7c86c-47gl7" event={"ID":"33482b7a-6aaa-48c4-9a46-c3de05cacebe","Type":"ContainerStarted","Data":"2dc8101d9674abdd24c05de873f7688ec41b9d4d3f53f5e093d45fe541e994c8"} Dec 11 02:22:22 crc kubenswrapper[4824]: I1211 02:22:22.724780 4824 generic.go:334] "Generic (PLEG): container finished" podID="0ae0553d-4ec8-425b-bdf4-b7992b7cbda6" containerID="f4363872c145810a21a2ffbe5e07af05ac5f320e0444647370eef2363c0baf7b" exitCode=0 Dec 11 02:22:22 crc kubenswrapper[4824]: I1211 02:22:22.724957 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-p9vh6" event={"ID":"0ae0553d-4ec8-425b-bdf4-b7992b7cbda6","Type":"ContainerDied","Data":"f4363872c145810a21a2ffbe5e07af05ac5f320e0444647370eef2363c0baf7b"} Dec 11 02:22:22 crc kubenswrapper[4824]: I1211 02:22:22.724987 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-p9vh6" event={"ID":"0ae0553d-4ec8-425b-bdf4-b7992b7cbda6","Type":"ContainerStarted","Data":"ddd7469743b9d8c5ea57db2c7c21542c66f174a6162db1d7ef78e056bc0b908b"} Dec 11 02:22:22 crc kubenswrapper[4824]: I1211 02:22:22.729123 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-664bd5db46-xt8tr" event={"ID":"cd747b97-a287-4df6-8b35-f30121c8c0b9","Type":"ContainerStarted","Data":"0d0fe29a89b0f2637570126908d4bb2626b58341f28df864218a287de730b2f3"} Dec 11 02:22:22 crc kubenswrapper[4824]: I1211 02:22:22.729152 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 11 02:22:22 crc kubenswrapper[4824]: I1211 02:22:22.729257 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 11 02:22:22 crc kubenswrapper[4824]: I1211 02:22:22.750528 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-7c49d4759d-9h88v" podStartSLOduration=1.7505114960000001 podStartE2EDuration="1.750511496s" podCreationTimestamp="2025-12-11 02:22:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:22:22.737303654 +0000 UTC m=+1284.426341043" watchObservedRunningTime="2025-12-11 02:22:22.750511496 +0000 UTC m=+1284.439548875" Dec 11 02:22:23 crc kubenswrapper[4824]: I1211 02:22:23.543215 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-tn658" Dec 11 02:22:23 crc kubenswrapper[4824]: I1211 02:22:23.653720 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vgswv\" (UniqueName: \"kubernetes.io/projected/8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7-kube-api-access-vgswv\") pod \"8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7\" (UID: \"8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7\") " Dec 11 02:22:23 crc kubenswrapper[4824]: I1211 02:22:23.653794 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7-combined-ca-bundle\") pod \"8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7\" (UID: \"8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7\") " Dec 11 02:22:23 crc kubenswrapper[4824]: I1211 02:22:23.653885 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7-config-data\") pod \"8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7\" (UID: \"8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7\") " Dec 11 02:22:23 crc kubenswrapper[4824]: I1211 02:22:23.654013 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7-scripts\") pod \"8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7\" (UID: \"8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7\") " Dec 11 02:22:23 crc kubenswrapper[4824]: I1211 02:22:23.654068 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7-db-sync-config-data\") pod \"8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7\" (UID: \"8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7\") " Dec 11 02:22:23 crc kubenswrapper[4824]: I1211 02:22:23.654098 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7-etc-machine-id\") pod \"8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7\" (UID: \"8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7\") " Dec 11 02:22:23 crc kubenswrapper[4824]: I1211 02:22:23.654761 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7" (UID: "8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 02:22:23 crc kubenswrapper[4824]: I1211 02:22:23.658762 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7-kube-api-access-vgswv" (OuterVolumeSpecName: "kube-api-access-vgswv") pod "8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7" (UID: "8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7"). InnerVolumeSpecName "kube-api-access-vgswv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:22:23 crc kubenswrapper[4824]: I1211 02:22:23.660557 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7-scripts" (OuterVolumeSpecName: "scripts") pod "8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7" (UID: "8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:22:23 crc kubenswrapper[4824]: I1211 02:22:23.660977 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7" (UID: "8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:22:23 crc kubenswrapper[4824]: I1211 02:22:23.695331 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7" (UID: "8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:22:23 crc kubenswrapper[4824]: I1211 02:22:23.699687 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7-config-data" (OuterVolumeSpecName: "config-data") pod "8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7" (UID: "8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:22:23 crc kubenswrapper[4824]: I1211 02:22:23.717673 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-c88ccd6d6-bg6lk"] Dec 11 02:22:23 crc kubenswrapper[4824]: E1211 02:22:23.718129 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7" containerName="cinder-db-sync" Dec 11 02:22:23 crc kubenswrapper[4824]: I1211 02:22:23.718149 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7" containerName="cinder-db-sync" Dec 11 02:22:23 crc kubenswrapper[4824]: I1211 02:22:23.718388 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7" containerName="cinder-db-sync" Dec 11 02:22:23 crc kubenswrapper[4824]: I1211 02:22:23.719555 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-c88ccd6d6-bg6lk" Dec 11 02:22:23 crc kubenswrapper[4824]: I1211 02:22:23.722782 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Dec 11 02:22:23 crc kubenswrapper[4824]: I1211 02:22:23.722851 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Dec 11 02:22:23 crc kubenswrapper[4824]: I1211 02:22:23.744006 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-tn658" Dec 11 02:22:23 crc kubenswrapper[4824]: I1211 02:22:23.744055 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-tn658" event={"ID":"8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7","Type":"ContainerDied","Data":"d84bd6f07a5f4326184457e98968c257a95a27bc22c12f9d7c4a5db2ad7953a7"} Dec 11 02:22:23 crc kubenswrapper[4824]: I1211 02:22:23.744079 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d84bd6f07a5f4326184457e98968c257a95a27bc22c12f9d7c4a5db2ad7953a7" Dec 11 02:22:23 crc kubenswrapper[4824]: I1211 02:22:23.764184 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vgswv\" (UniqueName: \"kubernetes.io/projected/8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7-kube-api-access-vgswv\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:23 crc kubenswrapper[4824]: I1211 02:22:23.764241 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:23 crc kubenswrapper[4824]: I1211 02:22:23.764252 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:23 crc kubenswrapper[4824]: I1211 02:22:23.764275 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:23 crc kubenswrapper[4824]: I1211 02:22:23.764285 4824 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:23 crc kubenswrapper[4824]: I1211 02:22:23.764294 4824 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:23 crc kubenswrapper[4824]: I1211 02:22:23.803293 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-c88ccd6d6-bg6lk"] Dec 11 02:22:23 crc kubenswrapper[4824]: I1211 02:22:23.878695 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bd669057-4834-42ed-83ea-f7a454c0a013-logs\") pod \"barbican-api-c88ccd6d6-bg6lk\" (UID: \"bd669057-4834-42ed-83ea-f7a454c0a013\") " pod="openstack/barbican-api-c88ccd6d6-bg6lk" Dec 11 02:22:23 crc kubenswrapper[4824]: I1211 02:22:23.878896 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b9trl\" (UniqueName: \"kubernetes.io/projected/bd669057-4834-42ed-83ea-f7a454c0a013-kube-api-access-b9trl\") pod \"barbican-api-c88ccd6d6-bg6lk\" (UID: \"bd669057-4834-42ed-83ea-f7a454c0a013\") " pod="openstack/barbican-api-c88ccd6d6-bg6lk" Dec 11 02:22:23 crc kubenswrapper[4824]: I1211 02:22:23.878927 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd669057-4834-42ed-83ea-f7a454c0a013-config-data\") pod \"barbican-api-c88ccd6d6-bg6lk\" (UID: \"bd669057-4834-42ed-83ea-f7a454c0a013\") " pod="openstack/barbican-api-c88ccd6d6-bg6lk" Dec 11 02:22:23 crc kubenswrapper[4824]: I1211 02:22:23.878985 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bd669057-4834-42ed-83ea-f7a454c0a013-config-data-custom\") pod \"barbican-api-c88ccd6d6-bg6lk\" (UID: \"bd669057-4834-42ed-83ea-f7a454c0a013\") " pod="openstack/barbican-api-c88ccd6d6-bg6lk" Dec 11 02:22:23 crc kubenswrapper[4824]: I1211 02:22:23.879009 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd669057-4834-42ed-83ea-f7a454c0a013-internal-tls-certs\") pod \"barbican-api-c88ccd6d6-bg6lk\" (UID: \"bd669057-4834-42ed-83ea-f7a454c0a013\") " pod="openstack/barbican-api-c88ccd6d6-bg6lk" Dec 11 02:22:23 crc kubenswrapper[4824]: I1211 02:22:23.879127 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd669057-4834-42ed-83ea-f7a454c0a013-combined-ca-bundle\") pod \"barbican-api-c88ccd6d6-bg6lk\" (UID: \"bd669057-4834-42ed-83ea-f7a454c0a013\") " pod="openstack/barbican-api-c88ccd6d6-bg6lk" Dec 11 02:22:23 crc kubenswrapper[4824]: I1211 02:22:23.879181 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd669057-4834-42ed-83ea-f7a454c0a013-public-tls-certs\") pod \"barbican-api-c88ccd6d6-bg6lk\" (UID: \"bd669057-4834-42ed-83ea-f7a454c0a013\") " pod="openstack/barbican-api-c88ccd6d6-bg6lk" Dec 11 02:22:23 crc kubenswrapper[4824]: I1211 02:22:23.929801 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 11 02:22:23 crc kubenswrapper[4824]: I1211 02:22:23.931281 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 11 02:22:23 crc kubenswrapper[4824]: I1211 02:22:23.937199 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-xsjrn" Dec 11 02:22:23 crc kubenswrapper[4824]: I1211 02:22:23.937345 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 11 02:22:23 crc kubenswrapper[4824]: I1211 02:22:23.937446 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 11 02:22:23 crc kubenswrapper[4824]: I1211 02:22:23.937682 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 11 02:22:23 crc kubenswrapper[4824]: I1211 02:22:23.937784 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 11 02:22:23 crc kubenswrapper[4824]: I1211 02:22:23.981947 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd669057-4834-42ed-83ea-f7a454c0a013-combined-ca-bundle\") pod \"barbican-api-c88ccd6d6-bg6lk\" (UID: \"bd669057-4834-42ed-83ea-f7a454c0a013\") " pod="openstack/barbican-api-c88ccd6d6-bg6lk" Dec 11 02:22:23 crc kubenswrapper[4824]: I1211 02:22:23.982002 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd669057-4834-42ed-83ea-f7a454c0a013-public-tls-certs\") pod \"barbican-api-c88ccd6d6-bg6lk\" (UID: \"bd669057-4834-42ed-83ea-f7a454c0a013\") " pod="openstack/barbican-api-c88ccd6d6-bg6lk" Dec 11 02:22:23 crc kubenswrapper[4824]: I1211 02:22:23.982044 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bd669057-4834-42ed-83ea-f7a454c0a013-logs\") pod \"barbican-api-c88ccd6d6-bg6lk\" (UID: \"bd669057-4834-42ed-83ea-f7a454c0a013\") " pod="openstack/barbican-api-c88ccd6d6-bg6lk" Dec 11 02:22:23 crc kubenswrapper[4824]: I1211 02:22:23.982101 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b9trl\" (UniqueName: \"kubernetes.io/projected/bd669057-4834-42ed-83ea-f7a454c0a013-kube-api-access-b9trl\") pod \"barbican-api-c88ccd6d6-bg6lk\" (UID: \"bd669057-4834-42ed-83ea-f7a454c0a013\") " pod="openstack/barbican-api-c88ccd6d6-bg6lk" Dec 11 02:22:23 crc kubenswrapper[4824]: I1211 02:22:23.982141 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd669057-4834-42ed-83ea-f7a454c0a013-config-data\") pod \"barbican-api-c88ccd6d6-bg6lk\" (UID: \"bd669057-4834-42ed-83ea-f7a454c0a013\") " pod="openstack/barbican-api-c88ccd6d6-bg6lk" Dec 11 02:22:23 crc kubenswrapper[4824]: I1211 02:22:23.982814 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bd669057-4834-42ed-83ea-f7a454c0a013-config-data-custom\") pod \"barbican-api-c88ccd6d6-bg6lk\" (UID: \"bd669057-4834-42ed-83ea-f7a454c0a013\") " pod="openstack/barbican-api-c88ccd6d6-bg6lk" Dec 11 02:22:23 crc kubenswrapper[4824]: I1211 02:22:23.982859 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd669057-4834-42ed-83ea-f7a454c0a013-internal-tls-certs\") pod \"barbican-api-c88ccd6d6-bg6lk\" (UID: \"bd669057-4834-42ed-83ea-f7a454c0a013\") " pod="openstack/barbican-api-c88ccd6d6-bg6lk" Dec 11 02:22:23 crc kubenswrapper[4824]: I1211 02:22:23.989252 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd669057-4834-42ed-83ea-f7a454c0a013-internal-tls-certs\") pod \"barbican-api-c88ccd6d6-bg6lk\" (UID: \"bd669057-4834-42ed-83ea-f7a454c0a013\") " pod="openstack/barbican-api-c88ccd6d6-bg6lk" Dec 11 02:22:23 crc kubenswrapper[4824]: I1211 02:22:23.989579 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bd669057-4834-42ed-83ea-f7a454c0a013-logs\") pod \"barbican-api-c88ccd6d6-bg6lk\" (UID: \"bd669057-4834-42ed-83ea-f7a454c0a013\") " pod="openstack/barbican-api-c88ccd6d6-bg6lk" Dec 11 02:22:23 crc kubenswrapper[4824]: I1211 02:22:23.989930 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd669057-4834-42ed-83ea-f7a454c0a013-public-tls-certs\") pod \"barbican-api-c88ccd6d6-bg6lk\" (UID: \"bd669057-4834-42ed-83ea-f7a454c0a013\") " pod="openstack/barbican-api-c88ccd6d6-bg6lk" Dec 11 02:22:24 crc kubenswrapper[4824]: I1211 02:22:23.993477 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-p9vh6"] Dec 11 02:22:24 crc kubenswrapper[4824]: I1211 02:22:24.001102 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd669057-4834-42ed-83ea-f7a454c0a013-combined-ca-bundle\") pod \"barbican-api-c88ccd6d6-bg6lk\" (UID: \"bd669057-4834-42ed-83ea-f7a454c0a013\") " pod="openstack/barbican-api-c88ccd6d6-bg6lk" Dec 11 02:22:24 crc kubenswrapper[4824]: I1211 02:22:24.003027 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bd669057-4834-42ed-83ea-f7a454c0a013-config-data-custom\") pod \"barbican-api-c88ccd6d6-bg6lk\" (UID: \"bd669057-4834-42ed-83ea-f7a454c0a013\") " pod="openstack/barbican-api-c88ccd6d6-bg6lk" Dec 11 02:22:24 crc kubenswrapper[4824]: I1211 02:22:24.013205 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd669057-4834-42ed-83ea-f7a454c0a013-config-data\") pod \"barbican-api-c88ccd6d6-bg6lk\" (UID: \"bd669057-4834-42ed-83ea-f7a454c0a013\") " pod="openstack/barbican-api-c88ccd6d6-bg6lk" Dec 11 02:22:24 crc kubenswrapper[4824]: I1211 02:22:24.019782 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b9trl\" (UniqueName: \"kubernetes.io/projected/bd669057-4834-42ed-83ea-f7a454c0a013-kube-api-access-b9trl\") pod \"barbican-api-c88ccd6d6-bg6lk\" (UID: \"bd669057-4834-42ed-83ea-f7a454c0a013\") " pod="openstack/barbican-api-c88ccd6d6-bg6lk" Dec 11 02:22:24 crc kubenswrapper[4824]: I1211 02:22:24.044485 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-rng95"] Dec 11 02:22:24 crc kubenswrapper[4824]: I1211 02:22:24.058901 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-rng95" Dec 11 02:22:24 crc kubenswrapper[4824]: I1211 02:22:24.084659 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/33fd5103-fcb0-404a-a730-4fdbbbc3577f-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"33fd5103-fcb0-404a-a730-4fdbbbc3577f\") " pod="openstack/cinder-scheduler-0" Dec 11 02:22:24 crc kubenswrapper[4824]: I1211 02:22:24.084740 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33fd5103-fcb0-404a-a730-4fdbbbc3577f-scripts\") pod \"cinder-scheduler-0\" (UID: \"33fd5103-fcb0-404a-a730-4fdbbbc3577f\") " pod="openstack/cinder-scheduler-0" Dec 11 02:22:24 crc kubenswrapper[4824]: I1211 02:22:24.084764 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/33fd5103-fcb0-404a-a730-4fdbbbc3577f-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"33fd5103-fcb0-404a-a730-4fdbbbc3577f\") " pod="openstack/cinder-scheduler-0" Dec 11 02:22:24 crc kubenswrapper[4824]: I1211 02:22:24.084782 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33fd5103-fcb0-404a-a730-4fdbbbc3577f-config-data\") pod \"cinder-scheduler-0\" (UID: \"33fd5103-fcb0-404a-a730-4fdbbbc3577f\") " pod="openstack/cinder-scheduler-0" Dec 11 02:22:24 crc kubenswrapper[4824]: I1211 02:22:24.084802 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c9tls\" (UniqueName: \"kubernetes.io/projected/33fd5103-fcb0-404a-a730-4fdbbbc3577f-kube-api-access-c9tls\") pod \"cinder-scheduler-0\" (UID: \"33fd5103-fcb0-404a-a730-4fdbbbc3577f\") " pod="openstack/cinder-scheduler-0" Dec 11 02:22:24 crc kubenswrapper[4824]: I1211 02:22:24.084884 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33fd5103-fcb0-404a-a730-4fdbbbc3577f-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"33fd5103-fcb0-404a-a730-4fdbbbc3577f\") " pod="openstack/cinder-scheduler-0" Dec 11 02:22:24 crc kubenswrapper[4824]: I1211 02:22:24.093355 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-rng95"] Dec 11 02:22:24 crc kubenswrapper[4824]: I1211 02:22:24.106012 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-c88ccd6d6-bg6lk" Dec 11 02:22:24 crc kubenswrapper[4824]: I1211 02:22:24.130326 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 11 02:22:24 crc kubenswrapper[4824]: I1211 02:22:24.131858 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 11 02:22:24 crc kubenswrapper[4824]: I1211 02:22:24.134358 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 11 02:22:24 crc kubenswrapper[4824]: I1211 02:22:24.143356 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 11 02:22:24 crc kubenswrapper[4824]: I1211 02:22:24.186410 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33fd5103-fcb0-404a-a730-4fdbbbc3577f-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"33fd5103-fcb0-404a-a730-4fdbbbc3577f\") " pod="openstack/cinder-scheduler-0" Dec 11 02:22:24 crc kubenswrapper[4824]: I1211 02:22:24.186457 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ae139cc3-a38e-45fa-a3b4-e507ad5a99d6-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-rng95\" (UID: \"ae139cc3-a38e-45fa-a3b4-e507ad5a99d6\") " pod="openstack/dnsmasq-dns-5c9776ccc5-rng95" Dec 11 02:22:24 crc kubenswrapper[4824]: I1211 02:22:24.186483 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/33fd5103-fcb0-404a-a730-4fdbbbc3577f-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"33fd5103-fcb0-404a-a730-4fdbbbc3577f\") " pod="openstack/cinder-scheduler-0" Dec 11 02:22:24 crc kubenswrapper[4824]: I1211 02:22:24.186520 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae139cc3-a38e-45fa-a3b4-e507ad5a99d6-config\") pod \"dnsmasq-dns-5c9776ccc5-rng95\" (UID: \"ae139cc3-a38e-45fa-a3b4-e507ad5a99d6\") " pod="openstack/dnsmasq-dns-5c9776ccc5-rng95" Dec 11 02:22:24 crc kubenswrapper[4824]: I1211 02:22:24.186540 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ae139cc3-a38e-45fa-a3b4-e507ad5a99d6-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-rng95\" (UID: \"ae139cc3-a38e-45fa-a3b4-e507ad5a99d6\") " pod="openstack/dnsmasq-dns-5c9776ccc5-rng95" Dec 11 02:22:24 crc kubenswrapper[4824]: I1211 02:22:24.186570 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33fd5103-fcb0-404a-a730-4fdbbbc3577f-scripts\") pod \"cinder-scheduler-0\" (UID: \"33fd5103-fcb0-404a-a730-4fdbbbc3577f\") " pod="openstack/cinder-scheduler-0" Dec 11 02:22:24 crc kubenswrapper[4824]: I1211 02:22:24.186591 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/33fd5103-fcb0-404a-a730-4fdbbbc3577f-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"33fd5103-fcb0-404a-a730-4fdbbbc3577f\") " pod="openstack/cinder-scheduler-0" Dec 11 02:22:24 crc kubenswrapper[4824]: I1211 02:22:24.186611 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33fd5103-fcb0-404a-a730-4fdbbbc3577f-config-data\") pod \"cinder-scheduler-0\" (UID: \"33fd5103-fcb0-404a-a730-4fdbbbc3577f\") " pod="openstack/cinder-scheduler-0" Dec 11 02:22:24 crc kubenswrapper[4824]: I1211 02:22:24.186630 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c9tls\" (UniqueName: \"kubernetes.io/projected/33fd5103-fcb0-404a-a730-4fdbbbc3577f-kube-api-access-c9tls\") pod \"cinder-scheduler-0\" (UID: \"33fd5103-fcb0-404a-a730-4fdbbbc3577f\") " pod="openstack/cinder-scheduler-0" Dec 11 02:22:24 crc kubenswrapper[4824]: I1211 02:22:24.186657 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ae139cc3-a38e-45fa-a3b4-e507ad5a99d6-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-rng95\" (UID: \"ae139cc3-a38e-45fa-a3b4-e507ad5a99d6\") " pod="openstack/dnsmasq-dns-5c9776ccc5-rng95" Dec 11 02:22:24 crc kubenswrapper[4824]: I1211 02:22:24.186682 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jh6td\" (UniqueName: \"kubernetes.io/projected/ae139cc3-a38e-45fa-a3b4-e507ad5a99d6-kube-api-access-jh6td\") pod \"dnsmasq-dns-5c9776ccc5-rng95\" (UID: \"ae139cc3-a38e-45fa-a3b4-e507ad5a99d6\") " pod="openstack/dnsmasq-dns-5c9776ccc5-rng95" Dec 11 02:22:24 crc kubenswrapper[4824]: I1211 02:22:24.186719 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ae139cc3-a38e-45fa-a3b4-e507ad5a99d6-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-rng95\" (UID: \"ae139cc3-a38e-45fa-a3b4-e507ad5a99d6\") " pod="openstack/dnsmasq-dns-5c9776ccc5-rng95" Dec 11 02:22:24 crc kubenswrapper[4824]: I1211 02:22:24.187535 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/33fd5103-fcb0-404a-a730-4fdbbbc3577f-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"33fd5103-fcb0-404a-a730-4fdbbbc3577f\") " pod="openstack/cinder-scheduler-0" Dec 11 02:22:24 crc kubenswrapper[4824]: I1211 02:22:24.190588 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/33fd5103-fcb0-404a-a730-4fdbbbc3577f-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"33fd5103-fcb0-404a-a730-4fdbbbc3577f\") " pod="openstack/cinder-scheduler-0" Dec 11 02:22:24 crc kubenswrapper[4824]: I1211 02:22:24.191723 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33fd5103-fcb0-404a-a730-4fdbbbc3577f-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"33fd5103-fcb0-404a-a730-4fdbbbc3577f\") " pod="openstack/cinder-scheduler-0" Dec 11 02:22:24 crc kubenswrapper[4824]: I1211 02:22:24.192662 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33fd5103-fcb0-404a-a730-4fdbbbc3577f-config-data\") pod \"cinder-scheduler-0\" (UID: \"33fd5103-fcb0-404a-a730-4fdbbbc3577f\") " pod="openstack/cinder-scheduler-0" Dec 11 02:22:24 crc kubenswrapper[4824]: I1211 02:22:24.207797 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33fd5103-fcb0-404a-a730-4fdbbbc3577f-scripts\") pod \"cinder-scheduler-0\" (UID: \"33fd5103-fcb0-404a-a730-4fdbbbc3577f\") " pod="openstack/cinder-scheduler-0" Dec 11 02:22:24 crc kubenswrapper[4824]: I1211 02:22:24.210409 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c9tls\" (UniqueName: \"kubernetes.io/projected/33fd5103-fcb0-404a-a730-4fdbbbc3577f-kube-api-access-c9tls\") pod \"cinder-scheduler-0\" (UID: \"33fd5103-fcb0-404a-a730-4fdbbbc3577f\") " pod="openstack/cinder-scheduler-0" Dec 11 02:22:24 crc kubenswrapper[4824]: I1211 02:22:24.249466 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 11 02:22:24 crc kubenswrapper[4824]: I1211 02:22:24.288298 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/970c513a-2a0b-4a15-b051-e7215330c566-etc-machine-id\") pod \"cinder-api-0\" (UID: \"970c513a-2a0b-4a15-b051-e7215330c566\") " pod="openstack/cinder-api-0" Dec 11 02:22:24 crc kubenswrapper[4824]: I1211 02:22:24.288624 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/970c513a-2a0b-4a15-b051-e7215330c566-config-data-custom\") pod \"cinder-api-0\" (UID: \"970c513a-2a0b-4a15-b051-e7215330c566\") " pod="openstack/cinder-api-0" Dec 11 02:22:24 crc kubenswrapper[4824]: I1211 02:22:24.288691 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ae139cc3-a38e-45fa-a3b4-e507ad5a99d6-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-rng95\" (UID: \"ae139cc3-a38e-45fa-a3b4-e507ad5a99d6\") " pod="openstack/dnsmasq-dns-5c9776ccc5-rng95" Dec 11 02:22:24 crc kubenswrapper[4824]: I1211 02:22:24.288712 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/970c513a-2a0b-4a15-b051-e7215330c566-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"970c513a-2a0b-4a15-b051-e7215330c566\") " pod="openstack/cinder-api-0" Dec 11 02:22:24 crc kubenswrapper[4824]: I1211 02:22:24.288756 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jh6td\" (UniqueName: \"kubernetes.io/projected/ae139cc3-a38e-45fa-a3b4-e507ad5a99d6-kube-api-access-jh6td\") pod \"dnsmasq-dns-5c9776ccc5-rng95\" (UID: \"ae139cc3-a38e-45fa-a3b4-e507ad5a99d6\") " pod="openstack/dnsmasq-dns-5c9776ccc5-rng95" Dec 11 02:22:24 crc kubenswrapper[4824]: I1211 02:22:24.288771 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/970c513a-2a0b-4a15-b051-e7215330c566-logs\") pod \"cinder-api-0\" (UID: \"970c513a-2a0b-4a15-b051-e7215330c566\") " pod="openstack/cinder-api-0" Dec 11 02:22:24 crc kubenswrapper[4824]: I1211 02:22:24.288874 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ae139cc3-a38e-45fa-a3b4-e507ad5a99d6-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-rng95\" (UID: \"ae139cc3-a38e-45fa-a3b4-e507ad5a99d6\") " pod="openstack/dnsmasq-dns-5c9776ccc5-rng95" Dec 11 02:22:24 crc kubenswrapper[4824]: I1211 02:22:24.289027 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ae139cc3-a38e-45fa-a3b4-e507ad5a99d6-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-rng95\" (UID: \"ae139cc3-a38e-45fa-a3b4-e507ad5a99d6\") " pod="openstack/dnsmasq-dns-5c9776ccc5-rng95" Dec 11 02:22:24 crc kubenswrapper[4824]: I1211 02:22:24.289320 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hxrnh\" (UniqueName: \"kubernetes.io/projected/970c513a-2a0b-4a15-b051-e7215330c566-kube-api-access-hxrnh\") pod \"cinder-api-0\" (UID: \"970c513a-2a0b-4a15-b051-e7215330c566\") " pod="openstack/cinder-api-0" Dec 11 02:22:24 crc kubenswrapper[4824]: I1211 02:22:24.290252 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae139cc3-a38e-45fa-a3b4-e507ad5a99d6-config\") pod \"dnsmasq-dns-5c9776ccc5-rng95\" (UID: \"ae139cc3-a38e-45fa-a3b4-e507ad5a99d6\") " pod="openstack/dnsmasq-dns-5c9776ccc5-rng95" Dec 11 02:22:24 crc kubenswrapper[4824]: I1211 02:22:24.291034 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ae139cc3-a38e-45fa-a3b4-e507ad5a99d6-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-rng95\" (UID: \"ae139cc3-a38e-45fa-a3b4-e507ad5a99d6\") " pod="openstack/dnsmasq-dns-5c9776ccc5-rng95" Dec 11 02:22:24 crc kubenswrapper[4824]: I1211 02:22:24.291192 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/970c513a-2a0b-4a15-b051-e7215330c566-scripts\") pod \"cinder-api-0\" (UID: \"970c513a-2a0b-4a15-b051-e7215330c566\") " pod="openstack/cinder-api-0" Dec 11 02:22:24 crc kubenswrapper[4824]: I1211 02:22:24.291323 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/970c513a-2a0b-4a15-b051-e7215330c566-config-data\") pod \"cinder-api-0\" (UID: \"970c513a-2a0b-4a15-b051-e7215330c566\") " pod="openstack/cinder-api-0" Dec 11 02:22:24 crc kubenswrapper[4824]: I1211 02:22:24.290986 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae139cc3-a38e-45fa-a3b4-e507ad5a99d6-config\") pod \"dnsmasq-dns-5c9776ccc5-rng95\" (UID: \"ae139cc3-a38e-45fa-a3b4-e507ad5a99d6\") " pod="openstack/dnsmasq-dns-5c9776ccc5-rng95" Dec 11 02:22:24 crc kubenswrapper[4824]: I1211 02:22:24.290195 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ae139cc3-a38e-45fa-a3b4-e507ad5a99d6-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-rng95\" (UID: \"ae139cc3-a38e-45fa-a3b4-e507ad5a99d6\") " pod="openstack/dnsmasq-dns-5c9776ccc5-rng95" Dec 11 02:22:24 crc kubenswrapper[4824]: I1211 02:22:24.290037 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ae139cc3-a38e-45fa-a3b4-e507ad5a99d6-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-rng95\" (UID: \"ae139cc3-a38e-45fa-a3b4-e507ad5a99d6\") " pod="openstack/dnsmasq-dns-5c9776ccc5-rng95" Dec 11 02:22:24 crc kubenswrapper[4824]: I1211 02:22:24.289866 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ae139cc3-a38e-45fa-a3b4-e507ad5a99d6-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-rng95\" (UID: \"ae139cc3-a38e-45fa-a3b4-e507ad5a99d6\") " pod="openstack/dnsmasq-dns-5c9776ccc5-rng95" Dec 11 02:22:24 crc kubenswrapper[4824]: I1211 02:22:24.292041 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ae139cc3-a38e-45fa-a3b4-e507ad5a99d6-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-rng95\" (UID: \"ae139cc3-a38e-45fa-a3b4-e507ad5a99d6\") " pod="openstack/dnsmasq-dns-5c9776ccc5-rng95" Dec 11 02:22:24 crc kubenswrapper[4824]: I1211 02:22:24.304264 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jh6td\" (UniqueName: \"kubernetes.io/projected/ae139cc3-a38e-45fa-a3b4-e507ad5a99d6-kube-api-access-jh6td\") pod \"dnsmasq-dns-5c9776ccc5-rng95\" (UID: \"ae139cc3-a38e-45fa-a3b4-e507ad5a99d6\") " pod="openstack/dnsmasq-dns-5c9776ccc5-rng95" Dec 11 02:22:24 crc kubenswrapper[4824]: I1211 02:22:24.393604 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hxrnh\" (UniqueName: \"kubernetes.io/projected/970c513a-2a0b-4a15-b051-e7215330c566-kube-api-access-hxrnh\") pod \"cinder-api-0\" (UID: \"970c513a-2a0b-4a15-b051-e7215330c566\") " pod="openstack/cinder-api-0" Dec 11 02:22:24 crc kubenswrapper[4824]: I1211 02:22:24.393762 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/970c513a-2a0b-4a15-b051-e7215330c566-scripts\") pod \"cinder-api-0\" (UID: \"970c513a-2a0b-4a15-b051-e7215330c566\") " pod="openstack/cinder-api-0" Dec 11 02:22:24 crc kubenswrapper[4824]: I1211 02:22:24.393792 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/970c513a-2a0b-4a15-b051-e7215330c566-config-data\") pod \"cinder-api-0\" (UID: \"970c513a-2a0b-4a15-b051-e7215330c566\") " pod="openstack/cinder-api-0" Dec 11 02:22:24 crc kubenswrapper[4824]: I1211 02:22:24.394411 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/970c513a-2a0b-4a15-b051-e7215330c566-etc-machine-id\") pod \"cinder-api-0\" (UID: \"970c513a-2a0b-4a15-b051-e7215330c566\") " pod="openstack/cinder-api-0" Dec 11 02:22:24 crc kubenswrapper[4824]: I1211 02:22:24.394465 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/970c513a-2a0b-4a15-b051-e7215330c566-config-data-custom\") pod \"cinder-api-0\" (UID: \"970c513a-2a0b-4a15-b051-e7215330c566\") " pod="openstack/cinder-api-0" Dec 11 02:22:24 crc kubenswrapper[4824]: I1211 02:22:24.394490 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/970c513a-2a0b-4a15-b051-e7215330c566-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"970c513a-2a0b-4a15-b051-e7215330c566\") " pod="openstack/cinder-api-0" Dec 11 02:22:24 crc kubenswrapper[4824]: I1211 02:22:24.394530 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/970c513a-2a0b-4a15-b051-e7215330c566-logs\") pod \"cinder-api-0\" (UID: \"970c513a-2a0b-4a15-b051-e7215330c566\") " pod="openstack/cinder-api-0" Dec 11 02:22:24 crc kubenswrapper[4824]: I1211 02:22:24.394963 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/970c513a-2a0b-4a15-b051-e7215330c566-etc-machine-id\") pod \"cinder-api-0\" (UID: \"970c513a-2a0b-4a15-b051-e7215330c566\") " pod="openstack/cinder-api-0" Dec 11 02:22:24 crc kubenswrapper[4824]: I1211 02:22:24.395279 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/970c513a-2a0b-4a15-b051-e7215330c566-logs\") pod \"cinder-api-0\" (UID: \"970c513a-2a0b-4a15-b051-e7215330c566\") " pod="openstack/cinder-api-0" Dec 11 02:22:24 crc kubenswrapper[4824]: I1211 02:22:24.397267 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/970c513a-2a0b-4a15-b051-e7215330c566-scripts\") pod \"cinder-api-0\" (UID: \"970c513a-2a0b-4a15-b051-e7215330c566\") " pod="openstack/cinder-api-0" Dec 11 02:22:24 crc kubenswrapper[4824]: I1211 02:22:24.397497 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/970c513a-2a0b-4a15-b051-e7215330c566-config-data-custom\") pod \"cinder-api-0\" (UID: \"970c513a-2a0b-4a15-b051-e7215330c566\") " pod="openstack/cinder-api-0" Dec 11 02:22:24 crc kubenswrapper[4824]: I1211 02:22:24.397764 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/970c513a-2a0b-4a15-b051-e7215330c566-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"970c513a-2a0b-4a15-b051-e7215330c566\") " pod="openstack/cinder-api-0" Dec 11 02:22:24 crc kubenswrapper[4824]: I1211 02:22:24.406584 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-rng95" Dec 11 02:22:24 crc kubenswrapper[4824]: I1211 02:22:24.410863 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/970c513a-2a0b-4a15-b051-e7215330c566-config-data\") pod \"cinder-api-0\" (UID: \"970c513a-2a0b-4a15-b051-e7215330c566\") " pod="openstack/cinder-api-0" Dec 11 02:22:24 crc kubenswrapper[4824]: I1211 02:22:24.416811 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hxrnh\" (UniqueName: \"kubernetes.io/projected/970c513a-2a0b-4a15-b051-e7215330c566-kube-api-access-hxrnh\") pod \"cinder-api-0\" (UID: \"970c513a-2a0b-4a15-b051-e7215330c566\") " pod="openstack/cinder-api-0" Dec 11 02:22:24 crc kubenswrapper[4824]: I1211 02:22:24.463456 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 11 02:22:24 crc kubenswrapper[4824]: I1211 02:22:24.770925 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-p9vh6" event={"ID":"0ae0553d-4ec8-425b-bdf4-b7992b7cbda6","Type":"ContainerStarted","Data":"c92f390155c0b971260f48951537e72226a56e5616e220e88db811a0c76ee79c"} Dec 11 02:22:24 crc kubenswrapper[4824]: I1211 02:22:24.770968 4824 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 11 02:22:24 crc kubenswrapper[4824]: I1211 02:22:24.771601 4824 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 11 02:22:24 crc kubenswrapper[4824]: I1211 02:22:24.877559 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-c88ccd6d6-bg6lk"] Dec 11 02:22:24 crc kubenswrapper[4824]: I1211 02:22:24.990698 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 11 02:22:25 crc kubenswrapper[4824]: I1211 02:22:25.154694 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 11 02:22:25 crc kubenswrapper[4824]: W1211 02:22:25.173699 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod970c513a_2a0b_4a15_b051_e7215330c566.slice/crio-4ac679b0f378001bfa880a71aa44a4197801a057e35536f7edfbce56a1773694 WatchSource:0}: Error finding container 4ac679b0f378001bfa880a71aa44a4197801a057e35536f7edfbce56a1773694: Status 404 returned error can't find the container with id 4ac679b0f378001bfa880a71aa44a4197801a057e35536f7edfbce56a1773694 Dec 11 02:22:25 crc kubenswrapper[4824]: I1211 02:22:25.236985 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-rng95"] Dec 11 02:22:25 crc kubenswrapper[4824]: W1211 02:22:25.257233 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podae139cc3_a38e_45fa_a3b4_e507ad5a99d6.slice/crio-c7f896f81d1071e759f97e85536990b6e57d9b8c2ae4db5ea3a96cc5b4a8e81d WatchSource:0}: Error finding container c7f896f81d1071e759f97e85536990b6e57d9b8c2ae4db5ea3a96cc5b4a8e81d: Status 404 returned error can't find the container with id c7f896f81d1071e759f97e85536990b6e57d9b8c2ae4db5ea3a96cc5b4a8e81d Dec 11 02:22:25 crc kubenswrapper[4824]: I1211 02:22:25.292618 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 11 02:22:25 crc kubenswrapper[4824]: I1211 02:22:25.293520 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 11 02:22:25 crc kubenswrapper[4824]: I1211 02:22:25.782695 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"33fd5103-fcb0-404a-a730-4fdbbbc3577f","Type":"ContainerStarted","Data":"46eb982f9d23d1911e3f768cf97fb3328e848220db8b231b201b8fc7a2cc8fb1"} Dec 11 02:22:25 crc kubenswrapper[4824]: I1211 02:22:25.784154 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"970c513a-2a0b-4a15-b051-e7215330c566","Type":"ContainerStarted","Data":"4ac679b0f378001bfa880a71aa44a4197801a057e35536f7edfbce56a1773694"} Dec 11 02:22:25 crc kubenswrapper[4824]: I1211 02:22:25.785737 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-664bd5db46-xt8tr" event={"ID":"cd747b97-a287-4df6-8b35-f30121c8c0b9","Type":"ContainerStarted","Data":"2524a99c5486ee4dc3ce3f3318c052d06862abd33ff6defa48d976a464ad848b"} Dec 11 02:22:25 crc kubenswrapper[4824]: I1211 02:22:25.785791 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-664bd5db46-xt8tr" event={"ID":"cd747b97-a287-4df6-8b35-f30121c8c0b9","Type":"ContainerStarted","Data":"04380f4e535b100ae65cad3c9fd59dff06575d4e4bc7ba45ddd1da4ba16971f8"} Dec 11 02:22:25 crc kubenswrapper[4824]: I1211 02:22:25.787541 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-c88ccd6d6-bg6lk" event={"ID":"bd669057-4834-42ed-83ea-f7a454c0a013","Type":"ContainerStarted","Data":"ff08cf39429177e7cda4f55e7b1daee39e9c18b451d43708ff8608b35e9a2724"} Dec 11 02:22:25 crc kubenswrapper[4824]: I1211 02:22:25.787587 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-c88ccd6d6-bg6lk" event={"ID":"bd669057-4834-42ed-83ea-f7a454c0a013","Type":"ContainerStarted","Data":"b321d10a3446b9a8d20d7f426357a09ebe7e576b123eab4ca03bf0b748bc890e"} Dec 11 02:22:25 crc kubenswrapper[4824]: I1211 02:22:25.787600 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-c88ccd6d6-bg6lk" event={"ID":"bd669057-4834-42ed-83ea-f7a454c0a013","Type":"ContainerStarted","Data":"d11df6596b14de9d76cc6a3bccccea37913e2fdd040f133e468802881d25b134"} Dec 11 02:22:25 crc kubenswrapper[4824]: I1211 02:22:25.788381 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-c88ccd6d6-bg6lk" Dec 11 02:22:25 crc kubenswrapper[4824]: I1211 02:22:25.788408 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-c88ccd6d6-bg6lk" Dec 11 02:22:25 crc kubenswrapper[4824]: I1211 02:22:25.789787 4824 generic.go:334] "Generic (PLEG): container finished" podID="ae139cc3-a38e-45fa-a3b4-e507ad5a99d6" containerID="4f180d0a5f22f977c7b9cda23a893b756fc73677013a093c38a98c141a1fb5d8" exitCode=0 Dec 11 02:22:25 crc kubenswrapper[4824]: I1211 02:22:25.789829 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-rng95" event={"ID":"ae139cc3-a38e-45fa-a3b4-e507ad5a99d6","Type":"ContainerDied","Data":"4f180d0a5f22f977c7b9cda23a893b756fc73677013a093c38a98c141a1fb5d8"} Dec 11 02:22:25 crc kubenswrapper[4824]: I1211 02:22:25.789846 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-rng95" event={"ID":"ae139cc3-a38e-45fa-a3b4-e507ad5a99d6","Type":"ContainerStarted","Data":"c7f896f81d1071e759f97e85536990b6e57d9b8c2ae4db5ea3a96cc5b4a8e81d"} Dec 11 02:22:25 crc kubenswrapper[4824]: I1211 02:22:25.802955 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-85ff748b95-p9vh6" podUID="0ae0553d-4ec8-425b-bdf4-b7992b7cbda6" containerName="dnsmasq-dns" containerID="cri-o://c92f390155c0b971260f48951537e72226a56e5616e220e88db811a0c76ee79c" gracePeriod=10 Dec 11 02:22:25 crc kubenswrapper[4824]: I1211 02:22:25.803777 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-6967b7c86c-47gl7" event={"ID":"33482b7a-6aaa-48c4-9a46-c3de05cacebe","Type":"ContainerStarted","Data":"7d45d262f5ce8a0f70436675c9ce72c30d418d6df08736cef2b4e7c892c74ea4"} Dec 11 02:22:25 crc kubenswrapper[4824]: I1211 02:22:25.803826 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-6967b7c86c-47gl7" event={"ID":"33482b7a-6aaa-48c4-9a46-c3de05cacebe","Type":"ContainerStarted","Data":"ded861ea0928e67868983d5b42caa0a47d3566fa28436a8cbbf73108545d67fa"} Dec 11 02:22:25 crc kubenswrapper[4824]: I1211 02:22:25.803844 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-85ff748b95-p9vh6" Dec 11 02:22:25 crc kubenswrapper[4824]: I1211 02:22:25.809343 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-664bd5db46-xt8tr" podStartSLOduration=3.076383123 podStartE2EDuration="5.809279733s" podCreationTimestamp="2025-12-11 02:22:20 +0000 UTC" firstStartedPulling="2025-12-11 02:22:21.769685737 +0000 UTC m=+1283.458723116" lastFinishedPulling="2025-12-11 02:22:24.502582347 +0000 UTC m=+1286.191619726" observedRunningTime="2025-12-11 02:22:25.803163886 +0000 UTC m=+1287.492201265" watchObservedRunningTime="2025-12-11 02:22:25.809279733 +0000 UTC m=+1287.498317112" Dec 11 02:22:25 crc kubenswrapper[4824]: I1211 02:22:25.844504 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-6967b7c86c-47gl7" podStartSLOduration=3.038483405 podStartE2EDuration="5.844486029s" podCreationTimestamp="2025-12-11 02:22:20 +0000 UTC" firstStartedPulling="2025-12-11 02:22:21.698019022 +0000 UTC m=+1283.387056391" lastFinishedPulling="2025-12-11 02:22:24.504021636 +0000 UTC m=+1286.193059015" observedRunningTime="2025-12-11 02:22:25.830726242 +0000 UTC m=+1287.519763621" watchObservedRunningTime="2025-12-11 02:22:25.844486029 +0000 UTC m=+1287.533523408" Dec 11 02:22:25 crc kubenswrapper[4824]: I1211 02:22:25.880628 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-85ff748b95-p9vh6" podStartSLOduration=5.88059661 podStartE2EDuration="5.88059661s" podCreationTimestamp="2025-12-11 02:22:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:22:25.856620272 +0000 UTC m=+1287.545657651" watchObservedRunningTime="2025-12-11 02:22:25.88059661 +0000 UTC m=+1287.569633989" Dec 11 02:22:25 crc kubenswrapper[4824]: I1211 02:22:25.887480 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-c88ccd6d6-bg6lk" podStartSLOduration=2.887466098 podStartE2EDuration="2.887466098s" podCreationTimestamp="2025-12-11 02:22:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:22:25.873478334 +0000 UTC m=+1287.562515713" watchObservedRunningTime="2025-12-11 02:22:25.887466098 +0000 UTC m=+1287.576503477" Dec 11 02:22:26 crc kubenswrapper[4824]: I1211 02:22:26.003105 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-74cc6c88d8-zvlk7" Dec 11 02:22:26 crc kubenswrapper[4824]: I1211 02:22:26.276094 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-787649dcdd-ts6qs" Dec 11 02:22:26 crc kubenswrapper[4824]: I1211 02:22:26.654089 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-p9vh6" Dec 11 02:22:26 crc kubenswrapper[4824]: I1211 02:22:26.766597 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0ae0553d-4ec8-425b-bdf4-b7992b7cbda6-dns-swift-storage-0\") pod \"0ae0553d-4ec8-425b-bdf4-b7992b7cbda6\" (UID: \"0ae0553d-4ec8-425b-bdf4-b7992b7cbda6\") " Dec 11 02:22:26 crc kubenswrapper[4824]: I1211 02:22:26.766720 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0ae0553d-4ec8-425b-bdf4-b7992b7cbda6-ovsdbserver-nb\") pod \"0ae0553d-4ec8-425b-bdf4-b7992b7cbda6\" (UID: \"0ae0553d-4ec8-425b-bdf4-b7992b7cbda6\") " Dec 11 02:22:26 crc kubenswrapper[4824]: I1211 02:22:26.766763 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0ae0553d-4ec8-425b-bdf4-b7992b7cbda6-dns-svc\") pod \"0ae0553d-4ec8-425b-bdf4-b7992b7cbda6\" (UID: \"0ae0553d-4ec8-425b-bdf4-b7992b7cbda6\") " Dec 11 02:22:26 crc kubenswrapper[4824]: I1211 02:22:26.766810 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0ae0553d-4ec8-425b-bdf4-b7992b7cbda6-config\") pod \"0ae0553d-4ec8-425b-bdf4-b7992b7cbda6\" (UID: \"0ae0553d-4ec8-425b-bdf4-b7992b7cbda6\") " Dec 11 02:22:26 crc kubenswrapper[4824]: I1211 02:22:26.766831 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gxr9p\" (UniqueName: \"kubernetes.io/projected/0ae0553d-4ec8-425b-bdf4-b7992b7cbda6-kube-api-access-gxr9p\") pod \"0ae0553d-4ec8-425b-bdf4-b7992b7cbda6\" (UID: \"0ae0553d-4ec8-425b-bdf4-b7992b7cbda6\") " Dec 11 02:22:26 crc kubenswrapper[4824]: I1211 02:22:26.766896 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0ae0553d-4ec8-425b-bdf4-b7992b7cbda6-ovsdbserver-sb\") pod \"0ae0553d-4ec8-425b-bdf4-b7992b7cbda6\" (UID: \"0ae0553d-4ec8-425b-bdf4-b7992b7cbda6\") " Dec 11 02:22:26 crc kubenswrapper[4824]: I1211 02:22:26.771400 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ae0553d-4ec8-425b-bdf4-b7992b7cbda6-kube-api-access-gxr9p" (OuterVolumeSpecName: "kube-api-access-gxr9p") pod "0ae0553d-4ec8-425b-bdf4-b7992b7cbda6" (UID: "0ae0553d-4ec8-425b-bdf4-b7992b7cbda6"). InnerVolumeSpecName "kube-api-access-gxr9p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:22:26 crc kubenswrapper[4824]: I1211 02:22:26.840349 4824 generic.go:334] "Generic (PLEG): container finished" podID="0ae0553d-4ec8-425b-bdf4-b7992b7cbda6" containerID="c92f390155c0b971260f48951537e72226a56e5616e220e88db811a0c76ee79c" exitCode=0 Dec 11 02:22:26 crc kubenswrapper[4824]: I1211 02:22:26.840735 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-p9vh6" event={"ID":"0ae0553d-4ec8-425b-bdf4-b7992b7cbda6","Type":"ContainerDied","Data":"c92f390155c0b971260f48951537e72226a56e5616e220e88db811a0c76ee79c"} Dec 11 02:22:26 crc kubenswrapper[4824]: I1211 02:22:26.840761 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-p9vh6" event={"ID":"0ae0553d-4ec8-425b-bdf4-b7992b7cbda6","Type":"ContainerDied","Data":"ddd7469743b9d8c5ea57db2c7c21542c66f174a6162db1d7ef78e056bc0b908b"} Dec 11 02:22:26 crc kubenswrapper[4824]: I1211 02:22:26.840777 4824 scope.go:117] "RemoveContainer" containerID="c92f390155c0b971260f48951537e72226a56e5616e220e88db811a0c76ee79c" Dec 11 02:22:26 crc kubenswrapper[4824]: I1211 02:22:26.840890 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-p9vh6" Dec 11 02:22:26 crc kubenswrapper[4824]: I1211 02:22:26.848720 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"970c513a-2a0b-4a15-b051-e7215330c566","Type":"ContainerStarted","Data":"2e16c5c6832b346a8f74ddd8a7659b5e7a58a48e4a87eb5bea7656164467cd5f"} Dec 11 02:22:26 crc kubenswrapper[4824]: I1211 02:22:26.869428 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gxr9p\" (UniqueName: \"kubernetes.io/projected/0ae0553d-4ec8-425b-bdf4-b7992b7cbda6-kube-api-access-gxr9p\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:26 crc kubenswrapper[4824]: I1211 02:22:26.873181 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-rng95" event={"ID":"ae139cc3-a38e-45fa-a3b4-e507ad5a99d6","Type":"ContainerStarted","Data":"60188aa532b0f0fa9043b6f15ce868702ccd2495595cff67f37a859aa64a0100"} Dec 11 02:22:26 crc kubenswrapper[4824]: I1211 02:22:26.873302 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c9776ccc5-rng95" Dec 11 02:22:26 crc kubenswrapper[4824]: I1211 02:22:26.894641 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c9776ccc5-rng95" podStartSLOduration=2.894616139 podStartE2EDuration="2.894616139s" podCreationTimestamp="2025-12-11 02:22:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:22:26.880583674 +0000 UTC m=+1288.569621053" watchObservedRunningTime="2025-12-11 02:22:26.894616139 +0000 UTC m=+1288.583653518" Dec 11 02:22:26 crc kubenswrapper[4824]: I1211 02:22:26.895654 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0ae0553d-4ec8-425b-bdf4-b7992b7cbda6-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "0ae0553d-4ec8-425b-bdf4-b7992b7cbda6" (UID: "0ae0553d-4ec8-425b-bdf4-b7992b7cbda6"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:22:26 crc kubenswrapper[4824]: I1211 02:22:26.916429 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 11 02:22:26 crc kubenswrapper[4824]: I1211 02:22:26.919535 4824 scope.go:117] "RemoveContainer" containerID="f4363872c145810a21a2ffbe5e07af05ac5f320e0444647370eef2363c0baf7b" Dec 11 02:22:26 crc kubenswrapper[4824]: I1211 02:22:26.947577 4824 scope.go:117] "RemoveContainer" containerID="c92f390155c0b971260f48951537e72226a56e5616e220e88db811a0c76ee79c" Dec 11 02:22:26 crc kubenswrapper[4824]: E1211 02:22:26.948138 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c92f390155c0b971260f48951537e72226a56e5616e220e88db811a0c76ee79c\": container with ID starting with c92f390155c0b971260f48951537e72226a56e5616e220e88db811a0c76ee79c not found: ID does not exist" containerID="c92f390155c0b971260f48951537e72226a56e5616e220e88db811a0c76ee79c" Dec 11 02:22:26 crc kubenswrapper[4824]: I1211 02:22:26.948183 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c92f390155c0b971260f48951537e72226a56e5616e220e88db811a0c76ee79c"} err="failed to get container status \"c92f390155c0b971260f48951537e72226a56e5616e220e88db811a0c76ee79c\": rpc error: code = NotFound desc = could not find container \"c92f390155c0b971260f48951537e72226a56e5616e220e88db811a0c76ee79c\": container with ID starting with c92f390155c0b971260f48951537e72226a56e5616e220e88db811a0c76ee79c not found: ID does not exist" Dec 11 02:22:26 crc kubenswrapper[4824]: I1211 02:22:26.948209 4824 scope.go:117] "RemoveContainer" containerID="f4363872c145810a21a2ffbe5e07af05ac5f320e0444647370eef2363c0baf7b" Dec 11 02:22:26 crc kubenswrapper[4824]: E1211 02:22:26.948483 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f4363872c145810a21a2ffbe5e07af05ac5f320e0444647370eef2363c0baf7b\": container with ID starting with f4363872c145810a21a2ffbe5e07af05ac5f320e0444647370eef2363c0baf7b not found: ID does not exist" containerID="f4363872c145810a21a2ffbe5e07af05ac5f320e0444647370eef2363c0baf7b" Dec 11 02:22:26 crc kubenswrapper[4824]: I1211 02:22:26.948512 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4363872c145810a21a2ffbe5e07af05ac5f320e0444647370eef2363c0baf7b"} err="failed to get container status \"f4363872c145810a21a2ffbe5e07af05ac5f320e0444647370eef2363c0baf7b\": rpc error: code = NotFound desc = could not find container \"f4363872c145810a21a2ffbe5e07af05ac5f320e0444647370eef2363c0baf7b\": container with ID starting with f4363872c145810a21a2ffbe5e07af05ac5f320e0444647370eef2363c0baf7b not found: ID does not exist" Dec 11 02:22:26 crc kubenswrapper[4824]: I1211 02:22:26.971484 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0ae0553d-4ec8-425b-bdf4-b7992b7cbda6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0ae0553d-4ec8-425b-bdf4-b7992b7cbda6" (UID: "0ae0553d-4ec8-425b-bdf4-b7992b7cbda6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:22:26 crc kubenswrapper[4824]: I1211 02:22:26.973683 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0ae0553d-4ec8-425b-bdf4-b7992b7cbda6-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:26 crc kubenswrapper[4824]: I1211 02:22:26.973715 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0ae0553d-4ec8-425b-bdf4-b7992b7cbda6-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:26 crc kubenswrapper[4824]: I1211 02:22:26.983622 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0ae0553d-4ec8-425b-bdf4-b7992b7cbda6-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "0ae0553d-4ec8-425b-bdf4-b7992b7cbda6" (UID: "0ae0553d-4ec8-425b-bdf4-b7992b7cbda6"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:22:26 crc kubenswrapper[4824]: I1211 02:22:26.997898 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0ae0553d-4ec8-425b-bdf4-b7992b7cbda6-config" (OuterVolumeSpecName: "config") pod "0ae0553d-4ec8-425b-bdf4-b7992b7cbda6" (UID: "0ae0553d-4ec8-425b-bdf4-b7992b7cbda6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:22:26 crc kubenswrapper[4824]: I1211 02:22:26.998191 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0ae0553d-4ec8-425b-bdf4-b7992b7cbda6-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "0ae0553d-4ec8-425b-bdf4-b7992b7cbda6" (UID: "0ae0553d-4ec8-425b-bdf4-b7992b7cbda6"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:22:27 crc kubenswrapper[4824]: I1211 02:22:27.074672 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0ae0553d-4ec8-425b-bdf4-b7992b7cbda6-config\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:27 crc kubenswrapper[4824]: I1211 02:22:27.074693 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0ae0553d-4ec8-425b-bdf4-b7992b7cbda6-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:27 crc kubenswrapper[4824]: I1211 02:22:27.074702 4824 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0ae0553d-4ec8-425b-bdf4-b7992b7cbda6-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:27 crc kubenswrapper[4824]: I1211 02:22:27.178392 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-p9vh6"] Dec 11 02:22:27 crc kubenswrapper[4824]: I1211 02:22:27.186212 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-p9vh6"] Dec 11 02:22:27 crc kubenswrapper[4824]: I1211 02:22:27.885335 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"33fd5103-fcb0-404a-a730-4fdbbbc3577f","Type":"ContainerStarted","Data":"b1d7c28f9270b8f42148267da1914aede650a0637bfb612e2541d29aef379dc7"} Dec 11 02:22:27 crc kubenswrapper[4824]: I1211 02:22:27.890389 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-74cc6c88d8-zvlk7" Dec 11 02:22:27 crc kubenswrapper[4824]: I1211 02:22:27.896129 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"970c513a-2a0b-4a15-b051-e7215330c566","Type":"ContainerStarted","Data":"2dc35fa571dce36d41341fae6229aeeb8fbc31be620b625f5d0ff140df1c8dc6"} Dec 11 02:22:27 crc kubenswrapper[4824]: I1211 02:22:27.896578 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="970c513a-2a0b-4a15-b051-e7215330c566" containerName="cinder-api-log" containerID="cri-o://2e16c5c6832b346a8f74ddd8a7659b5e7a58a48e4a87eb5bea7656164467cd5f" gracePeriod=30 Dec 11 02:22:27 crc kubenswrapper[4824]: I1211 02:22:27.896716 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="970c513a-2a0b-4a15-b051-e7215330c566" containerName="cinder-api" containerID="cri-o://2dc35fa571dce36d41341fae6229aeeb8fbc31be620b625f5d0ff140df1c8dc6" gracePeriod=30 Dec 11 02:22:28 crc kubenswrapper[4824]: I1211 02:22:28.025191 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-787649dcdd-ts6qs"] Dec 11 02:22:28 crc kubenswrapper[4824]: I1211 02:22:28.025895 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-787649dcdd-ts6qs" podUID="f15f9ed0-98f2-42fe-8006-2142098cc2b6" containerName="horizon-log" containerID="cri-o://e4de9df2d8c6dc730220bfb406d58ac47fefcc07528a72cb76ac4ea026179a39" gracePeriod=30 Dec 11 02:22:28 crc kubenswrapper[4824]: I1211 02:22:28.026051 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-787649dcdd-ts6qs" podUID="f15f9ed0-98f2-42fe-8006-2142098cc2b6" containerName="horizon" containerID="cri-o://588925740a1286914b8db0f80b0f9426464ecc031e105965e01b1abcbf40a9ba" gracePeriod=30 Dec 11 02:22:28 crc kubenswrapper[4824]: I1211 02:22:28.035595 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-787649dcdd-ts6qs" podUID="f15f9ed0-98f2-42fe-8006-2142098cc2b6" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": EOF" Dec 11 02:22:28 crc kubenswrapper[4824]: I1211 02:22:28.039551 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.039532058 podStartE2EDuration="4.039532058s" podCreationTimestamp="2025-12-11 02:22:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:22:27.986901436 +0000 UTC m=+1289.675938835" watchObservedRunningTime="2025-12-11 02:22:28.039532058 +0000 UTC m=+1289.728569437" Dec 11 02:22:28 crc kubenswrapper[4824]: I1211 02:22:28.667496 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0ae0553d-4ec8-425b-bdf4-b7992b7cbda6" path="/var/lib/kubelet/pods/0ae0553d-4ec8-425b-bdf4-b7992b7cbda6/volumes" Dec 11 02:22:28 crc kubenswrapper[4824]: I1211 02:22:28.681086 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 11 02:22:28 crc kubenswrapper[4824]: I1211 02:22:28.743229 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/970c513a-2a0b-4a15-b051-e7215330c566-config-data-custom\") pod \"970c513a-2a0b-4a15-b051-e7215330c566\" (UID: \"970c513a-2a0b-4a15-b051-e7215330c566\") " Dec 11 02:22:28 crc kubenswrapper[4824]: I1211 02:22:28.743316 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hxrnh\" (UniqueName: \"kubernetes.io/projected/970c513a-2a0b-4a15-b051-e7215330c566-kube-api-access-hxrnh\") pod \"970c513a-2a0b-4a15-b051-e7215330c566\" (UID: \"970c513a-2a0b-4a15-b051-e7215330c566\") " Dec 11 02:22:28 crc kubenswrapper[4824]: I1211 02:22:28.743357 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/970c513a-2a0b-4a15-b051-e7215330c566-scripts\") pod \"970c513a-2a0b-4a15-b051-e7215330c566\" (UID: \"970c513a-2a0b-4a15-b051-e7215330c566\") " Dec 11 02:22:28 crc kubenswrapper[4824]: I1211 02:22:28.743382 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/970c513a-2a0b-4a15-b051-e7215330c566-logs\") pod \"970c513a-2a0b-4a15-b051-e7215330c566\" (UID: \"970c513a-2a0b-4a15-b051-e7215330c566\") " Dec 11 02:22:28 crc kubenswrapper[4824]: I1211 02:22:28.743485 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/970c513a-2a0b-4a15-b051-e7215330c566-combined-ca-bundle\") pod \"970c513a-2a0b-4a15-b051-e7215330c566\" (UID: \"970c513a-2a0b-4a15-b051-e7215330c566\") " Dec 11 02:22:28 crc kubenswrapper[4824]: I1211 02:22:28.743502 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/970c513a-2a0b-4a15-b051-e7215330c566-config-data\") pod \"970c513a-2a0b-4a15-b051-e7215330c566\" (UID: \"970c513a-2a0b-4a15-b051-e7215330c566\") " Dec 11 02:22:28 crc kubenswrapper[4824]: I1211 02:22:28.745081 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/970c513a-2a0b-4a15-b051-e7215330c566-logs" (OuterVolumeSpecName: "logs") pod "970c513a-2a0b-4a15-b051-e7215330c566" (UID: "970c513a-2a0b-4a15-b051-e7215330c566"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:22:28 crc kubenswrapper[4824]: I1211 02:22:28.745389 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/970c513a-2a0b-4a15-b051-e7215330c566-etc-machine-id\") pod \"970c513a-2a0b-4a15-b051-e7215330c566\" (UID: \"970c513a-2a0b-4a15-b051-e7215330c566\") " Dec 11 02:22:28 crc kubenswrapper[4824]: I1211 02:22:28.745668 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/970c513a-2a0b-4a15-b051-e7215330c566-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "970c513a-2a0b-4a15-b051-e7215330c566" (UID: "970c513a-2a0b-4a15-b051-e7215330c566"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 02:22:28 crc kubenswrapper[4824]: I1211 02:22:28.746487 4824 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/970c513a-2a0b-4a15-b051-e7215330c566-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:28 crc kubenswrapper[4824]: I1211 02:22:28.746505 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/970c513a-2a0b-4a15-b051-e7215330c566-logs\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:28 crc kubenswrapper[4824]: I1211 02:22:28.751622 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/970c513a-2a0b-4a15-b051-e7215330c566-kube-api-access-hxrnh" (OuterVolumeSpecName: "kube-api-access-hxrnh") pod "970c513a-2a0b-4a15-b051-e7215330c566" (UID: "970c513a-2a0b-4a15-b051-e7215330c566"). InnerVolumeSpecName "kube-api-access-hxrnh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:22:28 crc kubenswrapper[4824]: I1211 02:22:28.752307 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/970c513a-2a0b-4a15-b051-e7215330c566-scripts" (OuterVolumeSpecName: "scripts") pod "970c513a-2a0b-4a15-b051-e7215330c566" (UID: "970c513a-2a0b-4a15-b051-e7215330c566"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:22:28 crc kubenswrapper[4824]: I1211 02:22:28.752361 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/970c513a-2a0b-4a15-b051-e7215330c566-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "970c513a-2a0b-4a15-b051-e7215330c566" (UID: "970c513a-2a0b-4a15-b051-e7215330c566"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:22:28 crc kubenswrapper[4824]: I1211 02:22:28.812303 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/970c513a-2a0b-4a15-b051-e7215330c566-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "970c513a-2a0b-4a15-b051-e7215330c566" (UID: "970c513a-2a0b-4a15-b051-e7215330c566"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:22:28 crc kubenswrapper[4824]: I1211 02:22:28.822616 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/970c513a-2a0b-4a15-b051-e7215330c566-config-data" (OuterVolumeSpecName: "config-data") pod "970c513a-2a0b-4a15-b051-e7215330c566" (UID: "970c513a-2a0b-4a15-b051-e7215330c566"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:22:28 crc kubenswrapper[4824]: I1211 02:22:28.848138 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/970c513a-2a0b-4a15-b051-e7215330c566-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:28 crc kubenswrapper[4824]: I1211 02:22:28.848184 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/970c513a-2a0b-4a15-b051-e7215330c566-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:28 crc kubenswrapper[4824]: I1211 02:22:28.848195 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/970c513a-2a0b-4a15-b051-e7215330c566-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:28 crc kubenswrapper[4824]: I1211 02:22:28.848205 4824 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/970c513a-2a0b-4a15-b051-e7215330c566-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:28 crc kubenswrapper[4824]: I1211 02:22:28.848213 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hxrnh\" (UniqueName: \"kubernetes.io/projected/970c513a-2a0b-4a15-b051-e7215330c566-kube-api-access-hxrnh\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:28 crc kubenswrapper[4824]: I1211 02:22:28.905987 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"33fd5103-fcb0-404a-a730-4fdbbbc3577f","Type":"ContainerStarted","Data":"949b83cb29da9cf73d2b516df9e3e9973f4de1056f4647fbc2e4300d89d522b1"} Dec 11 02:22:28 crc kubenswrapper[4824]: I1211 02:22:28.911104 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 11 02:22:28 crc kubenswrapper[4824]: I1211 02:22:28.911098 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"970c513a-2a0b-4a15-b051-e7215330c566","Type":"ContainerDied","Data":"2dc35fa571dce36d41341fae6229aeeb8fbc31be620b625f5d0ff140df1c8dc6"} Dec 11 02:22:28 crc kubenswrapper[4824]: I1211 02:22:28.911313 4824 scope.go:117] "RemoveContainer" containerID="2dc35fa571dce36d41341fae6229aeeb8fbc31be620b625f5d0ff140df1c8dc6" Dec 11 02:22:28 crc kubenswrapper[4824]: I1211 02:22:28.911117 4824 generic.go:334] "Generic (PLEG): container finished" podID="970c513a-2a0b-4a15-b051-e7215330c566" containerID="2dc35fa571dce36d41341fae6229aeeb8fbc31be620b625f5d0ff140df1c8dc6" exitCode=0 Dec 11 02:22:28 crc kubenswrapper[4824]: I1211 02:22:28.911391 4824 generic.go:334] "Generic (PLEG): container finished" podID="970c513a-2a0b-4a15-b051-e7215330c566" containerID="2e16c5c6832b346a8f74ddd8a7659b5e7a58a48e4a87eb5bea7656164467cd5f" exitCode=143 Dec 11 02:22:28 crc kubenswrapper[4824]: I1211 02:22:28.911431 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"970c513a-2a0b-4a15-b051-e7215330c566","Type":"ContainerDied","Data":"2e16c5c6832b346a8f74ddd8a7659b5e7a58a48e4a87eb5bea7656164467cd5f"} Dec 11 02:22:28 crc kubenswrapper[4824]: I1211 02:22:28.911453 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"970c513a-2a0b-4a15-b051-e7215330c566","Type":"ContainerDied","Data":"4ac679b0f378001bfa880a71aa44a4197801a057e35536f7edfbce56a1773694"} Dec 11 02:22:28 crc kubenswrapper[4824]: I1211 02:22:28.927086 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.338613035 podStartE2EDuration="5.927070909s" podCreationTimestamp="2025-12-11 02:22:23 +0000 UTC" firstStartedPulling="2025-12-11 02:22:25.019493403 +0000 UTC m=+1286.708530782" lastFinishedPulling="2025-12-11 02:22:26.607951277 +0000 UTC m=+1288.296988656" observedRunningTime="2025-12-11 02:22:28.923681166 +0000 UTC m=+1290.612718545" watchObservedRunningTime="2025-12-11 02:22:28.927070909 +0000 UTC m=+1290.616108288" Dec 11 02:22:28 crc kubenswrapper[4824]: I1211 02:22:28.940684 4824 scope.go:117] "RemoveContainer" containerID="2e16c5c6832b346a8f74ddd8a7659b5e7a58a48e4a87eb5bea7656164467cd5f" Dec 11 02:22:28 crc kubenswrapper[4824]: I1211 02:22:28.958268 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 11 02:22:28 crc kubenswrapper[4824]: I1211 02:22:28.981878 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Dec 11 02:22:28 crc kubenswrapper[4824]: I1211 02:22:28.988140 4824 scope.go:117] "RemoveContainer" containerID="2dc35fa571dce36d41341fae6229aeeb8fbc31be620b625f5d0ff140df1c8dc6" Dec 11 02:22:29 crc kubenswrapper[4824]: E1211 02:22:28.996657 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2dc35fa571dce36d41341fae6229aeeb8fbc31be620b625f5d0ff140df1c8dc6\": container with ID starting with 2dc35fa571dce36d41341fae6229aeeb8fbc31be620b625f5d0ff140df1c8dc6 not found: ID does not exist" containerID="2dc35fa571dce36d41341fae6229aeeb8fbc31be620b625f5d0ff140df1c8dc6" Dec 11 02:22:29 crc kubenswrapper[4824]: I1211 02:22:28.996700 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2dc35fa571dce36d41341fae6229aeeb8fbc31be620b625f5d0ff140df1c8dc6"} err="failed to get container status \"2dc35fa571dce36d41341fae6229aeeb8fbc31be620b625f5d0ff140df1c8dc6\": rpc error: code = NotFound desc = could not find container \"2dc35fa571dce36d41341fae6229aeeb8fbc31be620b625f5d0ff140df1c8dc6\": container with ID starting with 2dc35fa571dce36d41341fae6229aeeb8fbc31be620b625f5d0ff140df1c8dc6 not found: ID does not exist" Dec 11 02:22:29 crc kubenswrapper[4824]: I1211 02:22:28.996728 4824 scope.go:117] "RemoveContainer" containerID="2e16c5c6832b346a8f74ddd8a7659b5e7a58a48e4a87eb5bea7656164467cd5f" Dec 11 02:22:29 crc kubenswrapper[4824]: E1211 02:22:28.998846 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e16c5c6832b346a8f74ddd8a7659b5e7a58a48e4a87eb5bea7656164467cd5f\": container with ID starting with 2e16c5c6832b346a8f74ddd8a7659b5e7a58a48e4a87eb5bea7656164467cd5f not found: ID does not exist" containerID="2e16c5c6832b346a8f74ddd8a7659b5e7a58a48e4a87eb5bea7656164467cd5f" Dec 11 02:22:29 crc kubenswrapper[4824]: I1211 02:22:28.998869 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e16c5c6832b346a8f74ddd8a7659b5e7a58a48e4a87eb5bea7656164467cd5f"} err="failed to get container status \"2e16c5c6832b346a8f74ddd8a7659b5e7a58a48e4a87eb5bea7656164467cd5f\": rpc error: code = NotFound desc = could not find container \"2e16c5c6832b346a8f74ddd8a7659b5e7a58a48e4a87eb5bea7656164467cd5f\": container with ID starting with 2e16c5c6832b346a8f74ddd8a7659b5e7a58a48e4a87eb5bea7656164467cd5f not found: ID does not exist" Dec 11 02:22:29 crc kubenswrapper[4824]: I1211 02:22:28.998883 4824 scope.go:117] "RemoveContainer" containerID="2dc35fa571dce36d41341fae6229aeeb8fbc31be620b625f5d0ff140df1c8dc6" Dec 11 02:22:29 crc kubenswrapper[4824]: I1211 02:22:28.999265 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2dc35fa571dce36d41341fae6229aeeb8fbc31be620b625f5d0ff140df1c8dc6"} err="failed to get container status \"2dc35fa571dce36d41341fae6229aeeb8fbc31be620b625f5d0ff140df1c8dc6\": rpc error: code = NotFound desc = could not find container \"2dc35fa571dce36d41341fae6229aeeb8fbc31be620b625f5d0ff140df1c8dc6\": container with ID starting with 2dc35fa571dce36d41341fae6229aeeb8fbc31be620b625f5d0ff140df1c8dc6 not found: ID does not exist" Dec 11 02:22:29 crc kubenswrapper[4824]: I1211 02:22:28.999281 4824 scope.go:117] "RemoveContainer" containerID="2e16c5c6832b346a8f74ddd8a7659b5e7a58a48e4a87eb5bea7656164467cd5f" Dec 11 02:22:29 crc kubenswrapper[4824]: I1211 02:22:28.999676 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e16c5c6832b346a8f74ddd8a7659b5e7a58a48e4a87eb5bea7656164467cd5f"} err="failed to get container status \"2e16c5c6832b346a8f74ddd8a7659b5e7a58a48e4a87eb5bea7656164467cd5f\": rpc error: code = NotFound desc = could not find container \"2e16c5c6832b346a8f74ddd8a7659b5e7a58a48e4a87eb5bea7656164467cd5f\": container with ID starting with 2e16c5c6832b346a8f74ddd8a7659b5e7a58a48e4a87eb5bea7656164467cd5f not found: ID does not exist" Dec 11 02:22:29 crc kubenswrapper[4824]: E1211 02:22:29.008594 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod970c513a_2a0b_4a15_b051_e7215330c566.slice\": RecentStats: unable to find data in memory cache]" Dec 11 02:22:29 crc kubenswrapper[4824]: I1211 02:22:29.010411 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 11 02:22:29 crc kubenswrapper[4824]: E1211 02:22:29.011462 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="970c513a-2a0b-4a15-b051-e7215330c566" containerName="cinder-api-log" Dec 11 02:22:29 crc kubenswrapper[4824]: I1211 02:22:29.011477 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="970c513a-2a0b-4a15-b051-e7215330c566" containerName="cinder-api-log" Dec 11 02:22:29 crc kubenswrapper[4824]: E1211 02:22:29.011514 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ae0553d-4ec8-425b-bdf4-b7992b7cbda6" containerName="init" Dec 11 02:22:29 crc kubenswrapper[4824]: I1211 02:22:29.011520 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ae0553d-4ec8-425b-bdf4-b7992b7cbda6" containerName="init" Dec 11 02:22:29 crc kubenswrapper[4824]: E1211 02:22:29.011538 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="970c513a-2a0b-4a15-b051-e7215330c566" containerName="cinder-api" Dec 11 02:22:29 crc kubenswrapper[4824]: I1211 02:22:29.011789 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="970c513a-2a0b-4a15-b051-e7215330c566" containerName="cinder-api" Dec 11 02:22:29 crc kubenswrapper[4824]: E1211 02:22:29.011802 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ae0553d-4ec8-425b-bdf4-b7992b7cbda6" containerName="dnsmasq-dns" Dec 11 02:22:29 crc kubenswrapper[4824]: I1211 02:22:29.011808 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ae0553d-4ec8-425b-bdf4-b7992b7cbda6" containerName="dnsmasq-dns" Dec 11 02:22:29 crc kubenswrapper[4824]: I1211 02:22:29.011981 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="970c513a-2a0b-4a15-b051-e7215330c566" containerName="cinder-api" Dec 11 02:22:29 crc kubenswrapper[4824]: I1211 02:22:29.011996 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="970c513a-2a0b-4a15-b051-e7215330c566" containerName="cinder-api-log" Dec 11 02:22:29 crc kubenswrapper[4824]: I1211 02:22:29.012013 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ae0553d-4ec8-425b-bdf4-b7992b7cbda6" containerName="dnsmasq-dns" Dec 11 02:22:29 crc kubenswrapper[4824]: I1211 02:22:29.013491 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 11 02:22:29 crc kubenswrapper[4824]: I1211 02:22:29.015842 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Dec 11 02:22:29 crc kubenswrapper[4824]: I1211 02:22:29.015999 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Dec 11 02:22:29 crc kubenswrapper[4824]: I1211 02:22:29.016608 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 11 02:22:29 crc kubenswrapper[4824]: I1211 02:22:29.024773 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 11 02:22:29 crc kubenswrapper[4824]: I1211 02:22:29.153521 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/717f2bc0-db62-483f-83e6-4d2f247eb2d6-public-tls-certs\") pod \"cinder-api-0\" (UID: \"717f2bc0-db62-483f-83e6-4d2f247eb2d6\") " pod="openstack/cinder-api-0" Dec 11 02:22:29 crc kubenswrapper[4824]: I1211 02:22:29.153575 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/717f2bc0-db62-483f-83e6-4d2f247eb2d6-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"717f2bc0-db62-483f-83e6-4d2f247eb2d6\") " pod="openstack/cinder-api-0" Dec 11 02:22:29 crc kubenswrapper[4824]: I1211 02:22:29.153603 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/717f2bc0-db62-483f-83e6-4d2f247eb2d6-config-data\") pod \"cinder-api-0\" (UID: \"717f2bc0-db62-483f-83e6-4d2f247eb2d6\") " pod="openstack/cinder-api-0" Dec 11 02:22:29 crc kubenswrapper[4824]: I1211 02:22:29.153621 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/717f2bc0-db62-483f-83e6-4d2f247eb2d6-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"717f2bc0-db62-483f-83e6-4d2f247eb2d6\") " pod="openstack/cinder-api-0" Dec 11 02:22:29 crc kubenswrapper[4824]: I1211 02:22:29.153744 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/717f2bc0-db62-483f-83e6-4d2f247eb2d6-config-data-custom\") pod \"cinder-api-0\" (UID: \"717f2bc0-db62-483f-83e6-4d2f247eb2d6\") " pod="openstack/cinder-api-0" Dec 11 02:22:29 crc kubenswrapper[4824]: I1211 02:22:29.153768 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/717f2bc0-db62-483f-83e6-4d2f247eb2d6-etc-machine-id\") pod \"cinder-api-0\" (UID: \"717f2bc0-db62-483f-83e6-4d2f247eb2d6\") " pod="openstack/cinder-api-0" Dec 11 02:22:29 crc kubenswrapper[4824]: I1211 02:22:29.153848 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/717f2bc0-db62-483f-83e6-4d2f247eb2d6-scripts\") pod \"cinder-api-0\" (UID: \"717f2bc0-db62-483f-83e6-4d2f247eb2d6\") " pod="openstack/cinder-api-0" Dec 11 02:22:29 crc kubenswrapper[4824]: I1211 02:22:29.153865 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/717f2bc0-db62-483f-83e6-4d2f247eb2d6-logs\") pod \"cinder-api-0\" (UID: \"717f2bc0-db62-483f-83e6-4d2f247eb2d6\") " pod="openstack/cinder-api-0" Dec 11 02:22:29 crc kubenswrapper[4824]: I1211 02:22:29.153887 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cjcws\" (UniqueName: \"kubernetes.io/projected/717f2bc0-db62-483f-83e6-4d2f247eb2d6-kube-api-access-cjcws\") pod \"cinder-api-0\" (UID: \"717f2bc0-db62-483f-83e6-4d2f247eb2d6\") " pod="openstack/cinder-api-0" Dec 11 02:22:29 crc kubenswrapper[4824]: I1211 02:22:29.249935 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 11 02:22:29 crc kubenswrapper[4824]: I1211 02:22:29.255251 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/717f2bc0-db62-483f-83e6-4d2f247eb2d6-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"717f2bc0-db62-483f-83e6-4d2f247eb2d6\") " pod="openstack/cinder-api-0" Dec 11 02:22:29 crc kubenswrapper[4824]: I1211 02:22:29.255293 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/717f2bc0-db62-483f-83e6-4d2f247eb2d6-config-data\") pod \"cinder-api-0\" (UID: \"717f2bc0-db62-483f-83e6-4d2f247eb2d6\") " pod="openstack/cinder-api-0" Dec 11 02:22:29 crc kubenswrapper[4824]: I1211 02:22:29.255312 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/717f2bc0-db62-483f-83e6-4d2f247eb2d6-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"717f2bc0-db62-483f-83e6-4d2f247eb2d6\") " pod="openstack/cinder-api-0" Dec 11 02:22:29 crc kubenswrapper[4824]: I1211 02:22:29.255363 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/717f2bc0-db62-483f-83e6-4d2f247eb2d6-config-data-custom\") pod \"cinder-api-0\" (UID: \"717f2bc0-db62-483f-83e6-4d2f247eb2d6\") " pod="openstack/cinder-api-0" Dec 11 02:22:29 crc kubenswrapper[4824]: I1211 02:22:29.255379 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/717f2bc0-db62-483f-83e6-4d2f247eb2d6-etc-machine-id\") pod \"cinder-api-0\" (UID: \"717f2bc0-db62-483f-83e6-4d2f247eb2d6\") " pod="openstack/cinder-api-0" Dec 11 02:22:29 crc kubenswrapper[4824]: I1211 02:22:29.255437 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/717f2bc0-db62-483f-83e6-4d2f247eb2d6-scripts\") pod \"cinder-api-0\" (UID: \"717f2bc0-db62-483f-83e6-4d2f247eb2d6\") " pod="openstack/cinder-api-0" Dec 11 02:22:29 crc kubenswrapper[4824]: I1211 02:22:29.255451 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/717f2bc0-db62-483f-83e6-4d2f247eb2d6-logs\") pod \"cinder-api-0\" (UID: \"717f2bc0-db62-483f-83e6-4d2f247eb2d6\") " pod="openstack/cinder-api-0" Dec 11 02:22:29 crc kubenswrapper[4824]: I1211 02:22:29.255471 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cjcws\" (UniqueName: \"kubernetes.io/projected/717f2bc0-db62-483f-83e6-4d2f247eb2d6-kube-api-access-cjcws\") pod \"cinder-api-0\" (UID: \"717f2bc0-db62-483f-83e6-4d2f247eb2d6\") " pod="openstack/cinder-api-0" Dec 11 02:22:29 crc kubenswrapper[4824]: I1211 02:22:29.255508 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/717f2bc0-db62-483f-83e6-4d2f247eb2d6-public-tls-certs\") pod \"cinder-api-0\" (UID: \"717f2bc0-db62-483f-83e6-4d2f247eb2d6\") " pod="openstack/cinder-api-0" Dec 11 02:22:29 crc kubenswrapper[4824]: I1211 02:22:29.256537 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/717f2bc0-db62-483f-83e6-4d2f247eb2d6-logs\") pod \"cinder-api-0\" (UID: \"717f2bc0-db62-483f-83e6-4d2f247eb2d6\") " pod="openstack/cinder-api-0" Dec 11 02:22:29 crc kubenswrapper[4824]: I1211 02:22:29.256589 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/717f2bc0-db62-483f-83e6-4d2f247eb2d6-etc-machine-id\") pod \"cinder-api-0\" (UID: \"717f2bc0-db62-483f-83e6-4d2f247eb2d6\") " pod="openstack/cinder-api-0" Dec 11 02:22:29 crc kubenswrapper[4824]: I1211 02:22:29.259391 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/717f2bc0-db62-483f-83e6-4d2f247eb2d6-public-tls-certs\") pod \"cinder-api-0\" (UID: \"717f2bc0-db62-483f-83e6-4d2f247eb2d6\") " pod="openstack/cinder-api-0" Dec 11 02:22:29 crc kubenswrapper[4824]: I1211 02:22:29.259817 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/717f2bc0-db62-483f-83e6-4d2f247eb2d6-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"717f2bc0-db62-483f-83e6-4d2f247eb2d6\") " pod="openstack/cinder-api-0" Dec 11 02:22:29 crc kubenswrapper[4824]: I1211 02:22:29.260158 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/717f2bc0-db62-483f-83e6-4d2f247eb2d6-config-data-custom\") pod \"cinder-api-0\" (UID: \"717f2bc0-db62-483f-83e6-4d2f247eb2d6\") " pod="openstack/cinder-api-0" Dec 11 02:22:29 crc kubenswrapper[4824]: I1211 02:22:29.260522 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/717f2bc0-db62-483f-83e6-4d2f247eb2d6-scripts\") pod \"cinder-api-0\" (UID: \"717f2bc0-db62-483f-83e6-4d2f247eb2d6\") " pod="openstack/cinder-api-0" Dec 11 02:22:29 crc kubenswrapper[4824]: I1211 02:22:29.263253 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/717f2bc0-db62-483f-83e6-4d2f247eb2d6-config-data\") pod \"cinder-api-0\" (UID: \"717f2bc0-db62-483f-83e6-4d2f247eb2d6\") " pod="openstack/cinder-api-0" Dec 11 02:22:29 crc kubenswrapper[4824]: I1211 02:22:29.272787 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/717f2bc0-db62-483f-83e6-4d2f247eb2d6-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"717f2bc0-db62-483f-83e6-4d2f247eb2d6\") " pod="openstack/cinder-api-0" Dec 11 02:22:29 crc kubenswrapper[4824]: I1211 02:22:29.276161 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cjcws\" (UniqueName: \"kubernetes.io/projected/717f2bc0-db62-483f-83e6-4d2f247eb2d6-kube-api-access-cjcws\") pod \"cinder-api-0\" (UID: \"717f2bc0-db62-483f-83e6-4d2f247eb2d6\") " pod="openstack/cinder-api-0" Dec 11 02:22:29 crc kubenswrapper[4824]: I1211 02:22:29.367548 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 11 02:22:29 crc kubenswrapper[4824]: I1211 02:22:29.826817 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 11 02:22:29 crc kubenswrapper[4824]: W1211 02:22:29.829746 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod717f2bc0_db62_483f_83e6_4d2f247eb2d6.slice/crio-58e2e136416ea91e1b39e407b5e6c009afca50c87b211c20b65105d3ecb1767e WatchSource:0}: Error finding container 58e2e136416ea91e1b39e407b5e6c009afca50c87b211c20b65105d3ecb1767e: Status 404 returned error can't find the container with id 58e2e136416ea91e1b39e407b5e6c009afca50c87b211c20b65105d3ecb1767e Dec 11 02:22:29 crc kubenswrapper[4824]: I1211 02:22:29.922434 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"717f2bc0-db62-483f-83e6-4d2f247eb2d6","Type":"ContainerStarted","Data":"58e2e136416ea91e1b39e407b5e6c009afca50c87b211c20b65105d3ecb1767e"} Dec 11 02:22:30 crc kubenswrapper[4824]: I1211 02:22:30.654287 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="970c513a-2a0b-4a15-b051-e7215330c566" path="/var/lib/kubelet/pods/970c513a-2a0b-4a15-b051-e7215330c566/volumes" Dec 11 02:22:30 crc kubenswrapper[4824]: I1211 02:22:30.945942 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"717f2bc0-db62-483f-83e6-4d2f247eb2d6","Type":"ContainerStarted","Data":"802d82dd759ad41dae77fb8f6f97d28eca962f36972ba092f0014a9ec98c1b44"} Dec 11 02:22:32 crc kubenswrapper[4824]: I1211 02:22:32.527514 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-787649dcdd-ts6qs" podUID="f15f9ed0-98f2-42fe-8006-2142098cc2b6" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": read tcp 10.217.0.2:52696->10.217.0.147:8443: read: connection reset by peer" Dec 11 02:22:32 crc kubenswrapper[4824]: I1211 02:22:32.818325 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7c49d4759d-9h88v" Dec 11 02:22:32 crc kubenswrapper[4824]: I1211 02:22:32.856059 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7c49d4759d-9h88v" Dec 11 02:22:32 crc kubenswrapper[4824]: I1211 02:22:32.967340 4824 generic.go:334] "Generic (PLEG): container finished" podID="d7637730-463d-4742-a5fb-4ed6029edd61" containerID="f5723003639a4381041359444637575c08e0809c0fa337753ae1b52e305e1592" exitCode=137 Dec 11 02:22:32 crc kubenswrapper[4824]: I1211 02:22:32.967377 4824 generic.go:334] "Generic (PLEG): container finished" podID="d7637730-463d-4742-a5fb-4ed6029edd61" containerID="61677087639bc33f5f228d3d104a74c1b37ee383d8903ad663d4f217b067440a" exitCode=137 Dec 11 02:22:32 crc kubenswrapper[4824]: I1211 02:22:32.967419 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5f68f577dc-nj2zp" event={"ID":"d7637730-463d-4742-a5fb-4ed6029edd61","Type":"ContainerDied","Data":"f5723003639a4381041359444637575c08e0809c0fa337753ae1b52e305e1592"} Dec 11 02:22:32 crc kubenswrapper[4824]: I1211 02:22:32.967449 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5f68f577dc-nj2zp" event={"ID":"d7637730-463d-4742-a5fb-4ed6029edd61","Type":"ContainerDied","Data":"61677087639bc33f5f228d3d104a74c1b37ee383d8903ad663d4f217b067440a"} Dec 11 02:22:32 crc kubenswrapper[4824]: I1211 02:22:32.969018 4824 generic.go:334] "Generic (PLEG): container finished" podID="f15f9ed0-98f2-42fe-8006-2142098cc2b6" containerID="588925740a1286914b8db0f80b0f9426464ecc031e105965e01b1abcbf40a9ba" exitCode=0 Dec 11 02:22:32 crc kubenswrapper[4824]: I1211 02:22:32.969887 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-787649dcdd-ts6qs" event={"ID":"f15f9ed0-98f2-42fe-8006-2142098cc2b6","Type":"ContainerDied","Data":"588925740a1286914b8db0f80b0f9426464ecc031e105965e01b1abcbf40a9ba"} Dec 11 02:22:33 crc kubenswrapper[4824]: I1211 02:22:33.559304 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-787649dcdd-ts6qs" podUID="f15f9ed0-98f2-42fe-8006-2142098cc2b6" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.147:8443: connect: connection refused" Dec 11 02:22:34 crc kubenswrapper[4824]: I1211 02:22:34.408296 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5c9776ccc5-rng95" Dec 11 02:22:34 crc kubenswrapper[4824]: I1211 02:22:34.412934 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5f68f577dc-nj2zp" Dec 11 02:22:34 crc kubenswrapper[4824]: I1211 02:22:34.457096 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d7637730-463d-4742-a5fb-4ed6029edd61-config-data\") pod \"d7637730-463d-4742-a5fb-4ed6029edd61\" (UID: \"d7637730-463d-4742-a5fb-4ed6029edd61\") " Dec 11 02:22:34 crc kubenswrapper[4824]: I1211 02:22:34.457179 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d7637730-463d-4742-a5fb-4ed6029edd61-scripts\") pod \"d7637730-463d-4742-a5fb-4ed6029edd61\" (UID: \"d7637730-463d-4742-a5fb-4ed6029edd61\") " Dec 11 02:22:34 crc kubenswrapper[4824]: I1211 02:22:34.457251 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rwmt9\" (UniqueName: \"kubernetes.io/projected/d7637730-463d-4742-a5fb-4ed6029edd61-kube-api-access-rwmt9\") pod \"d7637730-463d-4742-a5fb-4ed6029edd61\" (UID: \"d7637730-463d-4742-a5fb-4ed6029edd61\") " Dec 11 02:22:34 crc kubenswrapper[4824]: I1211 02:22:34.457292 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d7637730-463d-4742-a5fb-4ed6029edd61-horizon-secret-key\") pod \"d7637730-463d-4742-a5fb-4ed6029edd61\" (UID: \"d7637730-463d-4742-a5fb-4ed6029edd61\") " Dec 11 02:22:34 crc kubenswrapper[4824]: I1211 02:22:34.457349 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d7637730-463d-4742-a5fb-4ed6029edd61-logs\") pod \"d7637730-463d-4742-a5fb-4ed6029edd61\" (UID: \"d7637730-463d-4742-a5fb-4ed6029edd61\") " Dec 11 02:22:34 crc kubenswrapper[4824]: I1211 02:22:34.461671 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d7637730-463d-4742-a5fb-4ed6029edd61-logs" (OuterVolumeSpecName: "logs") pod "d7637730-463d-4742-a5fb-4ed6029edd61" (UID: "d7637730-463d-4742-a5fb-4ed6029edd61"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:22:34 crc kubenswrapper[4824]: I1211 02:22:34.489646 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7637730-463d-4742-a5fb-4ed6029edd61-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "d7637730-463d-4742-a5fb-4ed6029edd61" (UID: "d7637730-463d-4742-a5fb-4ed6029edd61"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:22:34 crc kubenswrapper[4824]: I1211 02:22:34.491365 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7637730-463d-4742-a5fb-4ed6029edd61-kube-api-access-rwmt9" (OuterVolumeSpecName: "kube-api-access-rwmt9") pod "d7637730-463d-4742-a5fb-4ed6029edd61" (UID: "d7637730-463d-4742-a5fb-4ed6029edd61"). InnerVolumeSpecName "kube-api-access-rwmt9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:22:34 crc kubenswrapper[4824]: I1211 02:22:34.540849 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-6wc4l"] Dec 11 02:22:34 crc kubenswrapper[4824]: I1211 02:22:34.541772 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-55f844cf75-6wc4l" podUID="89133b02-fcf7-4fe9-b511-c406d1b0d08b" containerName="dnsmasq-dns" containerID="cri-o://02c2354e3bfffc0ef0c43117b0a6607d70e56a2fc12c37f28004c3744f830cb3" gracePeriod=10 Dec 11 02:22:34 crc kubenswrapper[4824]: I1211 02:22:34.544438 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7637730-463d-4742-a5fb-4ed6029edd61-scripts" (OuterVolumeSpecName: "scripts") pod "d7637730-463d-4742-a5fb-4ed6029edd61" (UID: "d7637730-463d-4742-a5fb-4ed6029edd61"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:22:34 crc kubenswrapper[4824]: I1211 02:22:34.544808 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 11 02:22:34 crc kubenswrapper[4824]: I1211 02:22:34.581027 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d7637730-463d-4742-a5fb-4ed6029edd61-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:34 crc kubenswrapper[4824]: I1211 02:22:34.581060 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rwmt9\" (UniqueName: \"kubernetes.io/projected/d7637730-463d-4742-a5fb-4ed6029edd61-kube-api-access-rwmt9\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:34 crc kubenswrapper[4824]: I1211 02:22:34.581072 4824 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d7637730-463d-4742-a5fb-4ed6029edd61-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:34 crc kubenswrapper[4824]: I1211 02:22:34.581080 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d7637730-463d-4742-a5fb-4ed6029edd61-logs\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:34 crc kubenswrapper[4824]: I1211 02:22:34.611994 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7637730-463d-4742-a5fb-4ed6029edd61-config-data" (OuterVolumeSpecName: "config-data") pod "d7637730-463d-4742-a5fb-4ed6029edd61" (UID: "d7637730-463d-4742-a5fb-4ed6029edd61"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:22:34 crc kubenswrapper[4824]: I1211 02:22:34.683100 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d7637730-463d-4742-a5fb-4ed6029edd61-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:34 crc kubenswrapper[4824]: I1211 02:22:34.703319 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 11 02:22:34 crc kubenswrapper[4824]: I1211 02:22:34.733718 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-55f844cf75-6wc4l" podUID="89133b02-fcf7-4fe9-b511-c406d1b0d08b" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.150:5353: connect: connection refused" Dec 11 02:22:34 crc kubenswrapper[4824]: I1211 02:22:34.828399 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-5fbc99977b-tggml" Dec 11 02:22:34 crc kubenswrapper[4824]: E1211 02:22:34.902803 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="efa62db6-8875-4ae6-99ea-007e08be02bd" Dec 11 02:22:35 crc kubenswrapper[4824]: I1211 02:22:35.017504 4824 generic.go:334] "Generic (PLEG): container finished" podID="89133b02-fcf7-4fe9-b511-c406d1b0d08b" containerID="02c2354e3bfffc0ef0c43117b0a6607d70e56a2fc12c37f28004c3744f830cb3" exitCode=0 Dec 11 02:22:35 crc kubenswrapper[4824]: I1211 02:22:35.017611 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-6wc4l" event={"ID":"89133b02-fcf7-4fe9-b511-c406d1b0d08b","Type":"ContainerDied","Data":"02c2354e3bfffc0ef0c43117b0a6607d70e56a2fc12c37f28004c3744f830cb3"} Dec 11 02:22:35 crc kubenswrapper[4824]: I1211 02:22:35.022400 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"efa62db6-8875-4ae6-99ea-007e08be02bd","Type":"ContainerStarted","Data":"7c7991b4e320b3dd6c8a11c0f63600fd2e2ddf2b8913551c68c904b58daf53f0"} Dec 11 02:22:35 crc kubenswrapper[4824]: I1211 02:22:35.022861 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="efa62db6-8875-4ae6-99ea-007e08be02bd" containerName="ceilometer-notification-agent" containerID="cri-o://765b4a0946857ba769223a1fb59d99286c294efe426c05d0a350f8fc43d2044e" gracePeriod=30 Dec 11 02:22:35 crc kubenswrapper[4824]: I1211 02:22:35.023252 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 11 02:22:35 crc kubenswrapper[4824]: I1211 02:22:35.023571 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="efa62db6-8875-4ae6-99ea-007e08be02bd" containerName="proxy-httpd" containerID="cri-o://7c7991b4e320b3dd6c8a11c0f63600fd2e2ddf2b8913551c68c904b58daf53f0" gracePeriod=30 Dec 11 02:22:35 crc kubenswrapper[4824]: I1211 02:22:35.023606 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="efa62db6-8875-4ae6-99ea-007e08be02bd" containerName="sg-core" containerID="cri-o://fec321104810659464b218840e399c47290eb82d1c7ea38755cdcea70b40557b" gracePeriod=30 Dec 11 02:22:35 crc kubenswrapper[4824]: I1211 02:22:35.026412 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-6wc4l" Dec 11 02:22:35 crc kubenswrapper[4824]: I1211 02:22:35.026607 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5f68f577dc-nj2zp" Dec 11 02:22:35 crc kubenswrapper[4824]: I1211 02:22:35.026597 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5f68f577dc-nj2zp" event={"ID":"d7637730-463d-4742-a5fb-4ed6029edd61","Type":"ContainerDied","Data":"cd385f1cb3484d0e9da65069df64c3791209f38b68d60f06475a2e9df506d95f"} Dec 11 02:22:35 crc kubenswrapper[4824]: I1211 02:22:35.026669 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="33fd5103-fcb0-404a-a730-4fdbbbc3577f" containerName="cinder-scheduler" containerID="cri-o://b1d7c28f9270b8f42148267da1914aede650a0637bfb612e2541d29aef379dc7" gracePeriod=30 Dec 11 02:22:35 crc kubenswrapper[4824]: I1211 02:22:35.026694 4824 scope.go:117] "RemoveContainer" containerID="f5723003639a4381041359444637575c08e0809c0fa337753ae1b52e305e1592" Dec 11 02:22:35 crc kubenswrapper[4824]: I1211 02:22:35.026759 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="33fd5103-fcb0-404a-a730-4fdbbbc3577f" containerName="probe" containerID="cri-o://949b83cb29da9cf73d2b516df9e3e9973f4de1056f4647fbc2e4300d89d522b1" gracePeriod=30 Dec 11 02:22:35 crc kubenswrapper[4824]: I1211 02:22:35.083930 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5f68f577dc-nj2zp"] Dec 11 02:22:35 crc kubenswrapper[4824]: I1211 02:22:35.092062 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-5f68f577dc-nj2zp"] Dec 11 02:22:35 crc kubenswrapper[4824]: I1211 02:22:35.093177 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mj9hw\" (UniqueName: \"kubernetes.io/projected/89133b02-fcf7-4fe9-b511-c406d1b0d08b-kube-api-access-mj9hw\") pod \"89133b02-fcf7-4fe9-b511-c406d1b0d08b\" (UID: \"89133b02-fcf7-4fe9-b511-c406d1b0d08b\") " Dec 11 02:22:35 crc kubenswrapper[4824]: I1211 02:22:35.093251 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/89133b02-fcf7-4fe9-b511-c406d1b0d08b-dns-swift-storage-0\") pod \"89133b02-fcf7-4fe9-b511-c406d1b0d08b\" (UID: \"89133b02-fcf7-4fe9-b511-c406d1b0d08b\") " Dec 11 02:22:35 crc kubenswrapper[4824]: I1211 02:22:35.093301 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/89133b02-fcf7-4fe9-b511-c406d1b0d08b-dns-svc\") pod \"89133b02-fcf7-4fe9-b511-c406d1b0d08b\" (UID: \"89133b02-fcf7-4fe9-b511-c406d1b0d08b\") " Dec 11 02:22:35 crc kubenswrapper[4824]: I1211 02:22:35.093355 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/89133b02-fcf7-4fe9-b511-c406d1b0d08b-ovsdbserver-sb\") pod \"89133b02-fcf7-4fe9-b511-c406d1b0d08b\" (UID: \"89133b02-fcf7-4fe9-b511-c406d1b0d08b\") " Dec 11 02:22:35 crc kubenswrapper[4824]: I1211 02:22:35.093409 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89133b02-fcf7-4fe9-b511-c406d1b0d08b-config\") pod \"89133b02-fcf7-4fe9-b511-c406d1b0d08b\" (UID: \"89133b02-fcf7-4fe9-b511-c406d1b0d08b\") " Dec 11 02:22:35 crc kubenswrapper[4824]: I1211 02:22:35.093439 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/89133b02-fcf7-4fe9-b511-c406d1b0d08b-ovsdbserver-nb\") pod \"89133b02-fcf7-4fe9-b511-c406d1b0d08b\" (UID: \"89133b02-fcf7-4fe9-b511-c406d1b0d08b\") " Dec 11 02:22:35 crc kubenswrapper[4824]: I1211 02:22:35.098004 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/89133b02-fcf7-4fe9-b511-c406d1b0d08b-kube-api-access-mj9hw" (OuterVolumeSpecName: "kube-api-access-mj9hw") pod "89133b02-fcf7-4fe9-b511-c406d1b0d08b" (UID: "89133b02-fcf7-4fe9-b511-c406d1b0d08b"). InnerVolumeSpecName "kube-api-access-mj9hw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:22:35 crc kubenswrapper[4824]: I1211 02:22:35.167808 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/89133b02-fcf7-4fe9-b511-c406d1b0d08b-config" (OuterVolumeSpecName: "config") pod "89133b02-fcf7-4fe9-b511-c406d1b0d08b" (UID: "89133b02-fcf7-4fe9-b511-c406d1b0d08b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:22:35 crc kubenswrapper[4824]: I1211 02:22:35.192768 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/89133b02-fcf7-4fe9-b511-c406d1b0d08b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "89133b02-fcf7-4fe9-b511-c406d1b0d08b" (UID: "89133b02-fcf7-4fe9-b511-c406d1b0d08b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:22:35 crc kubenswrapper[4824]: I1211 02:22:35.192931 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/89133b02-fcf7-4fe9-b511-c406d1b0d08b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "89133b02-fcf7-4fe9-b511-c406d1b0d08b" (UID: "89133b02-fcf7-4fe9-b511-c406d1b0d08b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:22:35 crc kubenswrapper[4824]: I1211 02:22:35.196746 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mj9hw\" (UniqueName: \"kubernetes.io/projected/89133b02-fcf7-4fe9-b511-c406d1b0d08b-kube-api-access-mj9hw\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:35 crc kubenswrapper[4824]: I1211 02:22:35.196783 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/89133b02-fcf7-4fe9-b511-c406d1b0d08b-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:35 crc kubenswrapper[4824]: I1211 02:22:35.196793 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/89133b02-fcf7-4fe9-b511-c406d1b0d08b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:35 crc kubenswrapper[4824]: I1211 02:22:35.196803 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89133b02-fcf7-4fe9-b511-c406d1b0d08b-config\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:35 crc kubenswrapper[4824]: I1211 02:22:35.203381 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/89133b02-fcf7-4fe9-b511-c406d1b0d08b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "89133b02-fcf7-4fe9-b511-c406d1b0d08b" (UID: "89133b02-fcf7-4fe9-b511-c406d1b0d08b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:22:35 crc kubenswrapper[4824]: I1211 02:22:35.205696 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/89133b02-fcf7-4fe9-b511-c406d1b0d08b-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "89133b02-fcf7-4fe9-b511-c406d1b0d08b" (UID: "89133b02-fcf7-4fe9-b511-c406d1b0d08b"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:22:35 crc kubenswrapper[4824]: I1211 02:22:35.262282 4824 scope.go:117] "RemoveContainer" containerID="61677087639bc33f5f228d3d104a74c1b37ee383d8903ad663d4f217b067440a" Dec 11 02:22:35 crc kubenswrapper[4824]: I1211 02:22:35.298126 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/89133b02-fcf7-4fe9-b511-c406d1b0d08b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:35 crc kubenswrapper[4824]: I1211 02:22:35.298170 4824 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/89133b02-fcf7-4fe9-b511-c406d1b0d08b-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:35 crc kubenswrapper[4824]: I1211 02:22:35.761920 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-c88ccd6d6-bg6lk" Dec 11 02:22:35 crc kubenswrapper[4824]: I1211 02:22:35.839930 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-c88ccd6d6-bg6lk" Dec 11 02:22:35 crc kubenswrapper[4824]: I1211 02:22:35.897671 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-7c49d4759d-9h88v"] Dec 11 02:22:35 crc kubenswrapper[4824]: I1211 02:22:35.897890 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-7c49d4759d-9h88v" podUID="f07eb1ab-ee38-41b4-9079-1ce271bdd3e9" containerName="barbican-api-log" containerID="cri-o://814eac34539186694805a58af437492cdbd9fe8b4d535c481361454f1f4495c1" gracePeriod=30 Dec 11 02:22:35 crc kubenswrapper[4824]: I1211 02:22:35.897983 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-7c49d4759d-9h88v" podUID="f07eb1ab-ee38-41b4-9079-1ce271bdd3e9" containerName="barbican-api" containerID="cri-o://7f36efde65203331d341ec6eb5de62d50e5c1ddbae4723e622401cc845678a61" gracePeriod=30 Dec 11 02:22:36 crc kubenswrapper[4824]: I1211 02:22:36.041931 4824 generic.go:334] "Generic (PLEG): container finished" podID="33fd5103-fcb0-404a-a730-4fdbbbc3577f" containerID="949b83cb29da9cf73d2b516df9e3e9973f4de1056f4647fbc2e4300d89d522b1" exitCode=0 Dec 11 02:22:36 crc kubenswrapper[4824]: I1211 02:22:36.042000 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"33fd5103-fcb0-404a-a730-4fdbbbc3577f","Type":"ContainerDied","Data":"949b83cb29da9cf73d2b516df9e3e9973f4de1056f4647fbc2e4300d89d522b1"} Dec 11 02:22:36 crc kubenswrapper[4824]: I1211 02:22:36.045598 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"717f2bc0-db62-483f-83e6-4d2f247eb2d6","Type":"ContainerStarted","Data":"75c619d385cf6cee2e9ee25f0b90ec39ae1e71692c0c00ec9b309fb36f781ea1"} Dec 11 02:22:36 crc kubenswrapper[4824]: I1211 02:22:36.046145 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 11 02:22:36 crc kubenswrapper[4824]: I1211 02:22:36.053457 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-6wc4l" event={"ID":"89133b02-fcf7-4fe9-b511-c406d1b0d08b","Type":"ContainerDied","Data":"b1196e415d722aae4e9d7d26febd55839579c84917b75a3911438ccc3b6ed0d2"} Dec 11 02:22:36 crc kubenswrapper[4824]: I1211 02:22:36.053500 4824 scope.go:117] "RemoveContainer" containerID="02c2354e3bfffc0ef0c43117b0a6607d70e56a2fc12c37f28004c3744f830cb3" Dec 11 02:22:36 crc kubenswrapper[4824]: I1211 02:22:36.053597 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-6wc4l" Dec 11 02:22:36 crc kubenswrapper[4824]: I1211 02:22:36.060276 4824 generic.go:334] "Generic (PLEG): container finished" podID="efa62db6-8875-4ae6-99ea-007e08be02bd" containerID="7c7991b4e320b3dd6c8a11c0f63600fd2e2ddf2b8913551c68c904b58daf53f0" exitCode=0 Dec 11 02:22:36 crc kubenswrapper[4824]: I1211 02:22:36.060304 4824 generic.go:334] "Generic (PLEG): container finished" podID="efa62db6-8875-4ae6-99ea-007e08be02bd" containerID="fec321104810659464b218840e399c47290eb82d1c7ea38755cdcea70b40557b" exitCode=2 Dec 11 02:22:36 crc kubenswrapper[4824]: I1211 02:22:36.060357 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"efa62db6-8875-4ae6-99ea-007e08be02bd","Type":"ContainerDied","Data":"7c7991b4e320b3dd6c8a11c0f63600fd2e2ddf2b8913551c68c904b58daf53f0"} Dec 11 02:22:36 crc kubenswrapper[4824]: I1211 02:22:36.060407 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"efa62db6-8875-4ae6-99ea-007e08be02bd","Type":"ContainerDied","Data":"fec321104810659464b218840e399c47290eb82d1c7ea38755cdcea70b40557b"} Dec 11 02:22:36 crc kubenswrapper[4824]: I1211 02:22:36.063493 4824 generic.go:334] "Generic (PLEG): container finished" podID="f07eb1ab-ee38-41b4-9079-1ce271bdd3e9" containerID="814eac34539186694805a58af437492cdbd9fe8b4d535c481361454f1f4495c1" exitCode=143 Dec 11 02:22:36 crc kubenswrapper[4824]: I1211 02:22:36.063556 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7c49d4759d-9h88v" event={"ID":"f07eb1ab-ee38-41b4-9079-1ce271bdd3e9","Type":"ContainerDied","Data":"814eac34539186694805a58af437492cdbd9fe8b4d535c481361454f1f4495c1"} Dec 11 02:22:36 crc kubenswrapper[4824]: I1211 02:22:36.082554 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=8.08253365 podStartE2EDuration="8.08253365s" podCreationTimestamp="2025-12-11 02:22:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:22:36.07598371 +0000 UTC m=+1297.765021109" watchObservedRunningTime="2025-12-11 02:22:36.08253365 +0000 UTC m=+1297.771571029" Dec 11 02:22:36 crc kubenswrapper[4824]: I1211 02:22:36.090231 4824 scope.go:117] "RemoveContainer" containerID="aaaa7aefcd04234f29f5d526421d68dd1abc945578a44da6db1f41f30c0fae32" Dec 11 02:22:36 crc kubenswrapper[4824]: I1211 02:22:36.108595 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-6wc4l"] Dec 11 02:22:36 crc kubenswrapper[4824]: I1211 02:22:36.115987 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-6wc4l"] Dec 11 02:22:36 crc kubenswrapper[4824]: I1211 02:22:36.641572 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="89133b02-fcf7-4fe9-b511-c406d1b0d08b" path="/var/lib/kubelet/pods/89133b02-fcf7-4fe9-b511-c406d1b0d08b/volumes" Dec 11 02:22:36 crc kubenswrapper[4824]: I1211 02:22:36.642230 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7637730-463d-4742-a5fb-4ed6029edd61" path="/var/lib/kubelet/pods/d7637730-463d-4742-a5fb-4ed6029edd61/volumes" Dec 11 02:22:36 crc kubenswrapper[4824]: I1211 02:22:36.926952 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-78b68944cb-4p8xq" Dec 11 02:22:37 crc kubenswrapper[4824]: I1211 02:22:37.013287 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-78b68944cb-4p8xq" Dec 11 02:22:37 crc kubenswrapper[4824]: I1211 02:22:37.250998 4824 patch_prober.go:28] interesting pod/machine-config-daemon-gx6xt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 02:22:37 crc kubenswrapper[4824]: I1211 02:22:37.251047 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 02:22:38 crc kubenswrapper[4824]: I1211 02:22:38.145652 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-5b766ddd67-lflrw" Dec 11 02:22:38 crc kubenswrapper[4824]: I1211 02:22:38.220437 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-5fbc99977b-tggml"] Dec 11 02:22:38 crc kubenswrapper[4824]: I1211 02:22:38.220699 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-5fbc99977b-tggml" podUID="0a39253e-8c29-41ae-a242-8c9fccfe2613" containerName="neutron-api" containerID="cri-o://4b69e28249b7383bb629b55abdf4c31bcabee14ef7807ca36836547df2950751" gracePeriod=30 Dec 11 02:22:38 crc kubenswrapper[4824]: I1211 02:22:38.221023 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-5fbc99977b-tggml" podUID="0a39253e-8c29-41ae-a242-8c9fccfe2613" containerName="neutron-httpd" containerID="cri-o://4f93a423b9aeeca2bc45f2bbf56b517529c486e408c84b916c2eb9ed430fa56f" gracePeriod=30 Dec 11 02:22:38 crc kubenswrapper[4824]: I1211 02:22:38.940521 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.072265 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/33fd5103-fcb0-404a-a730-4fdbbbc3577f-etc-machine-id\") pod \"33fd5103-fcb0-404a-a730-4fdbbbc3577f\" (UID: \"33fd5103-fcb0-404a-a730-4fdbbbc3577f\") " Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.072336 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33fd5103-fcb0-404a-a730-4fdbbbc3577f-config-data\") pod \"33fd5103-fcb0-404a-a730-4fdbbbc3577f\" (UID: \"33fd5103-fcb0-404a-a730-4fdbbbc3577f\") " Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.072386 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/33fd5103-fcb0-404a-a730-4fdbbbc3577f-config-data-custom\") pod \"33fd5103-fcb0-404a-a730-4fdbbbc3577f\" (UID: \"33fd5103-fcb0-404a-a730-4fdbbbc3577f\") " Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.072532 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c9tls\" (UniqueName: \"kubernetes.io/projected/33fd5103-fcb0-404a-a730-4fdbbbc3577f-kube-api-access-c9tls\") pod \"33fd5103-fcb0-404a-a730-4fdbbbc3577f\" (UID: \"33fd5103-fcb0-404a-a730-4fdbbbc3577f\") " Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.072563 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33fd5103-fcb0-404a-a730-4fdbbbc3577f-combined-ca-bundle\") pod \"33fd5103-fcb0-404a-a730-4fdbbbc3577f\" (UID: \"33fd5103-fcb0-404a-a730-4fdbbbc3577f\") " Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.072610 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33fd5103-fcb0-404a-a730-4fdbbbc3577f-scripts\") pod \"33fd5103-fcb0-404a-a730-4fdbbbc3577f\" (UID: \"33fd5103-fcb0-404a-a730-4fdbbbc3577f\") " Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.075199 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/33fd5103-fcb0-404a-a730-4fdbbbc3577f-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "33fd5103-fcb0-404a-a730-4fdbbbc3577f" (UID: "33fd5103-fcb0-404a-a730-4fdbbbc3577f"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.080036 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33fd5103-fcb0-404a-a730-4fdbbbc3577f-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "33fd5103-fcb0-404a-a730-4fdbbbc3577f" (UID: "33fd5103-fcb0-404a-a730-4fdbbbc3577f"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.080103 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33fd5103-fcb0-404a-a730-4fdbbbc3577f-scripts" (OuterVolumeSpecName: "scripts") pod "33fd5103-fcb0-404a-a730-4fdbbbc3577f" (UID: "33fd5103-fcb0-404a-a730-4fdbbbc3577f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.082730 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33fd5103-fcb0-404a-a730-4fdbbbc3577f-kube-api-access-c9tls" (OuterVolumeSpecName: "kube-api-access-c9tls") pod "33fd5103-fcb0-404a-a730-4fdbbbc3577f" (UID: "33fd5103-fcb0-404a-a730-4fdbbbc3577f"). InnerVolumeSpecName "kube-api-access-c9tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.088279 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-7c49d4759d-9h88v" podUID="f07eb1ab-ee38-41b4-9079-1ce271bdd3e9" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.160:9311/healthcheck\": read tcp 10.217.0.2:48426->10.217.0.160:9311: read: connection reset by peer" Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.088941 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-7c49d4759d-9h88v" podUID="f07eb1ab-ee38-41b4-9079-1ce271bdd3e9" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.160:9311/healthcheck\": read tcp 10.217.0.2:48432->10.217.0.160:9311: read: connection reset by peer" Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.108348 4824 generic.go:334] "Generic (PLEG): container finished" podID="33fd5103-fcb0-404a-a730-4fdbbbc3577f" containerID="b1d7c28f9270b8f42148267da1914aede650a0637bfb612e2541d29aef379dc7" exitCode=0 Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.108409 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"33fd5103-fcb0-404a-a730-4fdbbbc3577f","Type":"ContainerDied","Data":"b1d7c28f9270b8f42148267da1914aede650a0637bfb612e2541d29aef379dc7"} Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.108436 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"33fd5103-fcb0-404a-a730-4fdbbbc3577f","Type":"ContainerDied","Data":"46eb982f9d23d1911e3f768cf97fb3328e848220db8b231b201b8fc7a2cc8fb1"} Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.108451 4824 scope.go:117] "RemoveContainer" containerID="949b83cb29da9cf73d2b516df9e3e9973f4de1056f4647fbc2e4300d89d522b1" Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.108562 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.114065 4824 generic.go:334] "Generic (PLEG): container finished" podID="0a39253e-8c29-41ae-a242-8c9fccfe2613" containerID="4f93a423b9aeeca2bc45f2bbf56b517529c486e408c84b916c2eb9ed430fa56f" exitCode=0 Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.114111 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5fbc99977b-tggml" event={"ID":"0a39253e-8c29-41ae-a242-8c9fccfe2613","Type":"ContainerDied","Data":"4f93a423b9aeeca2bc45f2bbf56b517529c486e408c84b916c2eb9ed430fa56f"} Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.156686 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33fd5103-fcb0-404a-a730-4fdbbbc3577f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "33fd5103-fcb0-404a-a730-4fdbbbc3577f" (UID: "33fd5103-fcb0-404a-a730-4fdbbbc3577f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.185679 4824 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/33fd5103-fcb0-404a-a730-4fdbbbc3577f-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.185713 4824 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/33fd5103-fcb0-404a-a730-4fdbbbc3577f-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.185723 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c9tls\" (UniqueName: \"kubernetes.io/projected/33fd5103-fcb0-404a-a730-4fdbbbc3577f-kube-api-access-c9tls\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.185732 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33fd5103-fcb0-404a-a730-4fdbbbc3577f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.185743 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33fd5103-fcb0-404a-a730-4fdbbbc3577f-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.202492 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33fd5103-fcb0-404a-a730-4fdbbbc3577f-config-data" (OuterVolumeSpecName: "config-data") pod "33fd5103-fcb0-404a-a730-4fdbbbc3577f" (UID: "33fd5103-fcb0-404a-a730-4fdbbbc3577f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.208883 4824 scope.go:117] "RemoveContainer" containerID="b1d7c28f9270b8f42148267da1914aede650a0637bfb612e2541d29aef379dc7" Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.249631 4824 scope.go:117] "RemoveContainer" containerID="949b83cb29da9cf73d2b516df9e3e9973f4de1056f4647fbc2e4300d89d522b1" Dec 11 02:22:39 crc kubenswrapper[4824]: E1211 02:22:39.250076 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"949b83cb29da9cf73d2b516df9e3e9973f4de1056f4647fbc2e4300d89d522b1\": container with ID starting with 949b83cb29da9cf73d2b516df9e3e9973f4de1056f4647fbc2e4300d89d522b1 not found: ID does not exist" containerID="949b83cb29da9cf73d2b516df9e3e9973f4de1056f4647fbc2e4300d89d522b1" Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.250100 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"949b83cb29da9cf73d2b516df9e3e9973f4de1056f4647fbc2e4300d89d522b1"} err="failed to get container status \"949b83cb29da9cf73d2b516df9e3e9973f4de1056f4647fbc2e4300d89d522b1\": rpc error: code = NotFound desc = could not find container \"949b83cb29da9cf73d2b516df9e3e9973f4de1056f4647fbc2e4300d89d522b1\": container with ID starting with 949b83cb29da9cf73d2b516df9e3e9973f4de1056f4647fbc2e4300d89d522b1 not found: ID does not exist" Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.250144 4824 scope.go:117] "RemoveContainer" containerID="b1d7c28f9270b8f42148267da1914aede650a0637bfb612e2541d29aef379dc7" Dec 11 02:22:39 crc kubenswrapper[4824]: E1211 02:22:39.250334 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b1d7c28f9270b8f42148267da1914aede650a0637bfb612e2541d29aef379dc7\": container with ID starting with b1d7c28f9270b8f42148267da1914aede650a0637bfb612e2541d29aef379dc7 not found: ID does not exist" containerID="b1d7c28f9270b8f42148267da1914aede650a0637bfb612e2541d29aef379dc7" Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.250349 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1d7c28f9270b8f42148267da1914aede650a0637bfb612e2541d29aef379dc7"} err="failed to get container status \"b1d7c28f9270b8f42148267da1914aede650a0637bfb612e2541d29aef379dc7\": rpc error: code = NotFound desc = could not find container \"b1d7c28f9270b8f42148267da1914aede650a0637bfb612e2541d29aef379dc7\": container with ID starting with b1d7c28f9270b8f42148267da1914aede650a0637bfb612e2541d29aef379dc7 not found: ID does not exist" Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.287268 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33fd5103-fcb0-404a-a730-4fdbbbc3577f-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:39 crc kubenswrapper[4824]: E1211 02:22:39.298881 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf07eb1ab_ee38_41b4_9079_1ce271bdd3e9.slice/crio-7f36efde65203331d341ec6eb5de62d50e5c1ddbae4723e622401cc845678a61.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf07eb1ab_ee38_41b4_9079_1ce271bdd3e9.slice/crio-conmon-7f36efde65203331d341ec6eb5de62d50e5c1ddbae4723e622401cc845678a61.scope\": RecentStats: unable to find data in memory cache]" Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.427285 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7c49d4759d-9h88v" Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.445640 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.463679 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.489978 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f07eb1ab-ee38-41b4-9079-1ce271bdd3e9-logs\") pod \"f07eb1ab-ee38-41b4-9079-1ce271bdd3e9\" (UID: \"f07eb1ab-ee38-41b4-9079-1ce271bdd3e9\") " Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.490346 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vcprh\" (UniqueName: \"kubernetes.io/projected/f07eb1ab-ee38-41b4-9079-1ce271bdd3e9-kube-api-access-vcprh\") pod \"f07eb1ab-ee38-41b4-9079-1ce271bdd3e9\" (UID: \"f07eb1ab-ee38-41b4-9079-1ce271bdd3e9\") " Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.490392 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f07eb1ab-ee38-41b4-9079-1ce271bdd3e9-config-data-custom\") pod \"f07eb1ab-ee38-41b4-9079-1ce271bdd3e9\" (UID: \"f07eb1ab-ee38-41b4-9079-1ce271bdd3e9\") " Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.490468 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f07eb1ab-ee38-41b4-9079-1ce271bdd3e9-combined-ca-bundle\") pod \"f07eb1ab-ee38-41b4-9079-1ce271bdd3e9\" (UID: \"f07eb1ab-ee38-41b4-9079-1ce271bdd3e9\") " Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.490518 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f07eb1ab-ee38-41b4-9079-1ce271bdd3e9-config-data\") pod \"f07eb1ab-ee38-41b4-9079-1ce271bdd3e9\" (UID: \"f07eb1ab-ee38-41b4-9079-1ce271bdd3e9\") " Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.494015 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f07eb1ab-ee38-41b4-9079-1ce271bdd3e9-logs" (OuterVolumeSpecName: "logs") pod "f07eb1ab-ee38-41b4-9079-1ce271bdd3e9" (UID: "f07eb1ab-ee38-41b4-9079-1ce271bdd3e9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.500472 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.501364 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f07eb1ab-ee38-41b4-9079-1ce271bdd3e9-kube-api-access-vcprh" (OuterVolumeSpecName: "kube-api-access-vcprh") pod "f07eb1ab-ee38-41b4-9079-1ce271bdd3e9" (UID: "f07eb1ab-ee38-41b4-9079-1ce271bdd3e9"). InnerVolumeSpecName "kube-api-access-vcprh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.502820 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f07eb1ab-ee38-41b4-9079-1ce271bdd3e9-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "f07eb1ab-ee38-41b4-9079-1ce271bdd3e9" (UID: "f07eb1ab-ee38-41b4-9079-1ce271bdd3e9"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:22:39 crc kubenswrapper[4824]: E1211 02:22:39.503711 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7637730-463d-4742-a5fb-4ed6029edd61" containerName="horizon-log" Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.503731 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7637730-463d-4742-a5fb-4ed6029edd61" containerName="horizon-log" Dec 11 02:22:39 crc kubenswrapper[4824]: E1211 02:22:39.503755 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7637730-463d-4742-a5fb-4ed6029edd61" containerName="horizon" Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.503764 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7637730-463d-4742-a5fb-4ed6029edd61" containerName="horizon" Dec 11 02:22:39 crc kubenswrapper[4824]: E1211 02:22:39.503810 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33fd5103-fcb0-404a-a730-4fdbbbc3577f" containerName="cinder-scheduler" Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.503818 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="33fd5103-fcb0-404a-a730-4fdbbbc3577f" containerName="cinder-scheduler" Dec 11 02:22:39 crc kubenswrapper[4824]: E1211 02:22:39.503833 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f07eb1ab-ee38-41b4-9079-1ce271bdd3e9" containerName="barbican-api-log" Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.503840 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f07eb1ab-ee38-41b4-9079-1ce271bdd3e9" containerName="barbican-api-log" Dec 11 02:22:39 crc kubenswrapper[4824]: E1211 02:22:39.503861 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89133b02-fcf7-4fe9-b511-c406d1b0d08b" containerName="init" Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.503867 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="89133b02-fcf7-4fe9-b511-c406d1b0d08b" containerName="init" Dec 11 02:22:39 crc kubenswrapper[4824]: E1211 02:22:39.503884 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f07eb1ab-ee38-41b4-9079-1ce271bdd3e9" containerName="barbican-api" Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.503890 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f07eb1ab-ee38-41b4-9079-1ce271bdd3e9" containerName="barbican-api" Dec 11 02:22:39 crc kubenswrapper[4824]: E1211 02:22:39.503910 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33fd5103-fcb0-404a-a730-4fdbbbc3577f" containerName="probe" Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.503915 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="33fd5103-fcb0-404a-a730-4fdbbbc3577f" containerName="probe" Dec 11 02:22:39 crc kubenswrapper[4824]: E1211 02:22:39.503928 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89133b02-fcf7-4fe9-b511-c406d1b0d08b" containerName="dnsmasq-dns" Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.503933 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="89133b02-fcf7-4fe9-b511-c406d1b0d08b" containerName="dnsmasq-dns" Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.512818 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7637730-463d-4742-a5fb-4ed6029edd61" containerName="horizon-log" Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.512854 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7637730-463d-4742-a5fb-4ed6029edd61" containerName="horizon" Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.512873 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="89133b02-fcf7-4fe9-b511-c406d1b0d08b" containerName="dnsmasq-dns" Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.512900 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="33fd5103-fcb0-404a-a730-4fdbbbc3577f" containerName="probe" Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.512908 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f07eb1ab-ee38-41b4-9079-1ce271bdd3e9" containerName="barbican-api" Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.512927 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f07eb1ab-ee38-41b4-9079-1ce271bdd3e9" containerName="barbican-api-log" Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.512946 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="33fd5103-fcb0-404a-a730-4fdbbbc3577f" containerName="cinder-scheduler" Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.514717 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.517666 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.531517 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f07eb1ab-ee38-41b4-9079-1ce271bdd3e9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f07eb1ab-ee38-41b4-9079-1ce271bdd3e9" (UID: "f07eb1ab-ee38-41b4-9079-1ce271bdd3e9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.549248 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.562175 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f07eb1ab-ee38-41b4-9079-1ce271bdd3e9-config-data" (OuterVolumeSpecName: "config-data") pod "f07eb1ab-ee38-41b4-9079-1ce271bdd3e9" (UID: "f07eb1ab-ee38-41b4-9079-1ce271bdd3e9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.592206 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0dde14f9-85a8-4731-94ef-8199ea6d8a4b-scripts\") pod \"cinder-scheduler-0\" (UID: \"0dde14f9-85a8-4731-94ef-8199ea6d8a4b\") " pod="openstack/cinder-scheduler-0" Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.592467 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0dde14f9-85a8-4731-94ef-8199ea6d8a4b-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"0dde14f9-85a8-4731-94ef-8199ea6d8a4b\") " pod="openstack/cinder-scheduler-0" Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.592566 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0dde14f9-85a8-4731-94ef-8199ea6d8a4b-config-data\") pod \"cinder-scheduler-0\" (UID: \"0dde14f9-85a8-4731-94ef-8199ea6d8a4b\") " pod="openstack/cinder-scheduler-0" Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.592681 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0dde14f9-85a8-4731-94ef-8199ea6d8a4b-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"0dde14f9-85a8-4731-94ef-8199ea6d8a4b\") " pod="openstack/cinder-scheduler-0" Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.592760 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0dde14f9-85a8-4731-94ef-8199ea6d8a4b-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"0dde14f9-85a8-4731-94ef-8199ea6d8a4b\") " pod="openstack/cinder-scheduler-0" Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.592830 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lsf7l\" (UniqueName: \"kubernetes.io/projected/0dde14f9-85a8-4731-94ef-8199ea6d8a4b-kube-api-access-lsf7l\") pod \"cinder-scheduler-0\" (UID: \"0dde14f9-85a8-4731-94ef-8199ea6d8a4b\") " pod="openstack/cinder-scheduler-0" Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.592945 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f07eb1ab-ee38-41b4-9079-1ce271bdd3e9-logs\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.593009 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vcprh\" (UniqueName: \"kubernetes.io/projected/f07eb1ab-ee38-41b4-9079-1ce271bdd3e9-kube-api-access-vcprh\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.593062 4824 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f07eb1ab-ee38-41b4-9079-1ce271bdd3e9-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.593134 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f07eb1ab-ee38-41b4-9079-1ce271bdd3e9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.593187 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f07eb1ab-ee38-41b4-9079-1ce271bdd3e9-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.694523 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0dde14f9-85a8-4731-94ef-8199ea6d8a4b-config-data\") pod \"cinder-scheduler-0\" (UID: \"0dde14f9-85a8-4731-94ef-8199ea6d8a4b\") " pod="openstack/cinder-scheduler-0" Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.694615 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0dde14f9-85a8-4731-94ef-8199ea6d8a4b-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"0dde14f9-85a8-4731-94ef-8199ea6d8a4b\") " pod="openstack/cinder-scheduler-0" Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.694654 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0dde14f9-85a8-4731-94ef-8199ea6d8a4b-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"0dde14f9-85a8-4731-94ef-8199ea6d8a4b\") " pod="openstack/cinder-scheduler-0" Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.694687 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lsf7l\" (UniqueName: \"kubernetes.io/projected/0dde14f9-85a8-4731-94ef-8199ea6d8a4b-kube-api-access-lsf7l\") pod \"cinder-scheduler-0\" (UID: \"0dde14f9-85a8-4731-94ef-8199ea6d8a4b\") " pod="openstack/cinder-scheduler-0" Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.694798 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0dde14f9-85a8-4731-94ef-8199ea6d8a4b-scripts\") pod \"cinder-scheduler-0\" (UID: \"0dde14f9-85a8-4731-94ef-8199ea6d8a4b\") " pod="openstack/cinder-scheduler-0" Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.694832 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0dde14f9-85a8-4731-94ef-8199ea6d8a4b-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"0dde14f9-85a8-4731-94ef-8199ea6d8a4b\") " pod="openstack/cinder-scheduler-0" Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.694896 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0dde14f9-85a8-4731-94ef-8199ea6d8a4b-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"0dde14f9-85a8-4731-94ef-8199ea6d8a4b\") " pod="openstack/cinder-scheduler-0" Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.698572 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0dde14f9-85a8-4731-94ef-8199ea6d8a4b-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"0dde14f9-85a8-4731-94ef-8199ea6d8a4b\") " pod="openstack/cinder-scheduler-0" Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.698756 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0dde14f9-85a8-4731-94ef-8199ea6d8a4b-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"0dde14f9-85a8-4731-94ef-8199ea6d8a4b\") " pod="openstack/cinder-scheduler-0" Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.700102 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0dde14f9-85a8-4731-94ef-8199ea6d8a4b-config-data\") pod \"cinder-scheduler-0\" (UID: \"0dde14f9-85a8-4731-94ef-8199ea6d8a4b\") " pod="openstack/cinder-scheduler-0" Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.701240 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0dde14f9-85a8-4731-94ef-8199ea6d8a4b-scripts\") pod \"cinder-scheduler-0\" (UID: \"0dde14f9-85a8-4731-94ef-8199ea6d8a4b\") " pod="openstack/cinder-scheduler-0" Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.712180 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lsf7l\" (UniqueName: \"kubernetes.io/projected/0dde14f9-85a8-4731-94ef-8199ea6d8a4b-kube-api-access-lsf7l\") pod \"cinder-scheduler-0\" (UID: \"0dde14f9-85a8-4731-94ef-8199ea6d8a4b\") " pod="openstack/cinder-scheduler-0" Dec 11 02:22:39 crc kubenswrapper[4824]: I1211 02:22:39.933127 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.094132 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.128325 4824 generic.go:334] "Generic (PLEG): container finished" podID="efa62db6-8875-4ae6-99ea-007e08be02bd" containerID="765b4a0946857ba769223a1fb59d99286c294efe426c05d0a350f8fc43d2044e" exitCode=0 Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.128379 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"efa62db6-8875-4ae6-99ea-007e08be02bd","Type":"ContainerDied","Data":"765b4a0946857ba769223a1fb59d99286c294efe426c05d0a350f8fc43d2044e"} Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.128403 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"efa62db6-8875-4ae6-99ea-007e08be02bd","Type":"ContainerDied","Data":"7f115c2b8b658ee4f32895ec61a442af614b8c9df49d6fcd4f69ec87080063f2"} Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.128418 4824 scope.go:117] "RemoveContainer" containerID="7c7991b4e320b3dd6c8a11c0f63600fd2e2ddf2b8913551c68c904b58daf53f0" Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.128521 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.131306 4824 generic.go:334] "Generic (PLEG): container finished" podID="f07eb1ab-ee38-41b4-9079-1ce271bdd3e9" containerID="7f36efde65203331d341ec6eb5de62d50e5c1ddbae4723e622401cc845678a61" exitCode=0 Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.131346 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7c49d4759d-9h88v" event={"ID":"f07eb1ab-ee38-41b4-9079-1ce271bdd3e9","Type":"ContainerDied","Data":"7f36efde65203331d341ec6eb5de62d50e5c1ddbae4723e622401cc845678a61"} Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.131362 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7c49d4759d-9h88v" event={"ID":"f07eb1ab-ee38-41b4-9079-1ce271bdd3e9","Type":"ContainerDied","Data":"833ba186d8b42a0a51aaaa7a41831d8dd2640e3b5e0c4b7ba3b281197bf01212"} Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.131415 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7c49d4759d-9h88v" Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.165859 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-7c49d4759d-9h88v"] Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.172025 4824 scope.go:117] "RemoveContainer" containerID="fec321104810659464b218840e399c47290eb82d1c7ea38755cdcea70b40557b" Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.173553 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-7c49d4759d-9h88v"] Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.188537 4824 scope.go:117] "RemoveContainer" containerID="765b4a0946857ba769223a1fb59d99286c294efe426c05d0a350f8fc43d2044e" Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.204208 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/efa62db6-8875-4ae6-99ea-007e08be02bd-run-httpd\") pod \"efa62db6-8875-4ae6-99ea-007e08be02bd\" (UID: \"efa62db6-8875-4ae6-99ea-007e08be02bd\") " Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.204352 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cqww2\" (UniqueName: \"kubernetes.io/projected/efa62db6-8875-4ae6-99ea-007e08be02bd-kube-api-access-cqww2\") pod \"efa62db6-8875-4ae6-99ea-007e08be02bd\" (UID: \"efa62db6-8875-4ae6-99ea-007e08be02bd\") " Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.204381 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/efa62db6-8875-4ae6-99ea-007e08be02bd-scripts\") pod \"efa62db6-8875-4ae6-99ea-007e08be02bd\" (UID: \"efa62db6-8875-4ae6-99ea-007e08be02bd\") " Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.204449 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/efa62db6-8875-4ae6-99ea-007e08be02bd-combined-ca-bundle\") pod \"efa62db6-8875-4ae6-99ea-007e08be02bd\" (UID: \"efa62db6-8875-4ae6-99ea-007e08be02bd\") " Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.204473 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/efa62db6-8875-4ae6-99ea-007e08be02bd-config-data\") pod \"efa62db6-8875-4ae6-99ea-007e08be02bd\" (UID: \"efa62db6-8875-4ae6-99ea-007e08be02bd\") " Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.204491 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/efa62db6-8875-4ae6-99ea-007e08be02bd-log-httpd\") pod \"efa62db6-8875-4ae6-99ea-007e08be02bd\" (UID: \"efa62db6-8875-4ae6-99ea-007e08be02bd\") " Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.204584 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/efa62db6-8875-4ae6-99ea-007e08be02bd-sg-core-conf-yaml\") pod \"efa62db6-8875-4ae6-99ea-007e08be02bd\" (UID: \"efa62db6-8875-4ae6-99ea-007e08be02bd\") " Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.204669 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/efa62db6-8875-4ae6-99ea-007e08be02bd-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "efa62db6-8875-4ae6-99ea-007e08be02bd" (UID: "efa62db6-8875-4ae6-99ea-007e08be02bd"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.204930 4824 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/efa62db6-8875-4ae6-99ea-007e08be02bd-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.205485 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/efa62db6-8875-4ae6-99ea-007e08be02bd-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "efa62db6-8875-4ae6-99ea-007e08be02bd" (UID: "efa62db6-8875-4ae6-99ea-007e08be02bd"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.209307 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efa62db6-8875-4ae6-99ea-007e08be02bd-scripts" (OuterVolumeSpecName: "scripts") pod "efa62db6-8875-4ae6-99ea-007e08be02bd" (UID: "efa62db6-8875-4ae6-99ea-007e08be02bd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.210597 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efa62db6-8875-4ae6-99ea-007e08be02bd-kube-api-access-cqww2" (OuterVolumeSpecName: "kube-api-access-cqww2") pod "efa62db6-8875-4ae6-99ea-007e08be02bd" (UID: "efa62db6-8875-4ae6-99ea-007e08be02bd"). InnerVolumeSpecName "kube-api-access-cqww2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.245670 4824 scope.go:117] "RemoveContainer" containerID="7c7991b4e320b3dd6c8a11c0f63600fd2e2ddf2b8913551c68c904b58daf53f0" Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.247345 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efa62db6-8875-4ae6-99ea-007e08be02bd-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "efa62db6-8875-4ae6-99ea-007e08be02bd" (UID: "efa62db6-8875-4ae6-99ea-007e08be02bd"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:22:40 crc kubenswrapper[4824]: E1211 02:22:40.251497 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7c7991b4e320b3dd6c8a11c0f63600fd2e2ddf2b8913551c68c904b58daf53f0\": container with ID starting with 7c7991b4e320b3dd6c8a11c0f63600fd2e2ddf2b8913551c68c904b58daf53f0 not found: ID does not exist" containerID="7c7991b4e320b3dd6c8a11c0f63600fd2e2ddf2b8913551c68c904b58daf53f0" Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.251527 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c7991b4e320b3dd6c8a11c0f63600fd2e2ddf2b8913551c68c904b58daf53f0"} err="failed to get container status \"7c7991b4e320b3dd6c8a11c0f63600fd2e2ddf2b8913551c68c904b58daf53f0\": rpc error: code = NotFound desc = could not find container \"7c7991b4e320b3dd6c8a11c0f63600fd2e2ddf2b8913551c68c904b58daf53f0\": container with ID starting with 7c7991b4e320b3dd6c8a11c0f63600fd2e2ddf2b8913551c68c904b58daf53f0 not found: ID does not exist" Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.251547 4824 scope.go:117] "RemoveContainer" containerID="fec321104810659464b218840e399c47290eb82d1c7ea38755cdcea70b40557b" Dec 11 02:22:40 crc kubenswrapper[4824]: E1211 02:22:40.256465 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fec321104810659464b218840e399c47290eb82d1c7ea38755cdcea70b40557b\": container with ID starting with fec321104810659464b218840e399c47290eb82d1c7ea38755cdcea70b40557b not found: ID does not exist" containerID="fec321104810659464b218840e399c47290eb82d1c7ea38755cdcea70b40557b" Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.256490 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fec321104810659464b218840e399c47290eb82d1c7ea38755cdcea70b40557b"} err="failed to get container status \"fec321104810659464b218840e399c47290eb82d1c7ea38755cdcea70b40557b\": rpc error: code = NotFound desc = could not find container \"fec321104810659464b218840e399c47290eb82d1c7ea38755cdcea70b40557b\": container with ID starting with fec321104810659464b218840e399c47290eb82d1c7ea38755cdcea70b40557b not found: ID does not exist" Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.256503 4824 scope.go:117] "RemoveContainer" containerID="765b4a0946857ba769223a1fb59d99286c294efe426c05d0a350f8fc43d2044e" Dec 11 02:22:40 crc kubenswrapper[4824]: E1211 02:22:40.256921 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"765b4a0946857ba769223a1fb59d99286c294efe426c05d0a350f8fc43d2044e\": container with ID starting with 765b4a0946857ba769223a1fb59d99286c294efe426c05d0a350f8fc43d2044e not found: ID does not exist" containerID="765b4a0946857ba769223a1fb59d99286c294efe426c05d0a350f8fc43d2044e" Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.256946 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"765b4a0946857ba769223a1fb59d99286c294efe426c05d0a350f8fc43d2044e"} err="failed to get container status \"765b4a0946857ba769223a1fb59d99286c294efe426c05d0a350f8fc43d2044e\": rpc error: code = NotFound desc = could not find container \"765b4a0946857ba769223a1fb59d99286c294efe426c05d0a350f8fc43d2044e\": container with ID starting with 765b4a0946857ba769223a1fb59d99286c294efe426c05d0a350f8fc43d2044e not found: ID does not exist" Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.256961 4824 scope.go:117] "RemoveContainer" containerID="7f36efde65203331d341ec6eb5de62d50e5c1ddbae4723e622401cc845678a61" Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.272087 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efa62db6-8875-4ae6-99ea-007e08be02bd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "efa62db6-8875-4ae6-99ea-007e08be02bd" (UID: "efa62db6-8875-4ae6-99ea-007e08be02bd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.308314 4824 scope.go:117] "RemoveContainer" containerID="814eac34539186694805a58af437492cdbd9fe8b4d535c481361454f1f4495c1" Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.309222 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/efa62db6-8875-4ae6-99ea-007e08be02bd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.309239 4824 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/efa62db6-8875-4ae6-99ea-007e08be02bd-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.309249 4824 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/efa62db6-8875-4ae6-99ea-007e08be02bd-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.309258 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cqww2\" (UniqueName: \"kubernetes.io/projected/efa62db6-8875-4ae6-99ea-007e08be02bd-kube-api-access-cqww2\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.309267 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/efa62db6-8875-4ae6-99ea-007e08be02bd-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.348618 4824 scope.go:117] "RemoveContainer" containerID="7f36efde65203331d341ec6eb5de62d50e5c1ddbae4723e622401cc845678a61" Dec 11 02:22:40 crc kubenswrapper[4824]: E1211 02:22:40.354500 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7f36efde65203331d341ec6eb5de62d50e5c1ddbae4723e622401cc845678a61\": container with ID starting with 7f36efde65203331d341ec6eb5de62d50e5c1ddbae4723e622401cc845678a61 not found: ID does not exist" containerID="7f36efde65203331d341ec6eb5de62d50e5c1ddbae4723e622401cc845678a61" Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.354531 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f36efde65203331d341ec6eb5de62d50e5c1ddbae4723e622401cc845678a61"} err="failed to get container status \"7f36efde65203331d341ec6eb5de62d50e5c1ddbae4723e622401cc845678a61\": rpc error: code = NotFound desc = could not find container \"7f36efde65203331d341ec6eb5de62d50e5c1ddbae4723e622401cc845678a61\": container with ID starting with 7f36efde65203331d341ec6eb5de62d50e5c1ddbae4723e622401cc845678a61 not found: ID does not exist" Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.354564 4824 scope.go:117] "RemoveContainer" containerID="814eac34539186694805a58af437492cdbd9fe8b4d535c481361454f1f4495c1" Dec 11 02:22:40 crc kubenswrapper[4824]: E1211 02:22:40.372424 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"814eac34539186694805a58af437492cdbd9fe8b4d535c481361454f1f4495c1\": container with ID starting with 814eac34539186694805a58af437492cdbd9fe8b4d535c481361454f1f4495c1 not found: ID does not exist" containerID="814eac34539186694805a58af437492cdbd9fe8b4d535c481361454f1f4495c1" Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.372473 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"814eac34539186694805a58af437492cdbd9fe8b4d535c481361454f1f4495c1"} err="failed to get container status \"814eac34539186694805a58af437492cdbd9fe8b4d535c481361454f1f4495c1\": rpc error: code = NotFound desc = could not find container \"814eac34539186694805a58af437492cdbd9fe8b4d535c481361454f1f4495c1\": container with ID starting with 814eac34539186694805a58af437492cdbd9fe8b4d535c481361454f1f4495c1 not found: ID does not exist" Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.394267 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efa62db6-8875-4ae6-99ea-007e08be02bd-config-data" (OuterVolumeSpecName: "config-data") pod "efa62db6-8875-4ae6-99ea-007e08be02bd" (UID: "efa62db6-8875-4ae6-99ea-007e08be02bd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.412290 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/efa62db6-8875-4ae6-99ea-007e08be02bd-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.421981 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 11 02:22:40 crc kubenswrapper[4824]: W1211 02:22:40.422897 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0dde14f9_85a8_4731_94ef_8199ea6d8a4b.slice/crio-60eb1d52701f8a83c4f7109fa0193de14591d041966d444d7e1b98ae58c855a8 WatchSource:0}: Error finding container 60eb1d52701f8a83c4f7109fa0193de14591d041966d444d7e1b98ae58c855a8: Status 404 returned error can't find the container with id 60eb1d52701f8a83c4f7109fa0193de14591d041966d444d7e1b98ae58c855a8 Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.560357 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.566841 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.580756 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 11 02:22:40 crc kubenswrapper[4824]: E1211 02:22:40.581054 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efa62db6-8875-4ae6-99ea-007e08be02bd" containerName="sg-core" Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.581075 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="efa62db6-8875-4ae6-99ea-007e08be02bd" containerName="sg-core" Dec 11 02:22:40 crc kubenswrapper[4824]: E1211 02:22:40.581102 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efa62db6-8875-4ae6-99ea-007e08be02bd" containerName="ceilometer-notification-agent" Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.581113 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="efa62db6-8875-4ae6-99ea-007e08be02bd" containerName="ceilometer-notification-agent" Dec 11 02:22:40 crc kubenswrapper[4824]: E1211 02:22:40.581136 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efa62db6-8875-4ae6-99ea-007e08be02bd" containerName="proxy-httpd" Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.581143 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="efa62db6-8875-4ae6-99ea-007e08be02bd" containerName="proxy-httpd" Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.581335 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="efa62db6-8875-4ae6-99ea-007e08be02bd" containerName="ceilometer-notification-agent" Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.581374 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="efa62db6-8875-4ae6-99ea-007e08be02bd" containerName="sg-core" Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.581388 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="efa62db6-8875-4ae6-99ea-007e08be02bd" containerName="proxy-httpd" Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.582907 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.589559 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.589949 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.600884 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.642113 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="33fd5103-fcb0-404a-a730-4fdbbbc3577f" path="/var/lib/kubelet/pods/33fd5103-fcb0-404a-a730-4fdbbbc3577f/volumes" Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.642832 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efa62db6-8875-4ae6-99ea-007e08be02bd" path="/var/lib/kubelet/pods/efa62db6-8875-4ae6-99ea-007e08be02bd/volumes" Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.643491 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f07eb1ab-ee38-41b4-9079-1ce271bdd3e9" path="/var/lib/kubelet/pods/f07eb1ab-ee38-41b4-9079-1ce271bdd3e9/volumes" Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.717027 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9259b35-21b6-4971-ae1b-04d367d993de-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d9259b35-21b6-4971-ae1b-04d367d993de\") " pod="openstack/ceilometer-0" Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.717072 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9259b35-21b6-4971-ae1b-04d367d993de-config-data\") pod \"ceilometer-0\" (UID: \"d9259b35-21b6-4971-ae1b-04d367d993de\") " pod="openstack/ceilometer-0" Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.717113 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d9259b35-21b6-4971-ae1b-04d367d993de-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d9259b35-21b6-4971-ae1b-04d367d993de\") " pod="openstack/ceilometer-0" Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.717216 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xmkcd\" (UniqueName: \"kubernetes.io/projected/d9259b35-21b6-4971-ae1b-04d367d993de-kube-api-access-xmkcd\") pod \"ceilometer-0\" (UID: \"d9259b35-21b6-4971-ae1b-04d367d993de\") " pod="openstack/ceilometer-0" Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.717416 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d9259b35-21b6-4971-ae1b-04d367d993de-scripts\") pod \"ceilometer-0\" (UID: \"d9259b35-21b6-4971-ae1b-04d367d993de\") " pod="openstack/ceilometer-0" Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.717494 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d9259b35-21b6-4971-ae1b-04d367d993de-log-httpd\") pod \"ceilometer-0\" (UID: \"d9259b35-21b6-4971-ae1b-04d367d993de\") " pod="openstack/ceilometer-0" Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.717770 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d9259b35-21b6-4971-ae1b-04d367d993de-run-httpd\") pod \"ceilometer-0\" (UID: \"d9259b35-21b6-4971-ae1b-04d367d993de\") " pod="openstack/ceilometer-0" Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.820470 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d9259b35-21b6-4971-ae1b-04d367d993de-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d9259b35-21b6-4971-ae1b-04d367d993de\") " pod="openstack/ceilometer-0" Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.820542 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xmkcd\" (UniqueName: \"kubernetes.io/projected/d9259b35-21b6-4971-ae1b-04d367d993de-kube-api-access-xmkcd\") pod \"ceilometer-0\" (UID: \"d9259b35-21b6-4971-ae1b-04d367d993de\") " pod="openstack/ceilometer-0" Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.820593 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d9259b35-21b6-4971-ae1b-04d367d993de-scripts\") pod \"ceilometer-0\" (UID: \"d9259b35-21b6-4971-ae1b-04d367d993de\") " pod="openstack/ceilometer-0" Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.820627 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d9259b35-21b6-4971-ae1b-04d367d993de-log-httpd\") pod \"ceilometer-0\" (UID: \"d9259b35-21b6-4971-ae1b-04d367d993de\") " pod="openstack/ceilometer-0" Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.820700 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d9259b35-21b6-4971-ae1b-04d367d993de-run-httpd\") pod \"ceilometer-0\" (UID: \"d9259b35-21b6-4971-ae1b-04d367d993de\") " pod="openstack/ceilometer-0" Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.820799 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9259b35-21b6-4971-ae1b-04d367d993de-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d9259b35-21b6-4971-ae1b-04d367d993de\") " pod="openstack/ceilometer-0" Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.820824 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9259b35-21b6-4971-ae1b-04d367d993de-config-data\") pod \"ceilometer-0\" (UID: \"d9259b35-21b6-4971-ae1b-04d367d993de\") " pod="openstack/ceilometer-0" Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.822735 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d9259b35-21b6-4971-ae1b-04d367d993de-log-httpd\") pod \"ceilometer-0\" (UID: \"d9259b35-21b6-4971-ae1b-04d367d993de\") " pod="openstack/ceilometer-0" Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.825497 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d9259b35-21b6-4971-ae1b-04d367d993de-run-httpd\") pod \"ceilometer-0\" (UID: \"d9259b35-21b6-4971-ae1b-04d367d993de\") " pod="openstack/ceilometer-0" Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.827226 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9259b35-21b6-4971-ae1b-04d367d993de-config-data\") pod \"ceilometer-0\" (UID: \"d9259b35-21b6-4971-ae1b-04d367d993de\") " pod="openstack/ceilometer-0" Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.832016 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d9259b35-21b6-4971-ae1b-04d367d993de-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d9259b35-21b6-4971-ae1b-04d367d993de\") " pod="openstack/ceilometer-0" Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.832440 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d9259b35-21b6-4971-ae1b-04d367d993de-scripts\") pod \"ceilometer-0\" (UID: \"d9259b35-21b6-4971-ae1b-04d367d993de\") " pod="openstack/ceilometer-0" Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.832531 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9259b35-21b6-4971-ae1b-04d367d993de-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d9259b35-21b6-4971-ae1b-04d367d993de\") " pod="openstack/ceilometer-0" Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.850051 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xmkcd\" (UniqueName: \"kubernetes.io/projected/d9259b35-21b6-4971-ae1b-04d367d993de-kube-api-access-xmkcd\") pod \"ceilometer-0\" (UID: \"d9259b35-21b6-4971-ae1b-04d367d993de\") " pod="openstack/ceilometer-0" Dec 11 02:22:40 crc kubenswrapper[4824]: I1211 02:22:40.900423 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 11 02:22:41 crc kubenswrapper[4824]: I1211 02:22:41.161314 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"0dde14f9-85a8-4731-94ef-8199ea6d8a4b","Type":"ContainerStarted","Data":"bffb766da8c7e4a9b4fe7067a5a82cb4edd17ba13d78c32afd1ef5b59eb0292b"} Dec 11 02:22:41 crc kubenswrapper[4824]: I1211 02:22:41.161656 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"0dde14f9-85a8-4731-94ef-8199ea6d8a4b","Type":"ContainerStarted","Data":"60eb1d52701f8a83c4f7109fa0193de14591d041966d444d7e1b98ae58c855a8"} Dec 11 02:22:41 crc kubenswrapper[4824]: W1211 02:22:41.388389 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd9259b35_21b6_4971_ae1b_04d367d993de.slice/crio-98b801d0f37556ebdda6cf42556f67ed6ff84abdd37aaeb5f4e9930bdc620fce WatchSource:0}: Error finding container 98b801d0f37556ebdda6cf42556f67ed6ff84abdd37aaeb5f4e9930bdc620fce: Status 404 returned error can't find the container with id 98b801d0f37556ebdda6cf42556f67ed6ff84abdd37aaeb5f4e9930bdc620fce Dec 11 02:22:41 crc kubenswrapper[4824]: I1211 02:22:41.400767 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 11 02:22:42 crc kubenswrapper[4824]: I1211 02:22:42.186238 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d9259b35-21b6-4971-ae1b-04d367d993de","Type":"ContainerStarted","Data":"98b801d0f37556ebdda6cf42556f67ed6ff84abdd37aaeb5f4e9930bdc620fce"} Dec 11 02:22:42 crc kubenswrapper[4824]: I1211 02:22:42.203834 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"0dde14f9-85a8-4731-94ef-8199ea6d8a4b","Type":"ContainerStarted","Data":"1c2bbcc0d8bc44d1e7f6b8e78ddeb0667be207c2f5ae2dd0b7393c3f8968058d"} Dec 11 02:22:42 crc kubenswrapper[4824]: I1211 02:22:42.230901 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.230883066 podStartE2EDuration="3.230883066s" podCreationTimestamp="2025-12-11 02:22:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:22:42.223652805 +0000 UTC m=+1303.912690184" watchObservedRunningTime="2025-12-11 02:22:42.230883066 +0000 UTC m=+1303.919920445" Dec 11 02:22:43 crc kubenswrapper[4824]: I1211 02:22:43.141728 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5fbc99977b-tggml" Dec 11 02:22:43 crc kubenswrapper[4824]: I1211 02:22:43.220611 4824 generic.go:334] "Generic (PLEG): container finished" podID="0a39253e-8c29-41ae-a242-8c9fccfe2613" containerID="4b69e28249b7383bb629b55abdf4c31bcabee14ef7807ca36836547df2950751" exitCode=0 Dec 11 02:22:43 crc kubenswrapper[4824]: I1211 02:22:43.220700 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5fbc99977b-tggml" event={"ID":"0a39253e-8c29-41ae-a242-8c9fccfe2613","Type":"ContainerDied","Data":"4b69e28249b7383bb629b55abdf4c31bcabee14ef7807ca36836547df2950751"} Dec 11 02:22:43 crc kubenswrapper[4824]: I1211 02:22:43.220710 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5fbc99977b-tggml" Dec 11 02:22:43 crc kubenswrapper[4824]: I1211 02:22:43.220755 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5fbc99977b-tggml" event={"ID":"0a39253e-8c29-41ae-a242-8c9fccfe2613","Type":"ContainerDied","Data":"d9f58bff9914c49685dacc672e47422ad4e80328ce63f58355ad585e4c25788d"} Dec 11 02:22:43 crc kubenswrapper[4824]: I1211 02:22:43.220782 4824 scope.go:117] "RemoveContainer" containerID="4f93a423b9aeeca2bc45f2bbf56b517529c486e408c84b916c2eb9ed430fa56f" Dec 11 02:22:43 crc kubenswrapper[4824]: I1211 02:22:43.230759 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d9259b35-21b6-4971-ae1b-04d367d993de","Type":"ContainerStarted","Data":"e52b4e9af95e8babd1001beb6863b9e7df734a7ad25d0f34810a66eacde1e066"} Dec 11 02:22:43 crc kubenswrapper[4824]: I1211 02:22:43.252490 4824 scope.go:117] "RemoveContainer" containerID="4b69e28249b7383bb629b55abdf4c31bcabee14ef7807ca36836547df2950751" Dec 11 02:22:43 crc kubenswrapper[4824]: I1211 02:22:43.272324 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a39253e-8c29-41ae-a242-8c9fccfe2613-combined-ca-bundle\") pod \"0a39253e-8c29-41ae-a242-8c9fccfe2613\" (UID: \"0a39253e-8c29-41ae-a242-8c9fccfe2613\") " Dec 11 02:22:43 crc kubenswrapper[4824]: I1211 02:22:43.272419 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a39253e-8c29-41ae-a242-8c9fccfe2613-ovndb-tls-certs\") pod \"0a39253e-8c29-41ae-a242-8c9fccfe2613\" (UID: \"0a39253e-8c29-41ae-a242-8c9fccfe2613\") " Dec 11 02:22:43 crc kubenswrapper[4824]: I1211 02:22:43.272475 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/0a39253e-8c29-41ae-a242-8c9fccfe2613-httpd-config\") pod \"0a39253e-8c29-41ae-a242-8c9fccfe2613\" (UID: \"0a39253e-8c29-41ae-a242-8c9fccfe2613\") " Dec 11 02:22:43 crc kubenswrapper[4824]: I1211 02:22:43.272586 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/0a39253e-8c29-41ae-a242-8c9fccfe2613-config\") pod \"0a39253e-8c29-41ae-a242-8c9fccfe2613\" (UID: \"0a39253e-8c29-41ae-a242-8c9fccfe2613\") " Dec 11 02:22:43 crc kubenswrapper[4824]: I1211 02:22:43.272684 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g9bsk\" (UniqueName: \"kubernetes.io/projected/0a39253e-8c29-41ae-a242-8c9fccfe2613-kube-api-access-g9bsk\") pod \"0a39253e-8c29-41ae-a242-8c9fccfe2613\" (UID: \"0a39253e-8c29-41ae-a242-8c9fccfe2613\") " Dec 11 02:22:43 crc kubenswrapper[4824]: I1211 02:22:43.277651 4824 scope.go:117] "RemoveContainer" containerID="4f93a423b9aeeca2bc45f2bbf56b517529c486e408c84b916c2eb9ed430fa56f" Dec 11 02:22:43 crc kubenswrapper[4824]: E1211 02:22:43.278730 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f93a423b9aeeca2bc45f2bbf56b517529c486e408c84b916c2eb9ed430fa56f\": container with ID starting with 4f93a423b9aeeca2bc45f2bbf56b517529c486e408c84b916c2eb9ed430fa56f not found: ID does not exist" containerID="4f93a423b9aeeca2bc45f2bbf56b517529c486e408c84b916c2eb9ed430fa56f" Dec 11 02:22:43 crc kubenswrapper[4824]: I1211 02:22:43.278772 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f93a423b9aeeca2bc45f2bbf56b517529c486e408c84b916c2eb9ed430fa56f"} err="failed to get container status \"4f93a423b9aeeca2bc45f2bbf56b517529c486e408c84b916c2eb9ed430fa56f\": rpc error: code = NotFound desc = could not find container \"4f93a423b9aeeca2bc45f2bbf56b517529c486e408c84b916c2eb9ed430fa56f\": container with ID starting with 4f93a423b9aeeca2bc45f2bbf56b517529c486e408c84b916c2eb9ed430fa56f not found: ID does not exist" Dec 11 02:22:43 crc kubenswrapper[4824]: I1211 02:22:43.278796 4824 scope.go:117] "RemoveContainer" containerID="4b69e28249b7383bb629b55abdf4c31bcabee14ef7807ca36836547df2950751" Dec 11 02:22:43 crc kubenswrapper[4824]: E1211 02:22:43.279237 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4b69e28249b7383bb629b55abdf4c31bcabee14ef7807ca36836547df2950751\": container with ID starting with 4b69e28249b7383bb629b55abdf4c31bcabee14ef7807ca36836547df2950751 not found: ID does not exist" containerID="4b69e28249b7383bb629b55abdf4c31bcabee14ef7807ca36836547df2950751" Dec 11 02:22:43 crc kubenswrapper[4824]: I1211 02:22:43.279259 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b69e28249b7383bb629b55abdf4c31bcabee14ef7807ca36836547df2950751"} err="failed to get container status \"4b69e28249b7383bb629b55abdf4c31bcabee14ef7807ca36836547df2950751\": rpc error: code = NotFound desc = could not find container \"4b69e28249b7383bb629b55abdf4c31bcabee14ef7807ca36836547df2950751\": container with ID starting with 4b69e28249b7383bb629b55abdf4c31bcabee14ef7807ca36836547df2950751 not found: ID does not exist" Dec 11 02:22:43 crc kubenswrapper[4824]: I1211 02:22:43.279906 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a39253e-8c29-41ae-a242-8c9fccfe2613-kube-api-access-g9bsk" (OuterVolumeSpecName: "kube-api-access-g9bsk") pod "0a39253e-8c29-41ae-a242-8c9fccfe2613" (UID: "0a39253e-8c29-41ae-a242-8c9fccfe2613"). InnerVolumeSpecName "kube-api-access-g9bsk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:22:43 crc kubenswrapper[4824]: I1211 02:22:43.291587 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a39253e-8c29-41ae-a242-8c9fccfe2613-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "0a39253e-8c29-41ae-a242-8c9fccfe2613" (UID: "0a39253e-8c29-41ae-a242-8c9fccfe2613"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:22:43 crc kubenswrapper[4824]: I1211 02:22:43.327571 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a39253e-8c29-41ae-a242-8c9fccfe2613-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0a39253e-8c29-41ae-a242-8c9fccfe2613" (UID: "0a39253e-8c29-41ae-a242-8c9fccfe2613"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:22:43 crc kubenswrapper[4824]: I1211 02:22:43.363312 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a39253e-8c29-41ae-a242-8c9fccfe2613-config" (OuterVolumeSpecName: "config") pod "0a39253e-8c29-41ae-a242-8c9fccfe2613" (UID: "0a39253e-8c29-41ae-a242-8c9fccfe2613"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:22:43 crc kubenswrapper[4824]: I1211 02:22:43.376300 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a39253e-8c29-41ae-a242-8c9fccfe2613-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:43 crc kubenswrapper[4824]: I1211 02:22:43.376517 4824 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/0a39253e-8c29-41ae-a242-8c9fccfe2613-httpd-config\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:43 crc kubenswrapper[4824]: I1211 02:22:43.376658 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/0a39253e-8c29-41ae-a242-8c9fccfe2613-config\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:43 crc kubenswrapper[4824]: I1211 02:22:43.376762 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g9bsk\" (UniqueName: \"kubernetes.io/projected/0a39253e-8c29-41ae-a242-8c9fccfe2613-kube-api-access-g9bsk\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:43 crc kubenswrapper[4824]: I1211 02:22:43.385420 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a39253e-8c29-41ae-a242-8c9fccfe2613-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "0a39253e-8c29-41ae-a242-8c9fccfe2613" (UID: "0a39253e-8c29-41ae-a242-8c9fccfe2613"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:22:43 crc kubenswrapper[4824]: I1211 02:22:43.478928 4824 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a39253e-8c29-41ae-a242-8c9fccfe2613-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:43 crc kubenswrapper[4824]: I1211 02:22:43.559619 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-787649dcdd-ts6qs" podUID="f15f9ed0-98f2-42fe-8006-2142098cc2b6" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.147:8443: connect: connection refused" Dec 11 02:22:43 crc kubenswrapper[4824]: I1211 02:22:43.567285 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-5fbc99977b-tggml"] Dec 11 02:22:43 crc kubenswrapper[4824]: I1211 02:22:43.574633 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-5fbc99977b-tggml"] Dec 11 02:22:43 crc kubenswrapper[4824]: I1211 02:22:43.845452 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-6c99457c66-dcp4p" Dec 11 02:22:44 crc kubenswrapper[4824]: I1211 02:22:44.241928 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d9259b35-21b6-4971-ae1b-04d367d993de","Type":"ContainerStarted","Data":"5c21f9201fcef1fe44ad0114460c7e9d419836c72431082208bb4d6c4a529d23"} Dec 11 02:22:44 crc kubenswrapper[4824]: I1211 02:22:44.641874 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a39253e-8c29-41ae-a242-8c9fccfe2613" path="/var/lib/kubelet/pods/0a39253e-8c29-41ae-a242-8c9fccfe2613/volumes" Dec 11 02:22:44 crc kubenswrapper[4824]: I1211 02:22:44.933736 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 11 02:22:45 crc kubenswrapper[4824]: I1211 02:22:45.066035 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 11 02:22:45 crc kubenswrapper[4824]: E1211 02:22:45.066390 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a39253e-8c29-41ae-a242-8c9fccfe2613" containerName="neutron-api" Dec 11 02:22:45 crc kubenswrapper[4824]: I1211 02:22:45.066402 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a39253e-8c29-41ae-a242-8c9fccfe2613" containerName="neutron-api" Dec 11 02:22:45 crc kubenswrapper[4824]: E1211 02:22:45.066414 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a39253e-8c29-41ae-a242-8c9fccfe2613" containerName="neutron-httpd" Dec 11 02:22:45 crc kubenswrapper[4824]: I1211 02:22:45.066421 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a39253e-8c29-41ae-a242-8c9fccfe2613" containerName="neutron-httpd" Dec 11 02:22:45 crc kubenswrapper[4824]: I1211 02:22:45.066575 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a39253e-8c29-41ae-a242-8c9fccfe2613" containerName="neutron-httpd" Dec 11 02:22:45 crc kubenswrapper[4824]: I1211 02:22:45.066594 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a39253e-8c29-41ae-a242-8c9fccfe2613" containerName="neutron-api" Dec 11 02:22:45 crc kubenswrapper[4824]: I1211 02:22:45.067176 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 11 02:22:45 crc kubenswrapper[4824]: I1211 02:22:45.068811 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Dec 11 02:22:45 crc kubenswrapper[4824]: I1211 02:22:45.068818 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-jtrfn" Dec 11 02:22:45 crc kubenswrapper[4824]: I1211 02:22:45.070016 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Dec 11 02:22:45 crc kubenswrapper[4824]: I1211 02:22:45.079380 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 11 02:22:45 crc kubenswrapper[4824]: I1211 02:22:45.211260 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56cb0440-d57a-4237-a034-c23a5b116c74-combined-ca-bundle\") pod \"openstackclient\" (UID: \"56cb0440-d57a-4237-a034-c23a5b116c74\") " pod="openstack/openstackclient" Dec 11 02:22:45 crc kubenswrapper[4824]: I1211 02:22:45.211373 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jbz85\" (UniqueName: \"kubernetes.io/projected/56cb0440-d57a-4237-a034-c23a5b116c74-kube-api-access-jbz85\") pod \"openstackclient\" (UID: \"56cb0440-d57a-4237-a034-c23a5b116c74\") " pod="openstack/openstackclient" Dec 11 02:22:45 crc kubenswrapper[4824]: I1211 02:22:45.211444 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/56cb0440-d57a-4237-a034-c23a5b116c74-openstack-config-secret\") pod \"openstackclient\" (UID: \"56cb0440-d57a-4237-a034-c23a5b116c74\") " pod="openstack/openstackclient" Dec 11 02:22:45 crc kubenswrapper[4824]: I1211 02:22:45.211472 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/56cb0440-d57a-4237-a034-c23a5b116c74-openstack-config\") pod \"openstackclient\" (UID: \"56cb0440-d57a-4237-a034-c23a5b116c74\") " pod="openstack/openstackclient" Dec 11 02:22:45 crc kubenswrapper[4824]: I1211 02:22:45.255554 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d9259b35-21b6-4971-ae1b-04d367d993de","Type":"ContainerStarted","Data":"ed1dc24efac1fb0032a295396091f46900b22703e02321b6a39c6e82be838f84"} Dec 11 02:22:45 crc kubenswrapper[4824]: I1211 02:22:45.313308 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/56cb0440-d57a-4237-a034-c23a5b116c74-openstack-config-secret\") pod \"openstackclient\" (UID: \"56cb0440-d57a-4237-a034-c23a5b116c74\") " pod="openstack/openstackclient" Dec 11 02:22:45 crc kubenswrapper[4824]: I1211 02:22:45.313394 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/56cb0440-d57a-4237-a034-c23a5b116c74-openstack-config\") pod \"openstackclient\" (UID: \"56cb0440-d57a-4237-a034-c23a5b116c74\") " pod="openstack/openstackclient" Dec 11 02:22:45 crc kubenswrapper[4824]: I1211 02:22:45.313512 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56cb0440-d57a-4237-a034-c23a5b116c74-combined-ca-bundle\") pod \"openstackclient\" (UID: \"56cb0440-d57a-4237-a034-c23a5b116c74\") " pod="openstack/openstackclient" Dec 11 02:22:45 crc kubenswrapper[4824]: I1211 02:22:45.313612 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jbz85\" (UniqueName: \"kubernetes.io/projected/56cb0440-d57a-4237-a034-c23a5b116c74-kube-api-access-jbz85\") pod \"openstackclient\" (UID: \"56cb0440-d57a-4237-a034-c23a5b116c74\") " pod="openstack/openstackclient" Dec 11 02:22:45 crc kubenswrapper[4824]: I1211 02:22:45.314563 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/56cb0440-d57a-4237-a034-c23a5b116c74-openstack-config\") pod \"openstackclient\" (UID: \"56cb0440-d57a-4237-a034-c23a5b116c74\") " pod="openstack/openstackclient" Dec 11 02:22:45 crc kubenswrapper[4824]: I1211 02:22:45.320594 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/56cb0440-d57a-4237-a034-c23a5b116c74-openstack-config-secret\") pod \"openstackclient\" (UID: \"56cb0440-d57a-4237-a034-c23a5b116c74\") " pod="openstack/openstackclient" Dec 11 02:22:45 crc kubenswrapper[4824]: I1211 02:22:45.320869 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56cb0440-d57a-4237-a034-c23a5b116c74-combined-ca-bundle\") pod \"openstackclient\" (UID: \"56cb0440-d57a-4237-a034-c23a5b116c74\") " pod="openstack/openstackclient" Dec 11 02:22:45 crc kubenswrapper[4824]: I1211 02:22:45.334656 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jbz85\" (UniqueName: \"kubernetes.io/projected/56cb0440-d57a-4237-a034-c23a5b116c74-kube-api-access-jbz85\") pod \"openstackclient\" (UID: \"56cb0440-d57a-4237-a034-c23a5b116c74\") " pod="openstack/openstackclient" Dec 11 02:22:45 crc kubenswrapper[4824]: I1211 02:22:45.393953 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Dec 11 02:22:45 crc kubenswrapper[4824]: I1211 02:22:45.394597 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 11 02:22:45 crc kubenswrapper[4824]: I1211 02:22:45.401598 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Dec 11 02:22:45 crc kubenswrapper[4824]: I1211 02:22:45.420187 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 11 02:22:45 crc kubenswrapper[4824]: I1211 02:22:45.421233 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 11 02:22:45 crc kubenswrapper[4824]: I1211 02:22:45.430581 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 11 02:22:45 crc kubenswrapper[4824]: I1211 02:22:45.516865 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a66f0fef-8984-4c67-b797-6762b404cd36-openstack-config-secret\") pod \"openstackclient\" (UID: \"a66f0fef-8984-4c67-b797-6762b404cd36\") " pod="openstack/openstackclient" Dec 11 02:22:45 crc kubenswrapper[4824]: I1211 02:22:45.517201 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a66f0fef-8984-4c67-b797-6762b404cd36-combined-ca-bundle\") pod \"openstackclient\" (UID: \"a66f0fef-8984-4c67-b797-6762b404cd36\") " pod="openstack/openstackclient" Dec 11 02:22:45 crc kubenswrapper[4824]: I1211 02:22:45.517259 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9r7bq\" (UniqueName: \"kubernetes.io/projected/a66f0fef-8984-4c67-b797-6762b404cd36-kube-api-access-9r7bq\") pod \"openstackclient\" (UID: \"a66f0fef-8984-4c67-b797-6762b404cd36\") " pod="openstack/openstackclient" Dec 11 02:22:45 crc kubenswrapper[4824]: I1211 02:22:45.517292 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/a66f0fef-8984-4c67-b797-6762b404cd36-openstack-config\") pod \"openstackclient\" (UID: \"a66f0fef-8984-4c67-b797-6762b404cd36\") " pod="openstack/openstackclient" Dec 11 02:22:45 crc kubenswrapper[4824]: E1211 02:22:45.535029 4824 log.go:32] "RunPodSandbox from runtime service failed" err=< Dec 11 02:22:45 crc kubenswrapper[4824]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_56cb0440-d57a-4237-a034-c23a5b116c74_0(413fefe91a2368f272a385833f91ab424b0020b6fc32275a46a529e279460c0e): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"413fefe91a2368f272a385833f91ab424b0020b6fc32275a46a529e279460c0e" Netns:"/var/run/netns/c4b858e2-f21b-4ffb-afe6-58d5b294eb2d" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=413fefe91a2368f272a385833f91ab424b0020b6fc32275a46a529e279460c0e;K8S_POD_UID=56cb0440-d57a-4237-a034-c23a5b116c74" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/56cb0440-d57a-4237-a034-c23a5b116c74]: expected pod UID "56cb0440-d57a-4237-a034-c23a5b116c74" but got "a66f0fef-8984-4c67-b797-6762b404cd36" from Kube API Dec 11 02:22:45 crc kubenswrapper[4824]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 11 02:22:45 crc kubenswrapper[4824]: > Dec 11 02:22:45 crc kubenswrapper[4824]: E1211 02:22:45.535089 4824 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Dec 11 02:22:45 crc kubenswrapper[4824]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_56cb0440-d57a-4237-a034-c23a5b116c74_0(413fefe91a2368f272a385833f91ab424b0020b6fc32275a46a529e279460c0e): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"413fefe91a2368f272a385833f91ab424b0020b6fc32275a46a529e279460c0e" Netns:"/var/run/netns/c4b858e2-f21b-4ffb-afe6-58d5b294eb2d" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=413fefe91a2368f272a385833f91ab424b0020b6fc32275a46a529e279460c0e;K8S_POD_UID=56cb0440-d57a-4237-a034-c23a5b116c74" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/56cb0440-d57a-4237-a034-c23a5b116c74]: expected pod UID "56cb0440-d57a-4237-a034-c23a5b116c74" but got "a66f0fef-8984-4c67-b797-6762b404cd36" from Kube API Dec 11 02:22:45 crc kubenswrapper[4824]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 11 02:22:45 crc kubenswrapper[4824]: > pod="openstack/openstackclient" Dec 11 02:22:45 crc kubenswrapper[4824]: I1211 02:22:45.618481 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a66f0fef-8984-4c67-b797-6762b404cd36-openstack-config-secret\") pod \"openstackclient\" (UID: \"a66f0fef-8984-4c67-b797-6762b404cd36\") " pod="openstack/openstackclient" Dec 11 02:22:45 crc kubenswrapper[4824]: I1211 02:22:45.618584 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a66f0fef-8984-4c67-b797-6762b404cd36-combined-ca-bundle\") pod \"openstackclient\" (UID: \"a66f0fef-8984-4c67-b797-6762b404cd36\") " pod="openstack/openstackclient" Dec 11 02:22:45 crc kubenswrapper[4824]: I1211 02:22:45.618643 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9r7bq\" (UniqueName: \"kubernetes.io/projected/a66f0fef-8984-4c67-b797-6762b404cd36-kube-api-access-9r7bq\") pod \"openstackclient\" (UID: \"a66f0fef-8984-4c67-b797-6762b404cd36\") " pod="openstack/openstackclient" Dec 11 02:22:45 crc kubenswrapper[4824]: I1211 02:22:45.618681 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/a66f0fef-8984-4c67-b797-6762b404cd36-openstack-config\") pod \"openstackclient\" (UID: \"a66f0fef-8984-4c67-b797-6762b404cd36\") " pod="openstack/openstackclient" Dec 11 02:22:45 crc kubenswrapper[4824]: I1211 02:22:45.619643 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/a66f0fef-8984-4c67-b797-6762b404cd36-openstack-config\") pod \"openstackclient\" (UID: \"a66f0fef-8984-4c67-b797-6762b404cd36\") " pod="openstack/openstackclient" Dec 11 02:22:45 crc kubenswrapper[4824]: I1211 02:22:45.623188 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a66f0fef-8984-4c67-b797-6762b404cd36-combined-ca-bundle\") pod \"openstackclient\" (UID: \"a66f0fef-8984-4c67-b797-6762b404cd36\") " pod="openstack/openstackclient" Dec 11 02:22:45 crc kubenswrapper[4824]: I1211 02:22:45.623391 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a66f0fef-8984-4c67-b797-6762b404cd36-openstack-config-secret\") pod \"openstackclient\" (UID: \"a66f0fef-8984-4c67-b797-6762b404cd36\") " pod="openstack/openstackclient" Dec 11 02:22:45 crc kubenswrapper[4824]: I1211 02:22:45.635276 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9r7bq\" (UniqueName: \"kubernetes.io/projected/a66f0fef-8984-4c67-b797-6762b404cd36-kube-api-access-9r7bq\") pod \"openstackclient\" (UID: \"a66f0fef-8984-4c67-b797-6762b404cd36\") " pod="openstack/openstackclient" Dec 11 02:22:45 crc kubenswrapper[4824]: I1211 02:22:45.894983 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 11 02:22:46 crc kubenswrapper[4824]: I1211 02:22:46.286409 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 11 02:22:46 crc kubenswrapper[4824]: I1211 02:22:46.290023 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d9259b35-21b6-4971-ae1b-04d367d993de","Type":"ContainerStarted","Data":"6a65d85a224360efdc1c1ccd3fb02c778ea2c034d15593e608fd2424c05be1ed"} Dec 11 02:22:46 crc kubenswrapper[4824]: I1211 02:22:46.290060 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 11 02:22:46 crc kubenswrapper[4824]: I1211 02:22:46.292898 4824 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="56cb0440-d57a-4237-a034-c23a5b116c74" podUID="a66f0fef-8984-4c67-b797-6762b404cd36" Dec 11 02:22:46 crc kubenswrapper[4824]: I1211 02:22:46.297860 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 11 02:22:46 crc kubenswrapper[4824]: I1211 02:22:46.341453 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 11 02:22:46 crc kubenswrapper[4824]: I1211 02:22:46.342712 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.023715445 podStartE2EDuration="6.342696644s" podCreationTimestamp="2025-12-11 02:22:40 +0000 UTC" firstStartedPulling="2025-12-11 02:22:41.391639695 +0000 UTC m=+1303.080677114" lastFinishedPulling="2025-12-11 02:22:45.710620934 +0000 UTC m=+1307.399658313" observedRunningTime="2025-12-11 02:22:46.335491143 +0000 UTC m=+1308.024528522" watchObservedRunningTime="2025-12-11 02:22:46.342696644 +0000 UTC m=+1308.031734023" Dec 11 02:22:46 crc kubenswrapper[4824]: I1211 02:22:46.421834 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 11 02:22:46 crc kubenswrapper[4824]: I1211 02:22:46.437222 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jbz85\" (UniqueName: \"kubernetes.io/projected/56cb0440-d57a-4237-a034-c23a5b116c74-kube-api-access-jbz85\") pod \"56cb0440-d57a-4237-a034-c23a5b116c74\" (UID: \"56cb0440-d57a-4237-a034-c23a5b116c74\") " Dec 11 02:22:46 crc kubenswrapper[4824]: I1211 02:22:46.437271 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/56cb0440-d57a-4237-a034-c23a5b116c74-openstack-config-secret\") pod \"56cb0440-d57a-4237-a034-c23a5b116c74\" (UID: \"56cb0440-d57a-4237-a034-c23a5b116c74\") " Dec 11 02:22:46 crc kubenswrapper[4824]: I1211 02:22:46.437461 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56cb0440-d57a-4237-a034-c23a5b116c74-combined-ca-bundle\") pod \"56cb0440-d57a-4237-a034-c23a5b116c74\" (UID: \"56cb0440-d57a-4237-a034-c23a5b116c74\") " Dec 11 02:22:46 crc kubenswrapper[4824]: I1211 02:22:46.437526 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/56cb0440-d57a-4237-a034-c23a5b116c74-openstack-config\") pod \"56cb0440-d57a-4237-a034-c23a5b116c74\" (UID: \"56cb0440-d57a-4237-a034-c23a5b116c74\") " Dec 11 02:22:46 crc kubenswrapper[4824]: I1211 02:22:46.437988 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/56cb0440-d57a-4237-a034-c23a5b116c74-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "56cb0440-d57a-4237-a034-c23a5b116c74" (UID: "56cb0440-d57a-4237-a034-c23a5b116c74"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:22:46 crc kubenswrapper[4824]: I1211 02:22:46.445599 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56cb0440-d57a-4237-a034-c23a5b116c74-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "56cb0440-d57a-4237-a034-c23a5b116c74" (UID: "56cb0440-d57a-4237-a034-c23a5b116c74"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:22:46 crc kubenswrapper[4824]: I1211 02:22:46.445717 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56cb0440-d57a-4237-a034-c23a5b116c74-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "56cb0440-d57a-4237-a034-c23a5b116c74" (UID: "56cb0440-d57a-4237-a034-c23a5b116c74"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:22:46 crc kubenswrapper[4824]: I1211 02:22:46.465229 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56cb0440-d57a-4237-a034-c23a5b116c74-kube-api-access-jbz85" (OuterVolumeSpecName: "kube-api-access-jbz85") pod "56cb0440-d57a-4237-a034-c23a5b116c74" (UID: "56cb0440-d57a-4237-a034-c23a5b116c74"). InnerVolumeSpecName "kube-api-access-jbz85". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:22:46 crc kubenswrapper[4824]: I1211 02:22:46.539527 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jbz85\" (UniqueName: \"kubernetes.io/projected/56cb0440-d57a-4237-a034-c23a5b116c74-kube-api-access-jbz85\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:46 crc kubenswrapper[4824]: I1211 02:22:46.539568 4824 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/56cb0440-d57a-4237-a034-c23a5b116c74-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:46 crc kubenswrapper[4824]: I1211 02:22:46.539580 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56cb0440-d57a-4237-a034-c23a5b116c74-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:46 crc kubenswrapper[4824]: I1211 02:22:46.539592 4824 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/56cb0440-d57a-4237-a034-c23a5b116c74-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:46 crc kubenswrapper[4824]: I1211 02:22:46.641056 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56cb0440-d57a-4237-a034-c23a5b116c74" path="/var/lib/kubelet/pods/56cb0440-d57a-4237-a034-c23a5b116c74/volumes" Dec 11 02:22:47 crc kubenswrapper[4824]: I1211 02:22:47.300449 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 11 02:22:47 crc kubenswrapper[4824]: I1211 02:22:47.300442 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"a66f0fef-8984-4c67-b797-6762b404cd36","Type":"ContainerStarted","Data":"f47d77ae265f7a334b140bf704bd82a8f734312f36a1f8b9b4efa3b60f74f32a"} Dec 11 02:22:47 crc kubenswrapper[4824]: I1211 02:22:47.307558 4824 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="56cb0440-d57a-4237-a034-c23a5b116c74" podUID="a66f0fef-8984-4c67-b797-6762b404cd36" Dec 11 02:22:48 crc kubenswrapper[4824]: I1211 02:22:48.448183 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-6458cc7487-d2555"] Dec 11 02:22:48 crc kubenswrapper[4824]: I1211 02:22:48.450646 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-6458cc7487-d2555" Dec 11 02:22:48 crc kubenswrapper[4824]: I1211 02:22:48.456048 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Dec 11 02:22:48 crc kubenswrapper[4824]: I1211 02:22:48.456102 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Dec 11 02:22:48 crc kubenswrapper[4824]: I1211 02:22:48.458036 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 11 02:22:48 crc kubenswrapper[4824]: I1211 02:22:48.465654 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-6458cc7487-d2555"] Dec 11 02:22:48 crc kubenswrapper[4824]: I1211 02:22:48.578203 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ead63438-3159-4a5e-889f-5f7c8a8d6a30-combined-ca-bundle\") pod \"swift-proxy-6458cc7487-d2555\" (UID: \"ead63438-3159-4a5e-889f-5f7c8a8d6a30\") " pod="openstack/swift-proxy-6458cc7487-d2555" Dec 11 02:22:48 crc kubenswrapper[4824]: I1211 02:22:48.578257 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ead63438-3159-4a5e-889f-5f7c8a8d6a30-internal-tls-certs\") pod \"swift-proxy-6458cc7487-d2555\" (UID: \"ead63438-3159-4a5e-889f-5f7c8a8d6a30\") " pod="openstack/swift-proxy-6458cc7487-d2555" Dec 11 02:22:48 crc kubenswrapper[4824]: I1211 02:22:48.578290 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ead63438-3159-4a5e-889f-5f7c8a8d6a30-config-data\") pod \"swift-proxy-6458cc7487-d2555\" (UID: \"ead63438-3159-4a5e-889f-5f7c8a8d6a30\") " pod="openstack/swift-proxy-6458cc7487-d2555" Dec 11 02:22:48 crc kubenswrapper[4824]: I1211 02:22:48.578308 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xldhm\" (UniqueName: \"kubernetes.io/projected/ead63438-3159-4a5e-889f-5f7c8a8d6a30-kube-api-access-xldhm\") pod \"swift-proxy-6458cc7487-d2555\" (UID: \"ead63438-3159-4a5e-889f-5f7c8a8d6a30\") " pod="openstack/swift-proxy-6458cc7487-d2555" Dec 11 02:22:48 crc kubenswrapper[4824]: I1211 02:22:48.578324 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ead63438-3159-4a5e-889f-5f7c8a8d6a30-run-httpd\") pod \"swift-proxy-6458cc7487-d2555\" (UID: \"ead63438-3159-4a5e-889f-5f7c8a8d6a30\") " pod="openstack/swift-proxy-6458cc7487-d2555" Dec 11 02:22:48 crc kubenswrapper[4824]: I1211 02:22:48.578342 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ead63438-3159-4a5e-889f-5f7c8a8d6a30-public-tls-certs\") pod \"swift-proxy-6458cc7487-d2555\" (UID: \"ead63438-3159-4a5e-889f-5f7c8a8d6a30\") " pod="openstack/swift-proxy-6458cc7487-d2555" Dec 11 02:22:48 crc kubenswrapper[4824]: I1211 02:22:48.578367 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/ead63438-3159-4a5e-889f-5f7c8a8d6a30-etc-swift\") pod \"swift-proxy-6458cc7487-d2555\" (UID: \"ead63438-3159-4a5e-889f-5f7c8a8d6a30\") " pod="openstack/swift-proxy-6458cc7487-d2555" Dec 11 02:22:48 crc kubenswrapper[4824]: I1211 02:22:48.578437 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ead63438-3159-4a5e-889f-5f7c8a8d6a30-log-httpd\") pod \"swift-proxy-6458cc7487-d2555\" (UID: \"ead63438-3159-4a5e-889f-5f7c8a8d6a30\") " pod="openstack/swift-proxy-6458cc7487-d2555" Dec 11 02:22:48 crc kubenswrapper[4824]: I1211 02:22:48.679670 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ead63438-3159-4a5e-889f-5f7c8a8d6a30-internal-tls-certs\") pod \"swift-proxy-6458cc7487-d2555\" (UID: \"ead63438-3159-4a5e-889f-5f7c8a8d6a30\") " pod="openstack/swift-proxy-6458cc7487-d2555" Dec 11 02:22:48 crc kubenswrapper[4824]: I1211 02:22:48.679743 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ead63438-3159-4a5e-889f-5f7c8a8d6a30-config-data\") pod \"swift-proxy-6458cc7487-d2555\" (UID: \"ead63438-3159-4a5e-889f-5f7c8a8d6a30\") " pod="openstack/swift-proxy-6458cc7487-d2555" Dec 11 02:22:48 crc kubenswrapper[4824]: I1211 02:22:48.679765 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xldhm\" (UniqueName: \"kubernetes.io/projected/ead63438-3159-4a5e-889f-5f7c8a8d6a30-kube-api-access-xldhm\") pod \"swift-proxy-6458cc7487-d2555\" (UID: \"ead63438-3159-4a5e-889f-5f7c8a8d6a30\") " pod="openstack/swift-proxy-6458cc7487-d2555" Dec 11 02:22:48 crc kubenswrapper[4824]: I1211 02:22:48.679781 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ead63438-3159-4a5e-889f-5f7c8a8d6a30-run-httpd\") pod \"swift-proxy-6458cc7487-d2555\" (UID: \"ead63438-3159-4a5e-889f-5f7c8a8d6a30\") " pod="openstack/swift-proxy-6458cc7487-d2555" Dec 11 02:22:48 crc kubenswrapper[4824]: I1211 02:22:48.679807 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ead63438-3159-4a5e-889f-5f7c8a8d6a30-public-tls-certs\") pod \"swift-proxy-6458cc7487-d2555\" (UID: \"ead63438-3159-4a5e-889f-5f7c8a8d6a30\") " pod="openstack/swift-proxy-6458cc7487-d2555" Dec 11 02:22:48 crc kubenswrapper[4824]: I1211 02:22:48.679832 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/ead63438-3159-4a5e-889f-5f7c8a8d6a30-etc-swift\") pod \"swift-proxy-6458cc7487-d2555\" (UID: \"ead63438-3159-4a5e-889f-5f7c8a8d6a30\") " pod="openstack/swift-proxy-6458cc7487-d2555" Dec 11 02:22:48 crc kubenswrapper[4824]: I1211 02:22:48.679883 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ead63438-3159-4a5e-889f-5f7c8a8d6a30-log-httpd\") pod \"swift-proxy-6458cc7487-d2555\" (UID: \"ead63438-3159-4a5e-889f-5f7c8a8d6a30\") " pod="openstack/swift-proxy-6458cc7487-d2555" Dec 11 02:22:48 crc kubenswrapper[4824]: I1211 02:22:48.679959 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ead63438-3159-4a5e-889f-5f7c8a8d6a30-combined-ca-bundle\") pod \"swift-proxy-6458cc7487-d2555\" (UID: \"ead63438-3159-4a5e-889f-5f7c8a8d6a30\") " pod="openstack/swift-proxy-6458cc7487-d2555" Dec 11 02:22:48 crc kubenswrapper[4824]: I1211 02:22:48.681069 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ead63438-3159-4a5e-889f-5f7c8a8d6a30-run-httpd\") pod \"swift-proxy-6458cc7487-d2555\" (UID: \"ead63438-3159-4a5e-889f-5f7c8a8d6a30\") " pod="openstack/swift-proxy-6458cc7487-d2555" Dec 11 02:22:48 crc kubenswrapper[4824]: I1211 02:22:48.682921 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ead63438-3159-4a5e-889f-5f7c8a8d6a30-log-httpd\") pod \"swift-proxy-6458cc7487-d2555\" (UID: \"ead63438-3159-4a5e-889f-5f7c8a8d6a30\") " pod="openstack/swift-proxy-6458cc7487-d2555" Dec 11 02:22:48 crc kubenswrapper[4824]: I1211 02:22:48.686376 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ead63438-3159-4a5e-889f-5f7c8a8d6a30-internal-tls-certs\") pod \"swift-proxy-6458cc7487-d2555\" (UID: \"ead63438-3159-4a5e-889f-5f7c8a8d6a30\") " pod="openstack/swift-proxy-6458cc7487-d2555" Dec 11 02:22:48 crc kubenswrapper[4824]: I1211 02:22:48.687637 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ead63438-3159-4a5e-889f-5f7c8a8d6a30-config-data\") pod \"swift-proxy-6458cc7487-d2555\" (UID: \"ead63438-3159-4a5e-889f-5f7c8a8d6a30\") " pod="openstack/swift-proxy-6458cc7487-d2555" Dec 11 02:22:48 crc kubenswrapper[4824]: I1211 02:22:48.688232 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ead63438-3159-4a5e-889f-5f7c8a8d6a30-public-tls-certs\") pod \"swift-proxy-6458cc7487-d2555\" (UID: \"ead63438-3159-4a5e-889f-5f7c8a8d6a30\") " pod="openstack/swift-proxy-6458cc7487-d2555" Dec 11 02:22:48 crc kubenswrapper[4824]: I1211 02:22:48.689038 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ead63438-3159-4a5e-889f-5f7c8a8d6a30-combined-ca-bundle\") pod \"swift-proxy-6458cc7487-d2555\" (UID: \"ead63438-3159-4a5e-889f-5f7c8a8d6a30\") " pod="openstack/swift-proxy-6458cc7487-d2555" Dec 11 02:22:48 crc kubenswrapper[4824]: I1211 02:22:48.703097 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xldhm\" (UniqueName: \"kubernetes.io/projected/ead63438-3159-4a5e-889f-5f7c8a8d6a30-kube-api-access-xldhm\") pod \"swift-proxy-6458cc7487-d2555\" (UID: \"ead63438-3159-4a5e-889f-5f7c8a8d6a30\") " pod="openstack/swift-proxy-6458cc7487-d2555" Dec 11 02:22:48 crc kubenswrapper[4824]: I1211 02:22:48.703192 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/ead63438-3159-4a5e-889f-5f7c8a8d6a30-etc-swift\") pod \"swift-proxy-6458cc7487-d2555\" (UID: \"ead63438-3159-4a5e-889f-5f7c8a8d6a30\") " pod="openstack/swift-proxy-6458cc7487-d2555" Dec 11 02:22:48 crc kubenswrapper[4824]: I1211 02:22:48.765697 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-6458cc7487-d2555" Dec 11 02:22:49 crc kubenswrapper[4824]: I1211 02:22:49.341908 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-6458cc7487-d2555"] Dec 11 02:22:50 crc kubenswrapper[4824]: I1211 02:22:50.090700 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 11 02:22:50 crc kubenswrapper[4824]: I1211 02:22:50.091477 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d9259b35-21b6-4971-ae1b-04d367d993de" containerName="ceilometer-central-agent" containerID="cri-o://e52b4e9af95e8babd1001beb6863b9e7df734a7ad25d0f34810a66eacde1e066" gracePeriod=30 Dec 11 02:22:50 crc kubenswrapper[4824]: I1211 02:22:50.091544 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d9259b35-21b6-4971-ae1b-04d367d993de" containerName="ceilometer-notification-agent" containerID="cri-o://5c21f9201fcef1fe44ad0114460c7e9d419836c72431082208bb4d6c4a529d23" gracePeriod=30 Dec 11 02:22:50 crc kubenswrapper[4824]: I1211 02:22:50.091544 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d9259b35-21b6-4971-ae1b-04d367d993de" containerName="proxy-httpd" containerID="cri-o://6a65d85a224360efdc1c1ccd3fb02c778ea2c034d15593e608fd2424c05be1ed" gracePeriod=30 Dec 11 02:22:50 crc kubenswrapper[4824]: I1211 02:22:50.091546 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d9259b35-21b6-4971-ae1b-04d367d993de" containerName="sg-core" containerID="cri-o://ed1dc24efac1fb0032a295396091f46900b22703e02321b6a39c6e82be838f84" gracePeriod=30 Dec 11 02:22:50 crc kubenswrapper[4824]: I1211 02:22:50.208878 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 11 02:22:50 crc kubenswrapper[4824]: I1211 02:22:50.333477 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6458cc7487-d2555" event={"ID":"ead63438-3159-4a5e-889f-5f7c8a8d6a30","Type":"ContainerStarted","Data":"ef41f3205405b306b477a5182d65c7e61706ccb6b82b008d37902eb7de98ec9e"} Dec 11 02:22:50 crc kubenswrapper[4824]: I1211 02:22:50.333528 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6458cc7487-d2555" event={"ID":"ead63438-3159-4a5e-889f-5f7c8a8d6a30","Type":"ContainerStarted","Data":"effeb8bcbaa3a59da5fd52e97fdd9c9214064dadc2ceae3d04d7212ec43368a6"} Dec 11 02:22:51 crc kubenswrapper[4824]: I1211 02:22:51.401764 4824 generic.go:334] "Generic (PLEG): container finished" podID="d9259b35-21b6-4971-ae1b-04d367d993de" containerID="6a65d85a224360efdc1c1ccd3fb02c778ea2c034d15593e608fd2424c05be1ed" exitCode=0 Dec 11 02:22:51 crc kubenswrapper[4824]: I1211 02:22:51.402073 4824 generic.go:334] "Generic (PLEG): container finished" podID="d9259b35-21b6-4971-ae1b-04d367d993de" containerID="ed1dc24efac1fb0032a295396091f46900b22703e02321b6a39c6e82be838f84" exitCode=2 Dec 11 02:22:51 crc kubenswrapper[4824]: I1211 02:22:51.402083 4824 generic.go:334] "Generic (PLEG): container finished" podID="d9259b35-21b6-4971-ae1b-04d367d993de" containerID="5c21f9201fcef1fe44ad0114460c7e9d419836c72431082208bb4d6c4a529d23" exitCode=0 Dec 11 02:22:51 crc kubenswrapper[4824]: I1211 02:22:51.402090 4824 generic.go:334] "Generic (PLEG): container finished" podID="d9259b35-21b6-4971-ae1b-04d367d993de" containerID="e52b4e9af95e8babd1001beb6863b9e7df734a7ad25d0f34810a66eacde1e066" exitCode=0 Dec 11 02:22:51 crc kubenswrapper[4824]: I1211 02:22:51.401827 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d9259b35-21b6-4971-ae1b-04d367d993de","Type":"ContainerDied","Data":"6a65d85a224360efdc1c1ccd3fb02c778ea2c034d15593e608fd2424c05be1ed"} Dec 11 02:22:51 crc kubenswrapper[4824]: I1211 02:22:51.402175 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d9259b35-21b6-4971-ae1b-04d367d993de","Type":"ContainerDied","Data":"ed1dc24efac1fb0032a295396091f46900b22703e02321b6a39c6e82be838f84"} Dec 11 02:22:51 crc kubenswrapper[4824]: I1211 02:22:51.402190 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d9259b35-21b6-4971-ae1b-04d367d993de","Type":"ContainerDied","Data":"5c21f9201fcef1fe44ad0114460c7e9d419836c72431082208bb4d6c4a529d23"} Dec 11 02:22:51 crc kubenswrapper[4824]: I1211 02:22:51.402200 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d9259b35-21b6-4971-ae1b-04d367d993de","Type":"ContainerDied","Data":"e52b4e9af95e8babd1001beb6863b9e7df734a7ad25d0f34810a66eacde1e066"} Dec 11 02:22:51 crc kubenswrapper[4824]: I1211 02:22:51.403633 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6458cc7487-d2555" event={"ID":"ead63438-3159-4a5e-889f-5f7c8a8d6a30","Type":"ContainerStarted","Data":"82201a036d2fb043f5a4d368909398b890e53c28d9f027970bd76fc6ba964eb0"} Dec 11 02:22:51 crc kubenswrapper[4824]: I1211 02:22:51.403744 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-6458cc7487-d2555" Dec 11 02:22:51 crc kubenswrapper[4824]: I1211 02:22:51.403770 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-6458cc7487-d2555" Dec 11 02:22:53 crc kubenswrapper[4824]: I1211 02:22:53.559391 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-787649dcdd-ts6qs" podUID="f15f9ed0-98f2-42fe-8006-2142098cc2b6" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.147:8443: connect: connection refused" Dec 11 02:22:56 crc kubenswrapper[4824]: I1211 02:22:56.700185 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d9259b35-21b6-4971-ae1b-04d367d993de","Type":"ContainerDied","Data":"98b801d0f37556ebdda6cf42556f67ed6ff84abdd37aaeb5f4e9930bdc620fce"} Dec 11 02:22:56 crc kubenswrapper[4824]: I1211 02:22:56.700688 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="98b801d0f37556ebdda6cf42556f67ed6ff84abdd37aaeb5f4e9930bdc620fce" Dec 11 02:22:56 crc kubenswrapper[4824]: I1211 02:22:56.704322 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 11 02:22:56 crc kubenswrapper[4824]: I1211 02:22:56.738015 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-6458cc7487-d2555" podStartSLOduration=8.737997086 podStartE2EDuration="8.737997086s" podCreationTimestamp="2025-12-11 02:22:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:22:51.434263431 +0000 UTC m=+1313.123300800" watchObservedRunningTime="2025-12-11 02:22:56.737997086 +0000 UTC m=+1318.427034465" Dec 11 02:22:56 crc kubenswrapper[4824]: I1211 02:22:56.754070 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9259b35-21b6-4971-ae1b-04d367d993de-combined-ca-bundle\") pod \"d9259b35-21b6-4971-ae1b-04d367d993de\" (UID: \"d9259b35-21b6-4971-ae1b-04d367d993de\") " Dec 11 02:22:56 crc kubenswrapper[4824]: I1211 02:22:56.754378 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d9259b35-21b6-4971-ae1b-04d367d993de-sg-core-conf-yaml\") pod \"d9259b35-21b6-4971-ae1b-04d367d993de\" (UID: \"d9259b35-21b6-4971-ae1b-04d367d993de\") " Dec 11 02:22:56 crc kubenswrapper[4824]: I1211 02:22:56.754507 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9259b35-21b6-4971-ae1b-04d367d993de-config-data\") pod \"d9259b35-21b6-4971-ae1b-04d367d993de\" (UID: \"d9259b35-21b6-4971-ae1b-04d367d993de\") " Dec 11 02:22:56 crc kubenswrapper[4824]: I1211 02:22:56.754692 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d9259b35-21b6-4971-ae1b-04d367d993de-log-httpd\") pod \"d9259b35-21b6-4971-ae1b-04d367d993de\" (UID: \"d9259b35-21b6-4971-ae1b-04d367d993de\") " Dec 11 02:22:56 crc kubenswrapper[4824]: I1211 02:22:56.754856 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d9259b35-21b6-4971-ae1b-04d367d993de-scripts\") pod \"d9259b35-21b6-4971-ae1b-04d367d993de\" (UID: \"d9259b35-21b6-4971-ae1b-04d367d993de\") " Dec 11 02:22:56 crc kubenswrapper[4824]: I1211 02:22:56.754960 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d9259b35-21b6-4971-ae1b-04d367d993de-run-httpd\") pod \"d9259b35-21b6-4971-ae1b-04d367d993de\" (UID: \"d9259b35-21b6-4971-ae1b-04d367d993de\") " Dec 11 02:22:56 crc kubenswrapper[4824]: I1211 02:22:56.755087 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xmkcd\" (UniqueName: \"kubernetes.io/projected/d9259b35-21b6-4971-ae1b-04d367d993de-kube-api-access-xmkcd\") pod \"d9259b35-21b6-4971-ae1b-04d367d993de\" (UID: \"d9259b35-21b6-4971-ae1b-04d367d993de\") " Dec 11 02:22:56 crc kubenswrapper[4824]: I1211 02:22:56.766482 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d9259b35-21b6-4971-ae1b-04d367d993de-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "d9259b35-21b6-4971-ae1b-04d367d993de" (UID: "d9259b35-21b6-4971-ae1b-04d367d993de"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:22:56 crc kubenswrapper[4824]: I1211 02:22:56.766598 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d9259b35-21b6-4971-ae1b-04d367d993de-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "d9259b35-21b6-4971-ae1b-04d367d993de" (UID: "d9259b35-21b6-4971-ae1b-04d367d993de"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:22:56 crc kubenswrapper[4824]: I1211 02:22:56.778385 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9259b35-21b6-4971-ae1b-04d367d993de-scripts" (OuterVolumeSpecName: "scripts") pod "d9259b35-21b6-4971-ae1b-04d367d993de" (UID: "d9259b35-21b6-4971-ae1b-04d367d993de"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:22:56 crc kubenswrapper[4824]: I1211 02:22:56.779476 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9259b35-21b6-4971-ae1b-04d367d993de-kube-api-access-xmkcd" (OuterVolumeSpecName: "kube-api-access-xmkcd") pod "d9259b35-21b6-4971-ae1b-04d367d993de" (UID: "d9259b35-21b6-4971-ae1b-04d367d993de"). InnerVolumeSpecName "kube-api-access-xmkcd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:22:56 crc kubenswrapper[4824]: I1211 02:22:56.794240 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9259b35-21b6-4971-ae1b-04d367d993de-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "d9259b35-21b6-4971-ae1b-04d367d993de" (UID: "d9259b35-21b6-4971-ae1b-04d367d993de"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:22:56 crc kubenswrapper[4824]: I1211 02:22:56.862316 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d9259b35-21b6-4971-ae1b-04d367d993de-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:56 crc kubenswrapper[4824]: I1211 02:22:56.862353 4824 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d9259b35-21b6-4971-ae1b-04d367d993de-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:56 crc kubenswrapper[4824]: I1211 02:22:56.862368 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xmkcd\" (UniqueName: \"kubernetes.io/projected/d9259b35-21b6-4971-ae1b-04d367d993de-kube-api-access-xmkcd\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:56 crc kubenswrapper[4824]: I1211 02:22:56.862383 4824 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d9259b35-21b6-4971-ae1b-04d367d993de-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:56 crc kubenswrapper[4824]: I1211 02:22:56.862393 4824 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d9259b35-21b6-4971-ae1b-04d367d993de-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:56 crc kubenswrapper[4824]: I1211 02:22:56.872709 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9259b35-21b6-4971-ae1b-04d367d993de-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d9259b35-21b6-4971-ae1b-04d367d993de" (UID: "d9259b35-21b6-4971-ae1b-04d367d993de"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:22:56 crc kubenswrapper[4824]: I1211 02:22:56.904082 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9259b35-21b6-4971-ae1b-04d367d993de-config-data" (OuterVolumeSpecName: "config-data") pod "d9259b35-21b6-4971-ae1b-04d367d993de" (UID: "d9259b35-21b6-4971-ae1b-04d367d993de"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:22:56 crc kubenswrapper[4824]: I1211 02:22:56.963450 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9259b35-21b6-4971-ae1b-04d367d993de-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:56 crc kubenswrapper[4824]: I1211 02:22:56.963480 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9259b35-21b6-4971-ae1b-04d367d993de-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:57 crc kubenswrapper[4824]: I1211 02:22:57.095876 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 11 02:22:57 crc kubenswrapper[4824]: I1211 02:22:57.096094 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="ee999ccf-334d-49ff-85bf-9c6601b97bf1" containerName="glance-log" containerID="cri-o://2fa8b7e2e0ed1d8d5b5c08d6ab83a548819dfb7841d471d788f368f2ca2743ae" gracePeriod=30 Dec 11 02:22:57 crc kubenswrapper[4824]: I1211 02:22:57.096308 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="ee999ccf-334d-49ff-85bf-9c6601b97bf1" containerName="glance-httpd" containerID="cri-o://22b549fc7d47ca2a6cc9968758dbb602d55f41fdce561681cce9f4f80206b283" gracePeriod=30 Dec 11 02:22:57 crc kubenswrapper[4824]: I1211 02:22:57.678780 4824 generic.go:334] "Generic (PLEG): container finished" podID="ee999ccf-334d-49ff-85bf-9c6601b97bf1" containerID="2fa8b7e2e0ed1d8d5b5c08d6ab83a548819dfb7841d471d788f368f2ca2743ae" exitCode=143 Dec 11 02:22:57 crc kubenswrapper[4824]: I1211 02:22:57.678905 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ee999ccf-334d-49ff-85bf-9c6601b97bf1","Type":"ContainerDied","Data":"2fa8b7e2e0ed1d8d5b5c08d6ab83a548819dfb7841d471d788f368f2ca2743ae"} Dec 11 02:22:57 crc kubenswrapper[4824]: I1211 02:22:57.681140 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"a66f0fef-8984-4c67-b797-6762b404cd36","Type":"ContainerStarted","Data":"312ca37bdd9cca5d73eb657c9d2e0177c18037ccf30253acf0498925022ff047"} Dec 11 02:22:57 crc kubenswrapper[4824]: I1211 02:22:57.681272 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 11 02:22:57 crc kubenswrapper[4824]: I1211 02:22:57.720943 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=3.62192255 podStartE2EDuration="12.720918855s" podCreationTimestamp="2025-12-11 02:22:45 +0000 UTC" firstStartedPulling="2025-12-11 02:22:46.42853989 +0000 UTC m=+1308.117577269" lastFinishedPulling="2025-12-11 02:22:55.527536155 +0000 UTC m=+1317.216573574" observedRunningTime="2025-12-11 02:22:57.707516669 +0000 UTC m=+1319.396554058" watchObservedRunningTime="2025-12-11 02:22:57.720918855 +0000 UTC m=+1319.409956264" Dec 11 02:22:57 crc kubenswrapper[4824]: I1211 02:22:57.735313 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 11 02:22:57 crc kubenswrapper[4824]: I1211 02:22:57.744753 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 11 02:22:57 crc kubenswrapper[4824]: I1211 02:22:57.762687 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 11 02:22:57 crc kubenswrapper[4824]: E1211 02:22:57.763024 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9259b35-21b6-4971-ae1b-04d367d993de" containerName="proxy-httpd" Dec 11 02:22:57 crc kubenswrapper[4824]: I1211 02:22:57.763040 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9259b35-21b6-4971-ae1b-04d367d993de" containerName="proxy-httpd" Dec 11 02:22:57 crc kubenswrapper[4824]: E1211 02:22:57.763051 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9259b35-21b6-4971-ae1b-04d367d993de" containerName="ceilometer-notification-agent" Dec 11 02:22:57 crc kubenswrapper[4824]: I1211 02:22:57.763057 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9259b35-21b6-4971-ae1b-04d367d993de" containerName="ceilometer-notification-agent" Dec 11 02:22:57 crc kubenswrapper[4824]: E1211 02:22:57.763074 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9259b35-21b6-4971-ae1b-04d367d993de" containerName="sg-core" Dec 11 02:22:57 crc kubenswrapper[4824]: I1211 02:22:57.763081 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9259b35-21b6-4971-ae1b-04d367d993de" containerName="sg-core" Dec 11 02:22:57 crc kubenswrapper[4824]: E1211 02:22:57.763098 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9259b35-21b6-4971-ae1b-04d367d993de" containerName="ceilometer-central-agent" Dec 11 02:22:57 crc kubenswrapper[4824]: I1211 02:22:57.763104 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9259b35-21b6-4971-ae1b-04d367d993de" containerName="ceilometer-central-agent" Dec 11 02:22:57 crc kubenswrapper[4824]: I1211 02:22:57.763285 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9259b35-21b6-4971-ae1b-04d367d993de" containerName="sg-core" Dec 11 02:22:57 crc kubenswrapper[4824]: I1211 02:22:57.763300 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9259b35-21b6-4971-ae1b-04d367d993de" containerName="ceilometer-notification-agent" Dec 11 02:22:57 crc kubenswrapper[4824]: I1211 02:22:57.763313 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9259b35-21b6-4971-ae1b-04d367d993de" containerName="proxy-httpd" Dec 11 02:22:57 crc kubenswrapper[4824]: I1211 02:22:57.763325 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9259b35-21b6-4971-ae1b-04d367d993de" containerName="ceilometer-central-agent" Dec 11 02:22:57 crc kubenswrapper[4824]: I1211 02:22:57.765522 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 11 02:22:57 crc kubenswrapper[4824]: I1211 02:22:57.767581 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 11 02:22:57 crc kubenswrapper[4824]: I1211 02:22:57.767962 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 11 02:22:57 crc kubenswrapper[4824]: I1211 02:22:57.780851 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 11 02:22:57 crc kubenswrapper[4824]: I1211 02:22:57.877065 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d77732b8-d6e7-4793-a699-cc569bb1ff60-config-data\") pod \"ceilometer-0\" (UID: \"d77732b8-d6e7-4793-a699-cc569bb1ff60\") " pod="openstack/ceilometer-0" Dec 11 02:22:57 crc kubenswrapper[4824]: I1211 02:22:57.877496 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d77732b8-d6e7-4793-a699-cc569bb1ff60-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d77732b8-d6e7-4793-a699-cc569bb1ff60\") " pod="openstack/ceilometer-0" Dec 11 02:22:57 crc kubenswrapper[4824]: I1211 02:22:57.877524 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d77732b8-d6e7-4793-a699-cc569bb1ff60-scripts\") pod \"ceilometer-0\" (UID: \"d77732b8-d6e7-4793-a699-cc569bb1ff60\") " pod="openstack/ceilometer-0" Dec 11 02:22:57 crc kubenswrapper[4824]: I1211 02:22:57.877616 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d77732b8-d6e7-4793-a699-cc569bb1ff60-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d77732b8-d6e7-4793-a699-cc569bb1ff60\") " pod="openstack/ceilometer-0" Dec 11 02:22:57 crc kubenswrapper[4824]: I1211 02:22:57.877702 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d77732b8-d6e7-4793-a699-cc569bb1ff60-log-httpd\") pod \"ceilometer-0\" (UID: \"d77732b8-d6e7-4793-a699-cc569bb1ff60\") " pod="openstack/ceilometer-0" Dec 11 02:22:57 crc kubenswrapper[4824]: I1211 02:22:57.877767 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dnjjp\" (UniqueName: \"kubernetes.io/projected/d77732b8-d6e7-4793-a699-cc569bb1ff60-kube-api-access-dnjjp\") pod \"ceilometer-0\" (UID: \"d77732b8-d6e7-4793-a699-cc569bb1ff60\") " pod="openstack/ceilometer-0" Dec 11 02:22:57 crc kubenswrapper[4824]: I1211 02:22:57.877850 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d77732b8-d6e7-4793-a699-cc569bb1ff60-run-httpd\") pod \"ceilometer-0\" (UID: \"d77732b8-d6e7-4793-a699-cc569bb1ff60\") " pod="openstack/ceilometer-0" Dec 11 02:22:57 crc kubenswrapper[4824]: I1211 02:22:57.909354 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 11 02:22:57 crc kubenswrapper[4824]: E1211 02:22:57.909964 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle config-data kube-api-access-dnjjp log-httpd run-httpd scripts sg-core-conf-yaml], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/ceilometer-0" podUID="d77732b8-d6e7-4793-a699-cc569bb1ff60" Dec 11 02:22:57 crc kubenswrapper[4824]: I1211 02:22:57.979151 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d77732b8-d6e7-4793-a699-cc569bb1ff60-run-httpd\") pod \"ceilometer-0\" (UID: \"d77732b8-d6e7-4793-a699-cc569bb1ff60\") " pod="openstack/ceilometer-0" Dec 11 02:22:57 crc kubenswrapper[4824]: I1211 02:22:57.979209 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d77732b8-d6e7-4793-a699-cc569bb1ff60-config-data\") pod \"ceilometer-0\" (UID: \"d77732b8-d6e7-4793-a699-cc569bb1ff60\") " pod="openstack/ceilometer-0" Dec 11 02:22:57 crc kubenswrapper[4824]: I1211 02:22:57.979261 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d77732b8-d6e7-4793-a699-cc569bb1ff60-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d77732b8-d6e7-4793-a699-cc569bb1ff60\") " pod="openstack/ceilometer-0" Dec 11 02:22:57 crc kubenswrapper[4824]: I1211 02:22:57.979280 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d77732b8-d6e7-4793-a699-cc569bb1ff60-scripts\") pod \"ceilometer-0\" (UID: \"d77732b8-d6e7-4793-a699-cc569bb1ff60\") " pod="openstack/ceilometer-0" Dec 11 02:22:57 crc kubenswrapper[4824]: I1211 02:22:57.979329 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d77732b8-d6e7-4793-a699-cc569bb1ff60-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d77732b8-d6e7-4793-a699-cc569bb1ff60\") " pod="openstack/ceilometer-0" Dec 11 02:22:57 crc kubenswrapper[4824]: I1211 02:22:57.979372 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d77732b8-d6e7-4793-a699-cc569bb1ff60-log-httpd\") pod \"ceilometer-0\" (UID: \"d77732b8-d6e7-4793-a699-cc569bb1ff60\") " pod="openstack/ceilometer-0" Dec 11 02:22:57 crc kubenswrapper[4824]: I1211 02:22:57.979398 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dnjjp\" (UniqueName: \"kubernetes.io/projected/d77732b8-d6e7-4793-a699-cc569bb1ff60-kube-api-access-dnjjp\") pod \"ceilometer-0\" (UID: \"d77732b8-d6e7-4793-a699-cc569bb1ff60\") " pod="openstack/ceilometer-0" Dec 11 02:22:57 crc kubenswrapper[4824]: I1211 02:22:57.979606 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d77732b8-d6e7-4793-a699-cc569bb1ff60-run-httpd\") pod \"ceilometer-0\" (UID: \"d77732b8-d6e7-4793-a699-cc569bb1ff60\") " pod="openstack/ceilometer-0" Dec 11 02:22:57 crc kubenswrapper[4824]: I1211 02:22:57.979956 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d77732b8-d6e7-4793-a699-cc569bb1ff60-log-httpd\") pod \"ceilometer-0\" (UID: \"d77732b8-d6e7-4793-a699-cc569bb1ff60\") " pod="openstack/ceilometer-0" Dec 11 02:22:57 crc kubenswrapper[4824]: I1211 02:22:57.984427 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d77732b8-d6e7-4793-a699-cc569bb1ff60-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d77732b8-d6e7-4793-a699-cc569bb1ff60\") " pod="openstack/ceilometer-0" Dec 11 02:22:57 crc kubenswrapper[4824]: I1211 02:22:57.984559 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d77732b8-d6e7-4793-a699-cc569bb1ff60-scripts\") pod \"ceilometer-0\" (UID: \"d77732b8-d6e7-4793-a699-cc569bb1ff60\") " pod="openstack/ceilometer-0" Dec 11 02:22:57 crc kubenswrapper[4824]: I1211 02:22:57.987767 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d77732b8-d6e7-4793-a699-cc569bb1ff60-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d77732b8-d6e7-4793-a699-cc569bb1ff60\") " pod="openstack/ceilometer-0" Dec 11 02:22:57 crc kubenswrapper[4824]: I1211 02:22:57.989227 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d77732b8-d6e7-4793-a699-cc569bb1ff60-config-data\") pod \"ceilometer-0\" (UID: \"d77732b8-d6e7-4793-a699-cc569bb1ff60\") " pod="openstack/ceilometer-0" Dec 11 02:22:58 crc kubenswrapper[4824]: I1211 02:22:58.002729 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dnjjp\" (UniqueName: \"kubernetes.io/projected/d77732b8-d6e7-4793-a699-cc569bb1ff60-kube-api-access-dnjjp\") pod \"ceilometer-0\" (UID: \"d77732b8-d6e7-4793-a699-cc569bb1ff60\") " pod="openstack/ceilometer-0" Dec 11 02:22:58 crc kubenswrapper[4824]: I1211 02:22:58.393645 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-787649dcdd-ts6qs" Dec 11 02:22:58 crc kubenswrapper[4824]: I1211 02:22:58.490127 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f15f9ed0-98f2-42fe-8006-2142098cc2b6-config-data\") pod \"f15f9ed0-98f2-42fe-8006-2142098cc2b6\" (UID: \"f15f9ed0-98f2-42fe-8006-2142098cc2b6\") " Dec 11 02:22:58 crc kubenswrapper[4824]: I1211 02:22:58.490541 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/f15f9ed0-98f2-42fe-8006-2142098cc2b6-horizon-tls-certs\") pod \"f15f9ed0-98f2-42fe-8006-2142098cc2b6\" (UID: \"f15f9ed0-98f2-42fe-8006-2142098cc2b6\") " Dec 11 02:22:58 crc kubenswrapper[4824]: I1211 02:22:58.490630 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f15f9ed0-98f2-42fe-8006-2142098cc2b6-logs\") pod \"f15f9ed0-98f2-42fe-8006-2142098cc2b6\" (UID: \"f15f9ed0-98f2-42fe-8006-2142098cc2b6\") " Dec 11 02:22:58 crc kubenswrapper[4824]: I1211 02:22:58.490748 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t6pdv\" (UniqueName: \"kubernetes.io/projected/f15f9ed0-98f2-42fe-8006-2142098cc2b6-kube-api-access-t6pdv\") pod \"f15f9ed0-98f2-42fe-8006-2142098cc2b6\" (UID: \"f15f9ed0-98f2-42fe-8006-2142098cc2b6\") " Dec 11 02:22:58 crc kubenswrapper[4824]: I1211 02:22:58.490788 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f15f9ed0-98f2-42fe-8006-2142098cc2b6-combined-ca-bundle\") pod \"f15f9ed0-98f2-42fe-8006-2142098cc2b6\" (UID: \"f15f9ed0-98f2-42fe-8006-2142098cc2b6\") " Dec 11 02:22:58 crc kubenswrapper[4824]: I1211 02:22:58.490819 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f15f9ed0-98f2-42fe-8006-2142098cc2b6-horizon-secret-key\") pod \"f15f9ed0-98f2-42fe-8006-2142098cc2b6\" (UID: \"f15f9ed0-98f2-42fe-8006-2142098cc2b6\") " Dec 11 02:22:58 crc kubenswrapper[4824]: I1211 02:22:58.490951 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f15f9ed0-98f2-42fe-8006-2142098cc2b6-scripts\") pod \"f15f9ed0-98f2-42fe-8006-2142098cc2b6\" (UID: \"f15f9ed0-98f2-42fe-8006-2142098cc2b6\") " Dec 11 02:22:58 crc kubenswrapper[4824]: I1211 02:22:58.491355 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f15f9ed0-98f2-42fe-8006-2142098cc2b6-logs" (OuterVolumeSpecName: "logs") pod "f15f9ed0-98f2-42fe-8006-2142098cc2b6" (UID: "f15f9ed0-98f2-42fe-8006-2142098cc2b6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:22:58 crc kubenswrapper[4824]: I1211 02:22:58.492128 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f15f9ed0-98f2-42fe-8006-2142098cc2b6-logs\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:58 crc kubenswrapper[4824]: I1211 02:22:58.497702 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f15f9ed0-98f2-42fe-8006-2142098cc2b6-kube-api-access-t6pdv" (OuterVolumeSpecName: "kube-api-access-t6pdv") pod "f15f9ed0-98f2-42fe-8006-2142098cc2b6" (UID: "f15f9ed0-98f2-42fe-8006-2142098cc2b6"). InnerVolumeSpecName "kube-api-access-t6pdv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:22:58 crc kubenswrapper[4824]: I1211 02:22:58.501243 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f15f9ed0-98f2-42fe-8006-2142098cc2b6-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "f15f9ed0-98f2-42fe-8006-2142098cc2b6" (UID: "f15f9ed0-98f2-42fe-8006-2142098cc2b6"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:22:58 crc kubenswrapper[4824]: I1211 02:22:58.518436 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f15f9ed0-98f2-42fe-8006-2142098cc2b6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f15f9ed0-98f2-42fe-8006-2142098cc2b6" (UID: "f15f9ed0-98f2-42fe-8006-2142098cc2b6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:22:58 crc kubenswrapper[4824]: I1211 02:22:58.519047 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f15f9ed0-98f2-42fe-8006-2142098cc2b6-scripts" (OuterVolumeSpecName: "scripts") pod "f15f9ed0-98f2-42fe-8006-2142098cc2b6" (UID: "f15f9ed0-98f2-42fe-8006-2142098cc2b6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:22:58 crc kubenswrapper[4824]: I1211 02:22:58.521813 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f15f9ed0-98f2-42fe-8006-2142098cc2b6-config-data" (OuterVolumeSpecName: "config-data") pod "f15f9ed0-98f2-42fe-8006-2142098cc2b6" (UID: "f15f9ed0-98f2-42fe-8006-2142098cc2b6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:22:58 crc kubenswrapper[4824]: I1211 02:22:58.543008 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f15f9ed0-98f2-42fe-8006-2142098cc2b6-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "f15f9ed0-98f2-42fe-8006-2142098cc2b6" (UID: "f15f9ed0-98f2-42fe-8006-2142098cc2b6"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:22:58 crc kubenswrapper[4824]: I1211 02:22:58.593639 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f15f9ed0-98f2-42fe-8006-2142098cc2b6-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:58 crc kubenswrapper[4824]: I1211 02:22:58.594587 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f15f9ed0-98f2-42fe-8006-2142098cc2b6-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:58 crc kubenswrapper[4824]: I1211 02:22:58.594603 4824 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/f15f9ed0-98f2-42fe-8006-2142098cc2b6-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:58 crc kubenswrapper[4824]: I1211 02:22:58.594617 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t6pdv\" (UniqueName: \"kubernetes.io/projected/f15f9ed0-98f2-42fe-8006-2142098cc2b6-kube-api-access-t6pdv\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:58 crc kubenswrapper[4824]: I1211 02:22:58.594630 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f15f9ed0-98f2-42fe-8006-2142098cc2b6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:58 crc kubenswrapper[4824]: I1211 02:22:58.594642 4824 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f15f9ed0-98f2-42fe-8006-2142098cc2b6-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:58 crc kubenswrapper[4824]: I1211 02:22:58.646043 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d9259b35-21b6-4971-ae1b-04d367d993de" path="/var/lib/kubelet/pods/d9259b35-21b6-4971-ae1b-04d367d993de/volumes" Dec 11 02:22:58 crc kubenswrapper[4824]: I1211 02:22:58.693796 4824 generic.go:334] "Generic (PLEG): container finished" podID="f15f9ed0-98f2-42fe-8006-2142098cc2b6" containerID="e4de9df2d8c6dc730220bfb406d58ac47fefcc07528a72cb76ac4ea026179a39" exitCode=137 Dec 11 02:22:58 crc kubenswrapper[4824]: I1211 02:22:58.693887 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 11 02:22:58 crc kubenswrapper[4824]: I1211 02:22:58.694599 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-787649dcdd-ts6qs" Dec 11 02:22:58 crc kubenswrapper[4824]: I1211 02:22:58.695152 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-787649dcdd-ts6qs" event={"ID":"f15f9ed0-98f2-42fe-8006-2142098cc2b6","Type":"ContainerDied","Data":"e4de9df2d8c6dc730220bfb406d58ac47fefcc07528a72cb76ac4ea026179a39"} Dec 11 02:22:58 crc kubenswrapper[4824]: I1211 02:22:58.695198 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-787649dcdd-ts6qs" event={"ID":"f15f9ed0-98f2-42fe-8006-2142098cc2b6","Type":"ContainerDied","Data":"1be3ac260841695de97d022c86cb5be498b8cbff79af2a73fc724347cc80c50a"} Dec 11 02:22:58 crc kubenswrapper[4824]: I1211 02:22:58.695215 4824 scope.go:117] "RemoveContainer" containerID="588925740a1286914b8db0f80b0f9426464ecc031e105965e01b1abcbf40a9ba" Dec 11 02:22:58 crc kubenswrapper[4824]: I1211 02:22:58.706221 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 11 02:22:58 crc kubenswrapper[4824]: I1211 02:22:58.730073 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-787649dcdd-ts6qs"] Dec 11 02:22:58 crc kubenswrapper[4824]: I1211 02:22:58.738762 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-787649dcdd-ts6qs"] Dec 11 02:22:58 crc kubenswrapper[4824]: I1211 02:22:58.774421 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-6458cc7487-d2555" Dec 11 02:22:58 crc kubenswrapper[4824]: I1211 02:22:58.782079 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-6458cc7487-d2555" Dec 11 02:22:58 crc kubenswrapper[4824]: I1211 02:22:58.802224 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d77732b8-d6e7-4793-a699-cc569bb1ff60-scripts\") pod \"d77732b8-d6e7-4793-a699-cc569bb1ff60\" (UID: \"d77732b8-d6e7-4793-a699-cc569bb1ff60\") " Dec 11 02:22:58 crc kubenswrapper[4824]: I1211 02:22:58.802328 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d77732b8-d6e7-4793-a699-cc569bb1ff60-log-httpd\") pod \"d77732b8-d6e7-4793-a699-cc569bb1ff60\" (UID: \"d77732b8-d6e7-4793-a699-cc569bb1ff60\") " Dec 11 02:22:58 crc kubenswrapper[4824]: I1211 02:22:58.802358 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dnjjp\" (UniqueName: \"kubernetes.io/projected/d77732b8-d6e7-4793-a699-cc569bb1ff60-kube-api-access-dnjjp\") pod \"d77732b8-d6e7-4793-a699-cc569bb1ff60\" (UID: \"d77732b8-d6e7-4793-a699-cc569bb1ff60\") " Dec 11 02:22:58 crc kubenswrapper[4824]: I1211 02:22:58.802511 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d77732b8-d6e7-4793-a699-cc569bb1ff60-combined-ca-bundle\") pod \"d77732b8-d6e7-4793-a699-cc569bb1ff60\" (UID: \"d77732b8-d6e7-4793-a699-cc569bb1ff60\") " Dec 11 02:22:58 crc kubenswrapper[4824]: I1211 02:22:58.802567 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d77732b8-d6e7-4793-a699-cc569bb1ff60-config-data\") pod \"d77732b8-d6e7-4793-a699-cc569bb1ff60\" (UID: \"d77732b8-d6e7-4793-a699-cc569bb1ff60\") " Dec 11 02:22:58 crc kubenswrapper[4824]: I1211 02:22:58.802597 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d77732b8-d6e7-4793-a699-cc569bb1ff60-run-httpd\") pod \"d77732b8-d6e7-4793-a699-cc569bb1ff60\" (UID: \"d77732b8-d6e7-4793-a699-cc569bb1ff60\") " Dec 11 02:22:58 crc kubenswrapper[4824]: I1211 02:22:58.802627 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d77732b8-d6e7-4793-a699-cc569bb1ff60-sg-core-conf-yaml\") pod \"d77732b8-d6e7-4793-a699-cc569bb1ff60\" (UID: \"d77732b8-d6e7-4793-a699-cc569bb1ff60\") " Dec 11 02:22:58 crc kubenswrapper[4824]: I1211 02:22:58.803745 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d77732b8-d6e7-4793-a699-cc569bb1ff60-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "d77732b8-d6e7-4793-a699-cc569bb1ff60" (UID: "d77732b8-d6e7-4793-a699-cc569bb1ff60"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:22:58 crc kubenswrapper[4824]: I1211 02:22:58.804047 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d77732b8-d6e7-4793-a699-cc569bb1ff60-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "d77732b8-d6e7-4793-a699-cc569bb1ff60" (UID: "d77732b8-d6e7-4793-a699-cc569bb1ff60"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:22:58 crc kubenswrapper[4824]: I1211 02:22:58.812081 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d77732b8-d6e7-4793-a699-cc569bb1ff60-scripts" (OuterVolumeSpecName: "scripts") pod "d77732b8-d6e7-4793-a699-cc569bb1ff60" (UID: "d77732b8-d6e7-4793-a699-cc569bb1ff60"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:22:58 crc kubenswrapper[4824]: I1211 02:22:58.812146 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d77732b8-d6e7-4793-a699-cc569bb1ff60-config-data" (OuterVolumeSpecName: "config-data") pod "d77732b8-d6e7-4793-a699-cc569bb1ff60" (UID: "d77732b8-d6e7-4793-a699-cc569bb1ff60"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:22:58 crc kubenswrapper[4824]: I1211 02:22:58.812210 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d77732b8-d6e7-4793-a699-cc569bb1ff60-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "d77732b8-d6e7-4793-a699-cc569bb1ff60" (UID: "d77732b8-d6e7-4793-a699-cc569bb1ff60"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:22:58 crc kubenswrapper[4824]: I1211 02:22:58.812414 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d77732b8-d6e7-4793-a699-cc569bb1ff60-kube-api-access-dnjjp" (OuterVolumeSpecName: "kube-api-access-dnjjp") pod "d77732b8-d6e7-4793-a699-cc569bb1ff60" (UID: "d77732b8-d6e7-4793-a699-cc569bb1ff60"). InnerVolumeSpecName "kube-api-access-dnjjp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:22:58 crc kubenswrapper[4824]: I1211 02:22:58.817040 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d77732b8-d6e7-4793-a699-cc569bb1ff60-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d77732b8-d6e7-4793-a699-cc569bb1ff60" (UID: "d77732b8-d6e7-4793-a699-cc569bb1ff60"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:22:58 crc kubenswrapper[4824]: I1211 02:22:58.829165 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 11 02:22:58 crc kubenswrapper[4824]: I1211 02:22:58.829586 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="e7794eea-3d90-4bd6-8d69-732d80385243" containerName="glance-log" containerID="cri-o://8411d991de0195412273e5c06a0d59951db008bb73091da57111623429573c22" gracePeriod=30 Dec 11 02:22:58 crc kubenswrapper[4824]: I1211 02:22:58.829781 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="e7794eea-3d90-4bd6-8d69-732d80385243" containerName="glance-httpd" containerID="cri-o://4364c167b32c39ebe179b655d9e1d469552dee7e3454b999439b9e634cc9cf53" gracePeriod=30 Dec 11 02:22:58 crc kubenswrapper[4824]: I1211 02:22:58.890376 4824 scope.go:117] "RemoveContainer" containerID="e4de9df2d8c6dc730220bfb406d58ac47fefcc07528a72cb76ac4ea026179a39" Dec 11 02:22:58 crc kubenswrapper[4824]: I1211 02:22:58.904763 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d77732b8-d6e7-4793-a699-cc569bb1ff60-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:58 crc kubenswrapper[4824]: I1211 02:22:58.905025 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d77732b8-d6e7-4793-a699-cc569bb1ff60-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:58 crc kubenswrapper[4824]: I1211 02:22:58.905035 4824 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d77732b8-d6e7-4793-a699-cc569bb1ff60-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:58 crc kubenswrapper[4824]: I1211 02:22:58.905043 4824 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d77732b8-d6e7-4793-a699-cc569bb1ff60-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:58 crc kubenswrapper[4824]: I1211 02:22:58.905051 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d77732b8-d6e7-4793-a699-cc569bb1ff60-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:58 crc kubenswrapper[4824]: I1211 02:22:58.905059 4824 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d77732b8-d6e7-4793-a699-cc569bb1ff60-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:58 crc kubenswrapper[4824]: I1211 02:22:58.905067 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dnjjp\" (UniqueName: \"kubernetes.io/projected/d77732b8-d6e7-4793-a699-cc569bb1ff60-kube-api-access-dnjjp\") on node \"crc\" DevicePath \"\"" Dec 11 02:22:58 crc kubenswrapper[4824]: I1211 02:22:58.924798 4824 scope.go:117] "RemoveContainer" containerID="588925740a1286914b8db0f80b0f9426464ecc031e105965e01b1abcbf40a9ba" Dec 11 02:22:58 crc kubenswrapper[4824]: E1211 02:22:58.928539 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"588925740a1286914b8db0f80b0f9426464ecc031e105965e01b1abcbf40a9ba\": container with ID starting with 588925740a1286914b8db0f80b0f9426464ecc031e105965e01b1abcbf40a9ba not found: ID does not exist" containerID="588925740a1286914b8db0f80b0f9426464ecc031e105965e01b1abcbf40a9ba" Dec 11 02:22:58 crc kubenswrapper[4824]: I1211 02:22:58.928578 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"588925740a1286914b8db0f80b0f9426464ecc031e105965e01b1abcbf40a9ba"} err="failed to get container status \"588925740a1286914b8db0f80b0f9426464ecc031e105965e01b1abcbf40a9ba\": rpc error: code = NotFound desc = could not find container \"588925740a1286914b8db0f80b0f9426464ecc031e105965e01b1abcbf40a9ba\": container with ID starting with 588925740a1286914b8db0f80b0f9426464ecc031e105965e01b1abcbf40a9ba not found: ID does not exist" Dec 11 02:22:58 crc kubenswrapper[4824]: I1211 02:22:58.928601 4824 scope.go:117] "RemoveContainer" containerID="e4de9df2d8c6dc730220bfb406d58ac47fefcc07528a72cb76ac4ea026179a39" Dec 11 02:22:58 crc kubenswrapper[4824]: E1211 02:22:58.936099 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e4de9df2d8c6dc730220bfb406d58ac47fefcc07528a72cb76ac4ea026179a39\": container with ID starting with e4de9df2d8c6dc730220bfb406d58ac47fefcc07528a72cb76ac4ea026179a39 not found: ID does not exist" containerID="e4de9df2d8c6dc730220bfb406d58ac47fefcc07528a72cb76ac4ea026179a39" Dec 11 02:22:58 crc kubenswrapper[4824]: I1211 02:22:58.936146 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e4de9df2d8c6dc730220bfb406d58ac47fefcc07528a72cb76ac4ea026179a39"} err="failed to get container status \"e4de9df2d8c6dc730220bfb406d58ac47fefcc07528a72cb76ac4ea026179a39\": rpc error: code = NotFound desc = could not find container \"e4de9df2d8c6dc730220bfb406d58ac47fefcc07528a72cb76ac4ea026179a39\": container with ID starting with e4de9df2d8c6dc730220bfb406d58ac47fefcc07528a72cb76ac4ea026179a39 not found: ID does not exist" Dec 11 02:22:59 crc kubenswrapper[4824]: I1211 02:22:59.709024 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-kchr8"] Dec 11 02:22:59 crc kubenswrapper[4824]: E1211 02:22:59.709447 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f15f9ed0-98f2-42fe-8006-2142098cc2b6" containerName="horizon-log" Dec 11 02:22:59 crc kubenswrapper[4824]: I1211 02:22:59.709461 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f15f9ed0-98f2-42fe-8006-2142098cc2b6" containerName="horizon-log" Dec 11 02:22:59 crc kubenswrapper[4824]: E1211 02:22:59.709481 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f15f9ed0-98f2-42fe-8006-2142098cc2b6" containerName="horizon" Dec 11 02:22:59 crc kubenswrapper[4824]: I1211 02:22:59.709488 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f15f9ed0-98f2-42fe-8006-2142098cc2b6" containerName="horizon" Dec 11 02:22:59 crc kubenswrapper[4824]: I1211 02:22:59.709679 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f15f9ed0-98f2-42fe-8006-2142098cc2b6" containerName="horizon-log" Dec 11 02:22:59 crc kubenswrapper[4824]: I1211 02:22:59.709703 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f15f9ed0-98f2-42fe-8006-2142098cc2b6" containerName="horizon" Dec 11 02:22:59 crc kubenswrapper[4824]: I1211 02:22:59.710337 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-kchr8" Dec 11 02:22:59 crc kubenswrapper[4824]: I1211 02:22:59.721096 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-kchr8"] Dec 11 02:22:59 crc kubenswrapper[4824]: I1211 02:22:59.742080 4824 generic.go:334] "Generic (PLEG): container finished" podID="e7794eea-3d90-4bd6-8d69-732d80385243" containerID="8411d991de0195412273e5c06a0d59951db008bb73091da57111623429573c22" exitCode=143 Dec 11 02:22:59 crc kubenswrapper[4824]: I1211 02:22:59.742154 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e7794eea-3d90-4bd6-8d69-732d80385243","Type":"ContainerDied","Data":"8411d991de0195412273e5c06a0d59951db008bb73091da57111623429573c22"} Dec 11 02:22:59 crc kubenswrapper[4824]: I1211 02:22:59.746469 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 11 02:22:59 crc kubenswrapper[4824]: I1211 02:22:59.779428 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-tqtd7"] Dec 11 02:22:59 crc kubenswrapper[4824]: I1211 02:22:59.780513 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-tqtd7" Dec 11 02:22:59 crc kubenswrapper[4824]: I1211 02:22:59.792248 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-tqtd7"] Dec 11 02:22:59 crc kubenswrapper[4824]: I1211 02:22:59.810605 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 11 02:22:59 crc kubenswrapper[4824]: I1211 02:22:59.819670 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 11 02:22:59 crc kubenswrapper[4824]: I1211 02:22:59.820644 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vm67l\" (UniqueName: \"kubernetes.io/projected/eb82f620-fb10-4f76-b2eb-adde800579bf-kube-api-access-vm67l\") pod \"nova-api-db-create-kchr8\" (UID: \"eb82f620-fb10-4f76-b2eb-adde800579bf\") " pod="openstack/nova-api-db-create-kchr8" Dec 11 02:22:59 crc kubenswrapper[4824]: I1211 02:22:59.820674 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eb82f620-fb10-4f76-b2eb-adde800579bf-operator-scripts\") pod \"nova-api-db-create-kchr8\" (UID: \"eb82f620-fb10-4f76-b2eb-adde800579bf\") " pod="openstack/nova-api-db-create-kchr8" Dec 11 02:22:59 crc kubenswrapper[4824]: I1211 02:22:59.873698 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 11 02:22:59 crc kubenswrapper[4824]: I1211 02:22:59.883248 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 11 02:22:59 crc kubenswrapper[4824]: I1211 02:22:59.887783 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 11 02:22:59 crc kubenswrapper[4824]: I1211 02:22:59.888193 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 11 02:22:59 crc kubenswrapper[4824]: I1211 02:22:59.894330 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 11 02:22:59 crc kubenswrapper[4824]: I1211 02:22:59.902082 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-kpv29"] Dec 11 02:22:59 crc kubenswrapper[4824]: I1211 02:22:59.903449 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-kpv29" Dec 11 02:22:59 crc kubenswrapper[4824]: I1211 02:22:59.921007 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-32c8-account-create-update-stjs4"] Dec 11 02:22:59 crc kubenswrapper[4824]: I1211 02:22:59.922239 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/295fe4d0-5910-49d6-b8ac-56ff445666f9-operator-scripts\") pod \"nova-cell0-db-create-tqtd7\" (UID: \"295fe4d0-5910-49d6-b8ac-56ff445666f9\") " pod="openstack/nova-cell0-db-create-tqtd7" Dec 11 02:22:59 crc kubenswrapper[4824]: I1211 02:22:59.922287 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sn9sl\" (UniqueName: \"kubernetes.io/projected/295fe4d0-5910-49d6-b8ac-56ff445666f9-kube-api-access-sn9sl\") pod \"nova-cell0-db-create-tqtd7\" (UID: \"295fe4d0-5910-49d6-b8ac-56ff445666f9\") " pod="openstack/nova-cell0-db-create-tqtd7" Dec 11 02:22:59 crc kubenswrapper[4824]: I1211 02:22:59.922328 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/21e824ae-7668-42ad-a039-b7795c13b449-scripts\") pod \"ceilometer-0\" (UID: \"21e824ae-7668-42ad-a039-b7795c13b449\") " pod="openstack/ceilometer-0" Dec 11 02:22:59 crc kubenswrapper[4824]: I1211 02:22:59.922357 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21e824ae-7668-42ad-a039-b7795c13b449-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"21e824ae-7668-42ad-a039-b7795c13b449\") " pod="openstack/ceilometer-0" Dec 11 02:22:59 crc kubenswrapper[4824]: I1211 02:22:59.922383 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/21e824ae-7668-42ad-a039-b7795c13b449-log-httpd\") pod \"ceilometer-0\" (UID: \"21e824ae-7668-42ad-a039-b7795c13b449\") " pod="openstack/ceilometer-0" Dec 11 02:22:59 crc kubenswrapper[4824]: I1211 02:22:59.922421 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/21e824ae-7668-42ad-a039-b7795c13b449-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"21e824ae-7668-42ad-a039-b7795c13b449\") " pod="openstack/ceilometer-0" Dec 11 02:22:59 crc kubenswrapper[4824]: I1211 02:22:59.922454 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/21e824ae-7668-42ad-a039-b7795c13b449-run-httpd\") pod \"ceilometer-0\" (UID: \"21e824ae-7668-42ad-a039-b7795c13b449\") " pod="openstack/ceilometer-0" Dec 11 02:22:59 crc kubenswrapper[4824]: I1211 02:22:59.922473 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21e824ae-7668-42ad-a039-b7795c13b449-config-data\") pod \"ceilometer-0\" (UID: \"21e824ae-7668-42ad-a039-b7795c13b449\") " pod="openstack/ceilometer-0" Dec 11 02:22:59 crc kubenswrapper[4824]: I1211 02:22:59.922498 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vm67l\" (UniqueName: \"kubernetes.io/projected/eb82f620-fb10-4f76-b2eb-adde800579bf-kube-api-access-vm67l\") pod \"nova-api-db-create-kchr8\" (UID: \"eb82f620-fb10-4f76-b2eb-adde800579bf\") " pod="openstack/nova-api-db-create-kchr8" Dec 11 02:22:59 crc kubenswrapper[4824]: I1211 02:22:59.922549 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4xzbw\" (UniqueName: \"kubernetes.io/projected/21e824ae-7668-42ad-a039-b7795c13b449-kube-api-access-4xzbw\") pod \"ceilometer-0\" (UID: \"21e824ae-7668-42ad-a039-b7795c13b449\") " pod="openstack/ceilometer-0" Dec 11 02:22:59 crc kubenswrapper[4824]: I1211 02:22:59.922771 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eb82f620-fb10-4f76-b2eb-adde800579bf-operator-scripts\") pod \"nova-api-db-create-kchr8\" (UID: \"eb82f620-fb10-4f76-b2eb-adde800579bf\") " pod="openstack/nova-api-db-create-kchr8" Dec 11 02:22:59 crc kubenswrapper[4824]: I1211 02:22:59.923411 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eb82f620-fb10-4f76-b2eb-adde800579bf-operator-scripts\") pod \"nova-api-db-create-kchr8\" (UID: \"eb82f620-fb10-4f76-b2eb-adde800579bf\") " pod="openstack/nova-api-db-create-kchr8" Dec 11 02:22:59 crc kubenswrapper[4824]: I1211 02:22:59.923731 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-32c8-account-create-update-stjs4" Dec 11 02:22:59 crc kubenswrapper[4824]: I1211 02:22:59.925705 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Dec 11 02:22:59 crc kubenswrapper[4824]: I1211 02:22:59.942199 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-kpv29"] Dec 11 02:22:59 crc kubenswrapper[4824]: I1211 02:22:59.943465 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vm67l\" (UniqueName: \"kubernetes.io/projected/eb82f620-fb10-4f76-b2eb-adde800579bf-kube-api-access-vm67l\") pod \"nova-api-db-create-kchr8\" (UID: \"eb82f620-fb10-4f76-b2eb-adde800579bf\") " pod="openstack/nova-api-db-create-kchr8" Dec 11 02:22:59 crc kubenswrapper[4824]: I1211 02:22:59.953330 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-32c8-account-create-update-stjs4"] Dec 11 02:23:00 crc kubenswrapper[4824]: I1211 02:23:00.024912 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/21e824ae-7668-42ad-a039-b7795c13b449-log-httpd\") pod \"ceilometer-0\" (UID: \"21e824ae-7668-42ad-a039-b7795c13b449\") " pod="openstack/ceilometer-0" Dec 11 02:23:00 crc kubenswrapper[4824]: I1211 02:23:00.024960 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bpm9h\" (UniqueName: \"kubernetes.io/projected/f0fefd9f-c55d-4bb0-8d9c-b3f7c42ac4d4-kube-api-access-bpm9h\") pod \"nova-api-32c8-account-create-update-stjs4\" (UID: \"f0fefd9f-c55d-4bb0-8d9c-b3f7c42ac4d4\") " pod="openstack/nova-api-32c8-account-create-update-stjs4" Dec 11 02:23:00 crc kubenswrapper[4824]: I1211 02:23:00.025005 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/21e824ae-7668-42ad-a039-b7795c13b449-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"21e824ae-7668-42ad-a039-b7795c13b449\") " pod="openstack/ceilometer-0" Dec 11 02:23:00 crc kubenswrapper[4824]: I1211 02:23:00.025302 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/21e824ae-7668-42ad-a039-b7795c13b449-run-httpd\") pod \"ceilometer-0\" (UID: \"21e824ae-7668-42ad-a039-b7795c13b449\") " pod="openstack/ceilometer-0" Dec 11 02:23:00 crc kubenswrapper[4824]: I1211 02:23:00.025384 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21e824ae-7668-42ad-a039-b7795c13b449-config-data\") pod \"ceilometer-0\" (UID: \"21e824ae-7668-42ad-a039-b7795c13b449\") " pod="openstack/ceilometer-0" Dec 11 02:23:00 crc kubenswrapper[4824]: I1211 02:23:00.025500 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f0fefd9f-c55d-4bb0-8d9c-b3f7c42ac4d4-operator-scripts\") pod \"nova-api-32c8-account-create-update-stjs4\" (UID: \"f0fefd9f-c55d-4bb0-8d9c-b3f7c42ac4d4\") " pod="openstack/nova-api-32c8-account-create-update-stjs4" Dec 11 02:23:00 crc kubenswrapper[4824]: I1211 02:23:00.025608 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4xzbw\" (UniqueName: \"kubernetes.io/projected/21e824ae-7668-42ad-a039-b7795c13b449-kube-api-access-4xzbw\") pod \"ceilometer-0\" (UID: \"21e824ae-7668-42ad-a039-b7795c13b449\") " pod="openstack/ceilometer-0" Dec 11 02:23:00 crc kubenswrapper[4824]: I1211 02:23:00.025726 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7zk6s\" (UniqueName: \"kubernetes.io/projected/33034069-6622-45c9-9045-44a5e01682bd-kube-api-access-7zk6s\") pod \"nova-cell1-db-create-kpv29\" (UID: \"33034069-6622-45c9-9045-44a5e01682bd\") " pod="openstack/nova-cell1-db-create-kpv29" Dec 11 02:23:00 crc kubenswrapper[4824]: I1211 02:23:00.025860 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/295fe4d0-5910-49d6-b8ac-56ff445666f9-operator-scripts\") pod \"nova-cell0-db-create-tqtd7\" (UID: \"295fe4d0-5910-49d6-b8ac-56ff445666f9\") " pod="openstack/nova-cell0-db-create-tqtd7" Dec 11 02:23:00 crc kubenswrapper[4824]: I1211 02:23:00.025973 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sn9sl\" (UniqueName: \"kubernetes.io/projected/295fe4d0-5910-49d6-b8ac-56ff445666f9-kube-api-access-sn9sl\") pod \"nova-cell0-db-create-tqtd7\" (UID: \"295fe4d0-5910-49d6-b8ac-56ff445666f9\") " pod="openstack/nova-cell0-db-create-tqtd7" Dec 11 02:23:00 crc kubenswrapper[4824]: I1211 02:23:00.026086 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/33034069-6622-45c9-9045-44a5e01682bd-operator-scripts\") pod \"nova-cell1-db-create-kpv29\" (UID: \"33034069-6622-45c9-9045-44a5e01682bd\") " pod="openstack/nova-cell1-db-create-kpv29" Dec 11 02:23:00 crc kubenswrapper[4824]: I1211 02:23:00.026338 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/21e824ae-7668-42ad-a039-b7795c13b449-scripts\") pod \"ceilometer-0\" (UID: \"21e824ae-7668-42ad-a039-b7795c13b449\") " pod="openstack/ceilometer-0" Dec 11 02:23:00 crc kubenswrapper[4824]: I1211 02:23:00.026473 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21e824ae-7668-42ad-a039-b7795c13b449-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"21e824ae-7668-42ad-a039-b7795c13b449\") " pod="openstack/ceilometer-0" Dec 11 02:23:00 crc kubenswrapper[4824]: I1211 02:23:00.025771 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/21e824ae-7668-42ad-a039-b7795c13b449-run-httpd\") pod \"ceilometer-0\" (UID: \"21e824ae-7668-42ad-a039-b7795c13b449\") " pod="openstack/ceilometer-0" Dec 11 02:23:00 crc kubenswrapper[4824]: I1211 02:23:00.026603 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/295fe4d0-5910-49d6-b8ac-56ff445666f9-operator-scripts\") pod \"nova-cell0-db-create-tqtd7\" (UID: \"295fe4d0-5910-49d6-b8ac-56ff445666f9\") " pod="openstack/nova-cell0-db-create-tqtd7" Dec 11 02:23:00 crc kubenswrapper[4824]: I1211 02:23:00.025449 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/21e824ae-7668-42ad-a039-b7795c13b449-log-httpd\") pod \"ceilometer-0\" (UID: \"21e824ae-7668-42ad-a039-b7795c13b449\") " pod="openstack/ceilometer-0" Dec 11 02:23:00 crc kubenswrapper[4824]: I1211 02:23:00.029166 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21e824ae-7668-42ad-a039-b7795c13b449-config-data\") pod \"ceilometer-0\" (UID: \"21e824ae-7668-42ad-a039-b7795c13b449\") " pod="openstack/ceilometer-0" Dec 11 02:23:00 crc kubenswrapper[4824]: I1211 02:23:00.029845 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/21e824ae-7668-42ad-a039-b7795c13b449-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"21e824ae-7668-42ad-a039-b7795c13b449\") " pod="openstack/ceilometer-0" Dec 11 02:23:00 crc kubenswrapper[4824]: I1211 02:23:00.030078 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21e824ae-7668-42ad-a039-b7795c13b449-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"21e824ae-7668-42ad-a039-b7795c13b449\") " pod="openstack/ceilometer-0" Dec 11 02:23:00 crc kubenswrapper[4824]: I1211 02:23:00.038882 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/21e824ae-7668-42ad-a039-b7795c13b449-scripts\") pod \"ceilometer-0\" (UID: \"21e824ae-7668-42ad-a039-b7795c13b449\") " pod="openstack/ceilometer-0" Dec 11 02:23:00 crc kubenswrapper[4824]: I1211 02:23:00.039377 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-kchr8" Dec 11 02:23:00 crc kubenswrapper[4824]: I1211 02:23:00.042906 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sn9sl\" (UniqueName: \"kubernetes.io/projected/295fe4d0-5910-49d6-b8ac-56ff445666f9-kube-api-access-sn9sl\") pod \"nova-cell0-db-create-tqtd7\" (UID: \"295fe4d0-5910-49d6-b8ac-56ff445666f9\") " pod="openstack/nova-cell0-db-create-tqtd7" Dec 11 02:23:00 crc kubenswrapper[4824]: I1211 02:23:00.061009 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4xzbw\" (UniqueName: \"kubernetes.io/projected/21e824ae-7668-42ad-a039-b7795c13b449-kube-api-access-4xzbw\") pod \"ceilometer-0\" (UID: \"21e824ae-7668-42ad-a039-b7795c13b449\") " pod="openstack/ceilometer-0" Dec 11 02:23:00 crc kubenswrapper[4824]: I1211 02:23:00.097452 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-2225-account-create-update-f9zht"] Dec 11 02:23:00 crc kubenswrapper[4824]: I1211 02:23:00.097597 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-tqtd7" Dec 11 02:23:00 crc kubenswrapper[4824]: I1211 02:23:00.098624 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-2225-account-create-update-f9zht" Dec 11 02:23:00 crc kubenswrapper[4824]: I1211 02:23:00.101014 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Dec 11 02:23:00 crc kubenswrapper[4824]: I1211 02:23:00.131820 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bpm9h\" (UniqueName: \"kubernetes.io/projected/f0fefd9f-c55d-4bb0-8d9c-b3f7c42ac4d4-kube-api-access-bpm9h\") pod \"nova-api-32c8-account-create-update-stjs4\" (UID: \"f0fefd9f-c55d-4bb0-8d9c-b3f7c42ac4d4\") " pod="openstack/nova-api-32c8-account-create-update-stjs4" Dec 11 02:23:00 crc kubenswrapper[4824]: I1211 02:23:00.131924 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f0fefd9f-c55d-4bb0-8d9c-b3f7c42ac4d4-operator-scripts\") pod \"nova-api-32c8-account-create-update-stjs4\" (UID: \"f0fefd9f-c55d-4bb0-8d9c-b3f7c42ac4d4\") " pod="openstack/nova-api-32c8-account-create-update-stjs4" Dec 11 02:23:00 crc kubenswrapper[4824]: I1211 02:23:00.131962 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7zk6s\" (UniqueName: \"kubernetes.io/projected/33034069-6622-45c9-9045-44a5e01682bd-kube-api-access-7zk6s\") pod \"nova-cell1-db-create-kpv29\" (UID: \"33034069-6622-45c9-9045-44a5e01682bd\") " pod="openstack/nova-cell1-db-create-kpv29" Dec 11 02:23:00 crc kubenswrapper[4824]: I1211 02:23:00.132008 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/33034069-6622-45c9-9045-44a5e01682bd-operator-scripts\") pod \"nova-cell1-db-create-kpv29\" (UID: \"33034069-6622-45c9-9045-44a5e01682bd\") " pod="openstack/nova-cell1-db-create-kpv29" Dec 11 02:23:00 crc kubenswrapper[4824]: I1211 02:23:00.132844 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/33034069-6622-45c9-9045-44a5e01682bd-operator-scripts\") pod \"nova-cell1-db-create-kpv29\" (UID: \"33034069-6622-45c9-9045-44a5e01682bd\") " pod="openstack/nova-cell1-db-create-kpv29" Dec 11 02:23:00 crc kubenswrapper[4824]: I1211 02:23:00.135057 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f0fefd9f-c55d-4bb0-8d9c-b3f7c42ac4d4-operator-scripts\") pod \"nova-api-32c8-account-create-update-stjs4\" (UID: \"f0fefd9f-c55d-4bb0-8d9c-b3f7c42ac4d4\") " pod="openstack/nova-api-32c8-account-create-update-stjs4" Dec 11 02:23:00 crc kubenswrapper[4824]: I1211 02:23:00.135283 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-2225-account-create-update-f9zht"] Dec 11 02:23:00 crc kubenswrapper[4824]: I1211 02:23:00.150969 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bpm9h\" (UniqueName: \"kubernetes.io/projected/f0fefd9f-c55d-4bb0-8d9c-b3f7c42ac4d4-kube-api-access-bpm9h\") pod \"nova-api-32c8-account-create-update-stjs4\" (UID: \"f0fefd9f-c55d-4bb0-8d9c-b3f7c42ac4d4\") " pod="openstack/nova-api-32c8-account-create-update-stjs4" Dec 11 02:23:00 crc kubenswrapper[4824]: I1211 02:23:00.154626 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7zk6s\" (UniqueName: \"kubernetes.io/projected/33034069-6622-45c9-9045-44a5e01682bd-kube-api-access-7zk6s\") pod \"nova-cell1-db-create-kpv29\" (UID: \"33034069-6622-45c9-9045-44a5e01682bd\") " pod="openstack/nova-cell1-db-create-kpv29" Dec 11 02:23:00 crc kubenswrapper[4824]: I1211 02:23:00.220453 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 11 02:23:00 crc kubenswrapper[4824]: I1211 02:23:00.233960 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/14d1e80f-7c53-4c15-a002-eacde62186bc-operator-scripts\") pod \"nova-cell0-2225-account-create-update-f9zht\" (UID: \"14d1e80f-7c53-4c15-a002-eacde62186bc\") " pod="openstack/nova-cell0-2225-account-create-update-f9zht" Dec 11 02:23:00 crc kubenswrapper[4824]: I1211 02:23:00.234046 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-494lt\" (UniqueName: \"kubernetes.io/projected/14d1e80f-7c53-4c15-a002-eacde62186bc-kube-api-access-494lt\") pod \"nova-cell0-2225-account-create-update-f9zht\" (UID: \"14d1e80f-7c53-4c15-a002-eacde62186bc\") " pod="openstack/nova-cell0-2225-account-create-update-f9zht" Dec 11 02:23:00 crc kubenswrapper[4824]: I1211 02:23:00.243089 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-kpv29" Dec 11 02:23:00 crc kubenswrapper[4824]: I1211 02:23:00.291376 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-32c8-account-create-update-stjs4" Dec 11 02:23:00 crc kubenswrapper[4824]: I1211 02:23:00.297536 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-7a4e-account-create-update-c7h9j"] Dec 11 02:23:00 crc kubenswrapper[4824]: I1211 02:23:00.298863 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-7a4e-account-create-update-c7h9j" Dec 11 02:23:00 crc kubenswrapper[4824]: I1211 02:23:00.301719 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Dec 11 02:23:00 crc kubenswrapper[4824]: I1211 02:23:00.317869 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-7a4e-account-create-update-c7h9j"] Dec 11 02:23:00 crc kubenswrapper[4824]: I1211 02:23:00.341679 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2dchb\" (UniqueName: \"kubernetes.io/projected/78553f90-b619-4ff6-84b4-05835483a29e-kube-api-access-2dchb\") pod \"nova-cell1-7a4e-account-create-update-c7h9j\" (UID: \"78553f90-b619-4ff6-84b4-05835483a29e\") " pod="openstack/nova-cell1-7a4e-account-create-update-c7h9j" Dec 11 02:23:00 crc kubenswrapper[4824]: I1211 02:23:00.341786 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/14d1e80f-7c53-4c15-a002-eacde62186bc-operator-scripts\") pod \"nova-cell0-2225-account-create-update-f9zht\" (UID: \"14d1e80f-7c53-4c15-a002-eacde62186bc\") " pod="openstack/nova-cell0-2225-account-create-update-f9zht" Dec 11 02:23:00 crc kubenswrapper[4824]: I1211 02:23:00.341909 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-494lt\" (UniqueName: \"kubernetes.io/projected/14d1e80f-7c53-4c15-a002-eacde62186bc-kube-api-access-494lt\") pod \"nova-cell0-2225-account-create-update-f9zht\" (UID: \"14d1e80f-7c53-4c15-a002-eacde62186bc\") " pod="openstack/nova-cell0-2225-account-create-update-f9zht" Dec 11 02:23:00 crc kubenswrapper[4824]: I1211 02:23:00.341951 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/78553f90-b619-4ff6-84b4-05835483a29e-operator-scripts\") pod \"nova-cell1-7a4e-account-create-update-c7h9j\" (UID: \"78553f90-b619-4ff6-84b4-05835483a29e\") " pod="openstack/nova-cell1-7a4e-account-create-update-c7h9j" Dec 11 02:23:00 crc kubenswrapper[4824]: I1211 02:23:00.343236 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/14d1e80f-7c53-4c15-a002-eacde62186bc-operator-scripts\") pod \"nova-cell0-2225-account-create-update-f9zht\" (UID: \"14d1e80f-7c53-4c15-a002-eacde62186bc\") " pod="openstack/nova-cell0-2225-account-create-update-f9zht" Dec 11 02:23:00 crc kubenswrapper[4824]: I1211 02:23:00.364382 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-494lt\" (UniqueName: \"kubernetes.io/projected/14d1e80f-7c53-4c15-a002-eacde62186bc-kube-api-access-494lt\") pod \"nova-cell0-2225-account-create-update-f9zht\" (UID: \"14d1e80f-7c53-4c15-a002-eacde62186bc\") " pod="openstack/nova-cell0-2225-account-create-update-f9zht" Dec 11 02:23:00 crc kubenswrapper[4824]: I1211 02:23:00.445169 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2dchb\" (UniqueName: \"kubernetes.io/projected/78553f90-b619-4ff6-84b4-05835483a29e-kube-api-access-2dchb\") pod \"nova-cell1-7a4e-account-create-update-c7h9j\" (UID: \"78553f90-b619-4ff6-84b4-05835483a29e\") " pod="openstack/nova-cell1-7a4e-account-create-update-c7h9j" Dec 11 02:23:00 crc kubenswrapper[4824]: I1211 02:23:00.445660 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/78553f90-b619-4ff6-84b4-05835483a29e-operator-scripts\") pod \"nova-cell1-7a4e-account-create-update-c7h9j\" (UID: \"78553f90-b619-4ff6-84b4-05835483a29e\") " pod="openstack/nova-cell1-7a4e-account-create-update-c7h9j" Dec 11 02:23:00 crc kubenswrapper[4824]: I1211 02:23:00.446534 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/78553f90-b619-4ff6-84b4-05835483a29e-operator-scripts\") pod \"nova-cell1-7a4e-account-create-update-c7h9j\" (UID: \"78553f90-b619-4ff6-84b4-05835483a29e\") " pod="openstack/nova-cell1-7a4e-account-create-update-c7h9j" Dec 11 02:23:00 crc kubenswrapper[4824]: I1211 02:23:00.464741 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2dchb\" (UniqueName: \"kubernetes.io/projected/78553f90-b619-4ff6-84b4-05835483a29e-kube-api-access-2dchb\") pod \"nova-cell1-7a4e-account-create-update-c7h9j\" (UID: \"78553f90-b619-4ff6-84b4-05835483a29e\") " pod="openstack/nova-cell1-7a4e-account-create-update-c7h9j" Dec 11 02:23:00 crc kubenswrapper[4824]: I1211 02:23:00.517952 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-2225-account-create-update-f9zht" Dec 11 02:23:00 crc kubenswrapper[4824]: I1211 02:23:00.556103 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-kchr8"] Dec 11 02:23:00 crc kubenswrapper[4824]: I1211 02:23:00.647382 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d77732b8-d6e7-4793-a699-cc569bb1ff60" path="/var/lib/kubelet/pods/d77732b8-d6e7-4793-a699-cc569bb1ff60/volumes" Dec 11 02:23:00 crc kubenswrapper[4824]: I1211 02:23:00.647870 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f15f9ed0-98f2-42fe-8006-2142098cc2b6" path="/var/lib/kubelet/pods/f15f9ed0-98f2-42fe-8006-2142098cc2b6/volumes" Dec 11 02:23:00 crc kubenswrapper[4824]: I1211 02:23:00.666859 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-tqtd7"] Dec 11 02:23:00 crc kubenswrapper[4824]: I1211 02:23:00.667157 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-7a4e-account-create-update-c7h9j" Dec 11 02:23:00 crc kubenswrapper[4824]: I1211 02:23:00.806398 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 11 02:23:00 crc kubenswrapper[4824]: I1211 02:23:00.823148 4824 generic.go:334] "Generic (PLEG): container finished" podID="ee999ccf-334d-49ff-85bf-9c6601b97bf1" containerID="22b549fc7d47ca2a6cc9968758dbb602d55f41fdce561681cce9f4f80206b283" exitCode=0 Dec 11 02:23:00 crc kubenswrapper[4824]: I1211 02:23:00.823260 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ee999ccf-334d-49ff-85bf-9c6601b97bf1","Type":"ContainerDied","Data":"22b549fc7d47ca2a6cc9968758dbb602d55f41fdce561681cce9f4f80206b283"} Dec 11 02:23:00 crc kubenswrapper[4824]: I1211 02:23:00.833973 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-tqtd7" event={"ID":"295fe4d0-5910-49d6-b8ac-56ff445666f9","Type":"ContainerStarted","Data":"802dc4619fc2d3c4bf0328050f821b682a37918b3beb30d96dfb44df879a7286"} Dec 11 02:23:00 crc kubenswrapper[4824]: W1211 02:23:00.835102 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod21e824ae_7668_42ad_a039_b7795c13b449.slice/crio-daba60c06319fb10332c69c2556d0a520fc7052df83989937b08cb73639c1dc7 WatchSource:0}: Error finding container daba60c06319fb10332c69c2556d0a520fc7052df83989937b08cb73639c1dc7: Status 404 returned error can't find the container with id daba60c06319fb10332c69c2556d0a520fc7052df83989937b08cb73639c1dc7 Dec 11 02:23:00 crc kubenswrapper[4824]: I1211 02:23:00.836983 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-kchr8" event={"ID":"eb82f620-fb10-4f76-b2eb-adde800579bf","Type":"ContainerStarted","Data":"9b989c8ec4608d3afac27cb73f4c53f059b0fa135849cc91dbe60b75eaa17e85"} Dec 11 02:23:00 crc kubenswrapper[4824]: I1211 02:23:00.837026 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-kchr8" event={"ID":"eb82f620-fb10-4f76-b2eb-adde800579bf","Type":"ContainerStarted","Data":"735dc448330803a4639aaa351528f2bdd7b9ce2e202ee3cf99da602cf3c199b1"} Dec 11 02:23:00 crc kubenswrapper[4824]: I1211 02:23:00.861553 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 11 02:23:00 crc kubenswrapper[4824]: I1211 02:23:00.866932 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-db-create-kchr8" podStartSLOduration=1.866916844 podStartE2EDuration="1.866916844s" podCreationTimestamp="2025-12-11 02:22:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:23:00.854491021 +0000 UTC m=+1322.543528410" watchObservedRunningTime="2025-12-11 02:23:00.866916844 +0000 UTC m=+1322.555954223" Dec 11 02:23:00 crc kubenswrapper[4824]: I1211 02:23:00.928984 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-kpv29"] Dec 11 02:23:00 crc kubenswrapper[4824]: I1211 02:23:00.966435 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee999ccf-334d-49ff-85bf-9c6601b97bf1-combined-ca-bundle\") pod \"ee999ccf-334d-49ff-85bf-9c6601b97bf1\" (UID: \"ee999ccf-334d-49ff-85bf-9c6601b97bf1\") " Dec 11 02:23:00 crc kubenswrapper[4824]: I1211 02:23:00.966748 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee999ccf-334d-49ff-85bf-9c6601b97bf1-scripts\") pod \"ee999ccf-334d-49ff-85bf-9c6601b97bf1\" (UID: \"ee999ccf-334d-49ff-85bf-9c6601b97bf1\") " Dec 11 02:23:00 crc kubenswrapper[4824]: I1211 02:23:00.966831 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee999ccf-334d-49ff-85bf-9c6601b97bf1-public-tls-certs\") pod \"ee999ccf-334d-49ff-85bf-9c6601b97bf1\" (UID: \"ee999ccf-334d-49ff-85bf-9c6601b97bf1\") " Dec 11 02:23:00 crc kubenswrapper[4824]: I1211 02:23:00.966853 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ee999ccf-334d-49ff-85bf-9c6601b97bf1\" (UID: \"ee999ccf-334d-49ff-85bf-9c6601b97bf1\") " Dec 11 02:23:00 crc kubenswrapper[4824]: I1211 02:23:00.966907 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ee999ccf-334d-49ff-85bf-9c6601b97bf1-logs\") pod \"ee999ccf-334d-49ff-85bf-9c6601b97bf1\" (UID: \"ee999ccf-334d-49ff-85bf-9c6601b97bf1\") " Dec 11 02:23:00 crc kubenswrapper[4824]: I1211 02:23:00.966945 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hzlq2\" (UniqueName: \"kubernetes.io/projected/ee999ccf-334d-49ff-85bf-9c6601b97bf1-kube-api-access-hzlq2\") pod \"ee999ccf-334d-49ff-85bf-9c6601b97bf1\" (UID: \"ee999ccf-334d-49ff-85bf-9c6601b97bf1\") " Dec 11 02:23:00 crc kubenswrapper[4824]: I1211 02:23:00.966970 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee999ccf-334d-49ff-85bf-9c6601b97bf1-config-data\") pod \"ee999ccf-334d-49ff-85bf-9c6601b97bf1\" (UID: \"ee999ccf-334d-49ff-85bf-9c6601b97bf1\") " Dec 11 02:23:00 crc kubenswrapper[4824]: I1211 02:23:00.967047 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ee999ccf-334d-49ff-85bf-9c6601b97bf1-httpd-run\") pod \"ee999ccf-334d-49ff-85bf-9c6601b97bf1\" (UID: \"ee999ccf-334d-49ff-85bf-9c6601b97bf1\") " Dec 11 02:23:00 crc kubenswrapper[4824]: I1211 02:23:00.977080 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee999ccf-334d-49ff-85bf-9c6601b97bf1-logs" (OuterVolumeSpecName: "logs") pod "ee999ccf-334d-49ff-85bf-9c6601b97bf1" (UID: "ee999ccf-334d-49ff-85bf-9c6601b97bf1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:23:00 crc kubenswrapper[4824]: I1211 02:23:00.980314 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "glance") pod "ee999ccf-334d-49ff-85bf-9c6601b97bf1" (UID: "ee999ccf-334d-49ff-85bf-9c6601b97bf1"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 11 02:23:00 crc kubenswrapper[4824]: I1211 02:23:00.983278 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee999ccf-334d-49ff-85bf-9c6601b97bf1-scripts" (OuterVolumeSpecName: "scripts") pod "ee999ccf-334d-49ff-85bf-9c6601b97bf1" (UID: "ee999ccf-334d-49ff-85bf-9c6601b97bf1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:23:01 crc kubenswrapper[4824]: I1211 02:23:01.005279 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-32c8-account-create-update-stjs4"] Dec 11 02:23:01 crc kubenswrapper[4824]: I1211 02:23:01.010257 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee999ccf-334d-49ff-85bf-9c6601b97bf1-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "ee999ccf-334d-49ff-85bf-9c6601b97bf1" (UID: "ee999ccf-334d-49ff-85bf-9c6601b97bf1"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:23:01 crc kubenswrapper[4824]: I1211 02:23:01.056178 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee999ccf-334d-49ff-85bf-9c6601b97bf1-kube-api-access-hzlq2" (OuterVolumeSpecName: "kube-api-access-hzlq2") pod "ee999ccf-334d-49ff-85bf-9c6601b97bf1" (UID: "ee999ccf-334d-49ff-85bf-9c6601b97bf1"). InnerVolumeSpecName "kube-api-access-hzlq2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:23:01 crc kubenswrapper[4824]: I1211 02:23:01.060371 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee999ccf-334d-49ff-85bf-9c6601b97bf1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ee999ccf-334d-49ff-85bf-9c6601b97bf1" (UID: "ee999ccf-334d-49ff-85bf-9c6601b97bf1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:23:01 crc kubenswrapper[4824]: I1211 02:23:01.083282 4824 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ee999ccf-334d-49ff-85bf-9c6601b97bf1-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 11 02:23:01 crc kubenswrapper[4824]: I1211 02:23:01.083313 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee999ccf-334d-49ff-85bf-9c6601b97bf1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 02:23:01 crc kubenswrapper[4824]: I1211 02:23:01.083322 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee999ccf-334d-49ff-85bf-9c6601b97bf1-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 02:23:01 crc kubenswrapper[4824]: I1211 02:23:01.083344 4824 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Dec 11 02:23:01 crc kubenswrapper[4824]: I1211 02:23:01.083354 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ee999ccf-334d-49ff-85bf-9c6601b97bf1-logs\") on node \"crc\" DevicePath \"\"" Dec 11 02:23:01 crc kubenswrapper[4824]: I1211 02:23:01.083363 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hzlq2\" (UniqueName: \"kubernetes.io/projected/ee999ccf-334d-49ff-85bf-9c6601b97bf1-kube-api-access-hzlq2\") on node \"crc\" DevicePath \"\"" Dec 11 02:23:01 crc kubenswrapper[4824]: I1211 02:23:01.088607 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee999ccf-334d-49ff-85bf-9c6601b97bf1-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "ee999ccf-334d-49ff-85bf-9c6601b97bf1" (UID: "ee999ccf-334d-49ff-85bf-9c6601b97bf1"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:23:01 crc kubenswrapper[4824]: I1211 02:23:01.140054 4824 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Dec 11 02:23:01 crc kubenswrapper[4824]: I1211 02:23:01.146898 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-2225-account-create-update-f9zht"] Dec 11 02:23:01 crc kubenswrapper[4824]: I1211 02:23:01.160677 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee999ccf-334d-49ff-85bf-9c6601b97bf1-config-data" (OuterVolumeSpecName: "config-data") pod "ee999ccf-334d-49ff-85bf-9c6601b97bf1" (UID: "ee999ccf-334d-49ff-85bf-9c6601b97bf1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:23:01 crc kubenswrapper[4824]: I1211 02:23:01.185198 4824 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Dec 11 02:23:01 crc kubenswrapper[4824]: I1211 02:23:01.185228 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee999ccf-334d-49ff-85bf-9c6601b97bf1-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 02:23:01 crc kubenswrapper[4824]: I1211 02:23:01.185242 4824 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee999ccf-334d-49ff-85bf-9c6601b97bf1-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 11 02:23:01 crc kubenswrapper[4824]: I1211 02:23:01.357215 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-7a4e-account-create-update-c7h9j"] Dec 11 02:23:01 crc kubenswrapper[4824]: I1211 02:23:01.848018 4824 generic.go:334] "Generic (PLEG): container finished" podID="14d1e80f-7c53-4c15-a002-eacde62186bc" containerID="12e254d930693f63ccdfa5e852f7d30d78edf40699e481052f063ab4dc6cba73" exitCode=0 Dec 11 02:23:01 crc kubenswrapper[4824]: I1211 02:23:01.848075 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-2225-account-create-update-f9zht" event={"ID":"14d1e80f-7c53-4c15-a002-eacde62186bc","Type":"ContainerDied","Data":"12e254d930693f63ccdfa5e852f7d30d78edf40699e481052f063ab4dc6cba73"} Dec 11 02:23:01 crc kubenswrapper[4824]: I1211 02:23:01.848100 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-2225-account-create-update-f9zht" event={"ID":"14d1e80f-7c53-4c15-a002-eacde62186bc","Type":"ContainerStarted","Data":"8e9ea3d66f7f3114147086eaa93f181380d18267781a12b525b94a5e5a5b4612"} Dec 11 02:23:01 crc kubenswrapper[4824]: I1211 02:23:01.850638 4824 generic.go:334] "Generic (PLEG): container finished" podID="33034069-6622-45c9-9045-44a5e01682bd" containerID="b8e1b97c75d160789ea16a7148ba70c4ad22979eb01a893cc8e81de1b00e9d1d" exitCode=0 Dec 11 02:23:01 crc kubenswrapper[4824]: I1211 02:23:01.850687 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-kpv29" event={"ID":"33034069-6622-45c9-9045-44a5e01682bd","Type":"ContainerDied","Data":"b8e1b97c75d160789ea16a7148ba70c4ad22979eb01a893cc8e81de1b00e9d1d"} Dec 11 02:23:01 crc kubenswrapper[4824]: I1211 02:23:01.850703 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-kpv29" event={"ID":"33034069-6622-45c9-9045-44a5e01682bd","Type":"ContainerStarted","Data":"f1cbb8d7110d74e008c329d173d639a142b99bb56dd37e3b1fd6e2f6437dfc41"} Dec 11 02:23:01 crc kubenswrapper[4824]: I1211 02:23:01.855817 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"21e824ae-7668-42ad-a039-b7795c13b449","Type":"ContainerStarted","Data":"1de53b66a0c686e91e815f096e88be9dbe58faf50bdda3a0583df3ba0fe5527e"} Dec 11 02:23:01 crc kubenswrapper[4824]: I1211 02:23:01.855839 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"21e824ae-7668-42ad-a039-b7795c13b449","Type":"ContainerStarted","Data":"daba60c06319fb10332c69c2556d0a520fc7052df83989937b08cb73639c1dc7"} Dec 11 02:23:01 crc kubenswrapper[4824]: I1211 02:23:01.856944 4824 generic.go:334] "Generic (PLEG): container finished" podID="295fe4d0-5910-49d6-b8ac-56ff445666f9" containerID="84eddbf85534e881ea9bf082a39086269e957efa5e6265a2e9b4b74a8ae4e37b" exitCode=0 Dec 11 02:23:01 crc kubenswrapper[4824]: I1211 02:23:01.856988 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-tqtd7" event={"ID":"295fe4d0-5910-49d6-b8ac-56ff445666f9","Type":"ContainerDied","Data":"84eddbf85534e881ea9bf082a39086269e957efa5e6265a2e9b4b74a8ae4e37b"} Dec 11 02:23:01 crc kubenswrapper[4824]: I1211 02:23:01.880364 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ee999ccf-334d-49ff-85bf-9c6601b97bf1","Type":"ContainerDied","Data":"313644e840e2f94a47045b60e8e6a6b4395a3c55059af1cec115f6d6a77b43ce"} Dec 11 02:23:01 crc kubenswrapper[4824]: I1211 02:23:01.880420 4824 scope.go:117] "RemoveContainer" containerID="22b549fc7d47ca2a6cc9968758dbb602d55f41fdce561681cce9f4f80206b283" Dec 11 02:23:01 crc kubenswrapper[4824]: I1211 02:23:01.880552 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 11 02:23:01 crc kubenswrapper[4824]: I1211 02:23:01.905743 4824 generic.go:334] "Generic (PLEG): container finished" podID="f0fefd9f-c55d-4bb0-8d9c-b3f7c42ac4d4" containerID="6d0e5e492932c0331515a185fb06051ade0a660f2674731bbdf380d6d80fa182" exitCode=0 Dec 11 02:23:01 crc kubenswrapper[4824]: I1211 02:23:01.906022 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-32c8-account-create-update-stjs4" event={"ID":"f0fefd9f-c55d-4bb0-8d9c-b3f7c42ac4d4","Type":"ContainerDied","Data":"6d0e5e492932c0331515a185fb06051ade0a660f2674731bbdf380d6d80fa182"} Dec 11 02:23:01 crc kubenswrapper[4824]: I1211 02:23:01.906092 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-32c8-account-create-update-stjs4" event={"ID":"f0fefd9f-c55d-4bb0-8d9c-b3f7c42ac4d4","Type":"ContainerStarted","Data":"7c35c0f066068d53d320d966e076e1125251b7ea3a138e89eb217b9c4de37779"} Dec 11 02:23:01 crc kubenswrapper[4824]: I1211 02:23:01.926810 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-7a4e-account-create-update-c7h9j" event={"ID":"78553f90-b619-4ff6-84b4-05835483a29e","Type":"ContainerStarted","Data":"6f30f175ac348afce59696aa8ba26d931409e4ccddc9f33846b288313d383a58"} Dec 11 02:23:01 crc kubenswrapper[4824]: I1211 02:23:01.926854 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-7a4e-account-create-update-c7h9j" event={"ID":"78553f90-b619-4ff6-84b4-05835483a29e","Type":"ContainerStarted","Data":"5c342c01573ca4dc6466ca473d8d16cf7b8eff05738dce828bc268fdedb297e8"} Dec 11 02:23:01 crc kubenswrapper[4824]: I1211 02:23:01.936004 4824 generic.go:334] "Generic (PLEG): container finished" podID="eb82f620-fb10-4f76-b2eb-adde800579bf" containerID="9b989c8ec4608d3afac27cb73f4c53f059b0fa135849cc91dbe60b75eaa17e85" exitCode=0 Dec 11 02:23:01 crc kubenswrapper[4824]: I1211 02:23:01.936329 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-kchr8" event={"ID":"eb82f620-fb10-4f76-b2eb-adde800579bf","Type":"ContainerDied","Data":"9b989c8ec4608d3afac27cb73f4c53f059b0fa135849cc91dbe60b75eaa17e85"} Dec 11 02:23:01 crc kubenswrapper[4824]: I1211 02:23:01.974241 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-7a4e-account-create-update-c7h9j" podStartSLOduration=1.974217716 podStartE2EDuration="1.974217716s" podCreationTimestamp="2025-12-11 02:23:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:23:01.96044556 +0000 UTC m=+1323.649482939" watchObservedRunningTime="2025-12-11 02:23:01.974217716 +0000 UTC m=+1323.663255095" Dec 11 02:23:01 crc kubenswrapper[4824]: I1211 02:23:01.983371 4824 scope.go:117] "RemoveContainer" containerID="2fa8b7e2e0ed1d8d5b5c08d6ab83a548819dfb7841d471d788f368f2ca2743ae" Dec 11 02:23:02 crc kubenswrapper[4824]: I1211 02:23:02.088888 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 11 02:23:02 crc kubenswrapper[4824]: I1211 02:23:02.100282 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 11 02:23:02 crc kubenswrapper[4824]: I1211 02:23:02.120256 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 11 02:23:02 crc kubenswrapper[4824]: E1211 02:23:02.120645 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee999ccf-334d-49ff-85bf-9c6601b97bf1" containerName="glance-httpd" Dec 11 02:23:02 crc kubenswrapper[4824]: I1211 02:23:02.120658 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee999ccf-334d-49ff-85bf-9c6601b97bf1" containerName="glance-httpd" Dec 11 02:23:02 crc kubenswrapper[4824]: E1211 02:23:02.120673 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee999ccf-334d-49ff-85bf-9c6601b97bf1" containerName="glance-log" Dec 11 02:23:02 crc kubenswrapper[4824]: I1211 02:23:02.120679 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee999ccf-334d-49ff-85bf-9c6601b97bf1" containerName="glance-log" Dec 11 02:23:02 crc kubenswrapper[4824]: I1211 02:23:02.120891 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee999ccf-334d-49ff-85bf-9c6601b97bf1" containerName="glance-log" Dec 11 02:23:02 crc kubenswrapper[4824]: I1211 02:23:02.120907 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee999ccf-334d-49ff-85bf-9c6601b97bf1" containerName="glance-httpd" Dec 11 02:23:02 crc kubenswrapper[4824]: I1211 02:23:02.122666 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 11 02:23:02 crc kubenswrapper[4824]: I1211 02:23:02.124705 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 11 02:23:02 crc kubenswrapper[4824]: I1211 02:23:02.125510 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 11 02:23:02 crc kubenswrapper[4824]: I1211 02:23:02.128558 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 11 02:23:02 crc kubenswrapper[4824]: I1211 02:23:02.206972 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac4513fb-74e7-4bf2-8766-ef96066f6c13-config-data\") pod \"glance-default-external-api-0\" (UID: \"ac4513fb-74e7-4bf2-8766-ef96066f6c13\") " pod="openstack/glance-default-external-api-0" Dec 11 02:23:02 crc kubenswrapper[4824]: I1211 02:23:02.207011 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ac4513fb-74e7-4bf2-8766-ef96066f6c13-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"ac4513fb-74e7-4bf2-8766-ef96066f6c13\") " pod="openstack/glance-default-external-api-0" Dec 11 02:23:02 crc kubenswrapper[4824]: I1211 02:23:02.207048 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kp88j\" (UniqueName: \"kubernetes.io/projected/ac4513fb-74e7-4bf2-8766-ef96066f6c13-kube-api-access-kp88j\") pod \"glance-default-external-api-0\" (UID: \"ac4513fb-74e7-4bf2-8766-ef96066f6c13\") " pod="openstack/glance-default-external-api-0" Dec 11 02:23:02 crc kubenswrapper[4824]: I1211 02:23:02.207085 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac4513fb-74e7-4bf2-8766-ef96066f6c13-scripts\") pod \"glance-default-external-api-0\" (UID: \"ac4513fb-74e7-4bf2-8766-ef96066f6c13\") " pod="openstack/glance-default-external-api-0" Dec 11 02:23:02 crc kubenswrapper[4824]: I1211 02:23:02.207144 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac4513fb-74e7-4bf2-8766-ef96066f6c13-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"ac4513fb-74e7-4bf2-8766-ef96066f6c13\") " pod="openstack/glance-default-external-api-0" Dec 11 02:23:02 crc kubenswrapper[4824]: I1211 02:23:02.207163 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ac4513fb-74e7-4bf2-8766-ef96066f6c13-logs\") pod \"glance-default-external-api-0\" (UID: \"ac4513fb-74e7-4bf2-8766-ef96066f6c13\") " pod="openstack/glance-default-external-api-0" Dec 11 02:23:02 crc kubenswrapper[4824]: I1211 02:23:02.207210 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac4513fb-74e7-4bf2-8766-ef96066f6c13-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"ac4513fb-74e7-4bf2-8766-ef96066f6c13\") " pod="openstack/glance-default-external-api-0" Dec 11 02:23:02 crc kubenswrapper[4824]: I1211 02:23:02.207310 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"ac4513fb-74e7-4bf2-8766-ef96066f6c13\") " pod="openstack/glance-default-external-api-0" Dec 11 02:23:02 crc kubenswrapper[4824]: I1211 02:23:02.239031 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 11 02:23:02 crc kubenswrapper[4824]: I1211 02:23:02.314047 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac4513fb-74e7-4bf2-8766-ef96066f6c13-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"ac4513fb-74e7-4bf2-8766-ef96066f6c13\") " pod="openstack/glance-default-external-api-0" Dec 11 02:23:02 crc kubenswrapper[4824]: I1211 02:23:02.314937 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"ac4513fb-74e7-4bf2-8766-ef96066f6c13\") " pod="openstack/glance-default-external-api-0" Dec 11 02:23:02 crc kubenswrapper[4824]: I1211 02:23:02.314974 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac4513fb-74e7-4bf2-8766-ef96066f6c13-config-data\") pod \"glance-default-external-api-0\" (UID: \"ac4513fb-74e7-4bf2-8766-ef96066f6c13\") " pod="openstack/glance-default-external-api-0" Dec 11 02:23:02 crc kubenswrapper[4824]: I1211 02:23:02.314993 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ac4513fb-74e7-4bf2-8766-ef96066f6c13-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"ac4513fb-74e7-4bf2-8766-ef96066f6c13\") " pod="openstack/glance-default-external-api-0" Dec 11 02:23:02 crc kubenswrapper[4824]: I1211 02:23:02.315012 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kp88j\" (UniqueName: \"kubernetes.io/projected/ac4513fb-74e7-4bf2-8766-ef96066f6c13-kube-api-access-kp88j\") pod \"glance-default-external-api-0\" (UID: \"ac4513fb-74e7-4bf2-8766-ef96066f6c13\") " pod="openstack/glance-default-external-api-0" Dec 11 02:23:02 crc kubenswrapper[4824]: I1211 02:23:02.315048 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac4513fb-74e7-4bf2-8766-ef96066f6c13-scripts\") pod \"glance-default-external-api-0\" (UID: \"ac4513fb-74e7-4bf2-8766-ef96066f6c13\") " pod="openstack/glance-default-external-api-0" Dec 11 02:23:02 crc kubenswrapper[4824]: I1211 02:23:02.315075 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac4513fb-74e7-4bf2-8766-ef96066f6c13-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"ac4513fb-74e7-4bf2-8766-ef96066f6c13\") " pod="openstack/glance-default-external-api-0" Dec 11 02:23:02 crc kubenswrapper[4824]: I1211 02:23:02.315094 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ac4513fb-74e7-4bf2-8766-ef96066f6c13-logs\") pod \"glance-default-external-api-0\" (UID: \"ac4513fb-74e7-4bf2-8766-ef96066f6c13\") " pod="openstack/glance-default-external-api-0" Dec 11 02:23:02 crc kubenswrapper[4824]: I1211 02:23:02.315420 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ac4513fb-74e7-4bf2-8766-ef96066f6c13-logs\") pod \"glance-default-external-api-0\" (UID: \"ac4513fb-74e7-4bf2-8766-ef96066f6c13\") " pod="openstack/glance-default-external-api-0" Dec 11 02:23:02 crc kubenswrapper[4824]: I1211 02:23:02.315684 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"ac4513fb-74e7-4bf2-8766-ef96066f6c13\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/glance-default-external-api-0" Dec 11 02:23:02 crc kubenswrapper[4824]: I1211 02:23:02.323900 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac4513fb-74e7-4bf2-8766-ef96066f6c13-scripts\") pod \"glance-default-external-api-0\" (UID: \"ac4513fb-74e7-4bf2-8766-ef96066f6c13\") " pod="openstack/glance-default-external-api-0" Dec 11 02:23:02 crc kubenswrapper[4824]: I1211 02:23:02.323973 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac4513fb-74e7-4bf2-8766-ef96066f6c13-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"ac4513fb-74e7-4bf2-8766-ef96066f6c13\") " pod="openstack/glance-default-external-api-0" Dec 11 02:23:02 crc kubenswrapper[4824]: I1211 02:23:02.332347 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac4513fb-74e7-4bf2-8766-ef96066f6c13-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"ac4513fb-74e7-4bf2-8766-ef96066f6c13\") " pod="openstack/glance-default-external-api-0" Dec 11 02:23:02 crc kubenswrapper[4824]: I1211 02:23:02.332499 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ac4513fb-74e7-4bf2-8766-ef96066f6c13-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"ac4513fb-74e7-4bf2-8766-ef96066f6c13\") " pod="openstack/glance-default-external-api-0" Dec 11 02:23:02 crc kubenswrapper[4824]: I1211 02:23:02.337642 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kp88j\" (UniqueName: \"kubernetes.io/projected/ac4513fb-74e7-4bf2-8766-ef96066f6c13-kube-api-access-kp88j\") pod \"glance-default-external-api-0\" (UID: \"ac4513fb-74e7-4bf2-8766-ef96066f6c13\") " pod="openstack/glance-default-external-api-0" Dec 11 02:23:02 crc kubenswrapper[4824]: I1211 02:23:02.340979 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac4513fb-74e7-4bf2-8766-ef96066f6c13-config-data\") pod \"glance-default-external-api-0\" (UID: \"ac4513fb-74e7-4bf2-8766-ef96066f6c13\") " pod="openstack/glance-default-external-api-0" Dec 11 02:23:02 crc kubenswrapper[4824]: I1211 02:23:02.378335 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"ac4513fb-74e7-4bf2-8766-ef96066f6c13\") " pod="openstack/glance-default-external-api-0" Dec 11 02:23:02 crc kubenswrapper[4824]: I1211 02:23:02.456963 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 11 02:23:02 crc kubenswrapper[4824]: I1211 02:23:02.499517 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 11 02:23:02 crc kubenswrapper[4824]: I1211 02:23:02.619556 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e7794eea-3d90-4bd6-8d69-732d80385243-scripts\") pod \"e7794eea-3d90-4bd6-8d69-732d80385243\" (UID: \"e7794eea-3d90-4bd6-8d69-732d80385243\") " Dec 11 02:23:02 crc kubenswrapper[4824]: I1211 02:23:02.619814 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e7794eea-3d90-4bd6-8d69-732d80385243-internal-tls-certs\") pod \"e7794eea-3d90-4bd6-8d69-732d80385243\" (UID: \"e7794eea-3d90-4bd6-8d69-732d80385243\") " Dec 11 02:23:02 crc kubenswrapper[4824]: I1211 02:23:02.619866 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7794eea-3d90-4bd6-8d69-732d80385243-config-data\") pod \"e7794eea-3d90-4bd6-8d69-732d80385243\" (UID: \"e7794eea-3d90-4bd6-8d69-732d80385243\") " Dec 11 02:23:02 crc kubenswrapper[4824]: I1211 02:23:02.619895 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"e7794eea-3d90-4bd6-8d69-732d80385243\" (UID: \"e7794eea-3d90-4bd6-8d69-732d80385243\") " Dec 11 02:23:02 crc kubenswrapper[4824]: I1211 02:23:02.619922 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-65qz9\" (UniqueName: \"kubernetes.io/projected/e7794eea-3d90-4bd6-8d69-732d80385243-kube-api-access-65qz9\") pod \"e7794eea-3d90-4bd6-8d69-732d80385243\" (UID: \"e7794eea-3d90-4bd6-8d69-732d80385243\") " Dec 11 02:23:02 crc kubenswrapper[4824]: I1211 02:23:02.619955 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e7794eea-3d90-4bd6-8d69-732d80385243-logs\") pod \"e7794eea-3d90-4bd6-8d69-732d80385243\" (UID: \"e7794eea-3d90-4bd6-8d69-732d80385243\") " Dec 11 02:23:02 crc kubenswrapper[4824]: I1211 02:23:02.620034 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e7794eea-3d90-4bd6-8d69-732d80385243-httpd-run\") pod \"e7794eea-3d90-4bd6-8d69-732d80385243\" (UID: \"e7794eea-3d90-4bd6-8d69-732d80385243\") " Dec 11 02:23:02 crc kubenswrapper[4824]: I1211 02:23:02.620081 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7794eea-3d90-4bd6-8d69-732d80385243-combined-ca-bundle\") pod \"e7794eea-3d90-4bd6-8d69-732d80385243\" (UID: \"e7794eea-3d90-4bd6-8d69-732d80385243\") " Dec 11 02:23:02 crc kubenswrapper[4824]: I1211 02:23:02.622142 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e7794eea-3d90-4bd6-8d69-732d80385243-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "e7794eea-3d90-4bd6-8d69-732d80385243" (UID: "e7794eea-3d90-4bd6-8d69-732d80385243"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:23:02 crc kubenswrapper[4824]: I1211 02:23:02.622792 4824 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e7794eea-3d90-4bd6-8d69-732d80385243-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 11 02:23:02 crc kubenswrapper[4824]: I1211 02:23:02.624694 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e7794eea-3d90-4bd6-8d69-732d80385243-logs" (OuterVolumeSpecName: "logs") pod "e7794eea-3d90-4bd6-8d69-732d80385243" (UID: "e7794eea-3d90-4bd6-8d69-732d80385243"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:23:02 crc kubenswrapper[4824]: I1211 02:23:02.630660 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "e7794eea-3d90-4bd6-8d69-732d80385243" (UID: "e7794eea-3d90-4bd6-8d69-732d80385243"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 11 02:23:02 crc kubenswrapper[4824]: I1211 02:23:02.630666 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7794eea-3d90-4bd6-8d69-732d80385243-kube-api-access-65qz9" (OuterVolumeSpecName: "kube-api-access-65qz9") pod "e7794eea-3d90-4bd6-8d69-732d80385243" (UID: "e7794eea-3d90-4bd6-8d69-732d80385243"). InnerVolumeSpecName "kube-api-access-65qz9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:23:02 crc kubenswrapper[4824]: I1211 02:23:02.637021 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7794eea-3d90-4bd6-8d69-732d80385243-scripts" (OuterVolumeSpecName: "scripts") pod "e7794eea-3d90-4bd6-8d69-732d80385243" (UID: "e7794eea-3d90-4bd6-8d69-732d80385243"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:23:02 crc kubenswrapper[4824]: I1211 02:23:02.649735 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee999ccf-334d-49ff-85bf-9c6601b97bf1" path="/var/lib/kubelet/pods/ee999ccf-334d-49ff-85bf-9c6601b97bf1/volumes" Dec 11 02:23:02 crc kubenswrapper[4824]: I1211 02:23:02.684278 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7794eea-3d90-4bd6-8d69-732d80385243-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e7794eea-3d90-4bd6-8d69-732d80385243" (UID: "e7794eea-3d90-4bd6-8d69-732d80385243"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:23:02 crc kubenswrapper[4824]: I1211 02:23:02.690865 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7794eea-3d90-4bd6-8d69-732d80385243-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "e7794eea-3d90-4bd6-8d69-732d80385243" (UID: "e7794eea-3d90-4bd6-8d69-732d80385243"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:23:02 crc kubenswrapper[4824]: I1211 02:23:02.711870 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7794eea-3d90-4bd6-8d69-732d80385243-config-data" (OuterVolumeSpecName: "config-data") pod "e7794eea-3d90-4bd6-8d69-732d80385243" (UID: "e7794eea-3d90-4bd6-8d69-732d80385243"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:23:02 crc kubenswrapper[4824]: I1211 02:23:02.724172 4824 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Dec 11 02:23:02 crc kubenswrapper[4824]: I1211 02:23:02.724225 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-65qz9\" (UniqueName: \"kubernetes.io/projected/e7794eea-3d90-4bd6-8d69-732d80385243-kube-api-access-65qz9\") on node \"crc\" DevicePath \"\"" Dec 11 02:23:02 crc kubenswrapper[4824]: I1211 02:23:02.724237 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e7794eea-3d90-4bd6-8d69-732d80385243-logs\") on node \"crc\" DevicePath \"\"" Dec 11 02:23:02 crc kubenswrapper[4824]: I1211 02:23:02.724245 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7794eea-3d90-4bd6-8d69-732d80385243-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 02:23:02 crc kubenswrapper[4824]: I1211 02:23:02.724254 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e7794eea-3d90-4bd6-8d69-732d80385243-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 02:23:02 crc kubenswrapper[4824]: I1211 02:23:02.724264 4824 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e7794eea-3d90-4bd6-8d69-732d80385243-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 11 02:23:02 crc kubenswrapper[4824]: I1211 02:23:02.724271 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7794eea-3d90-4bd6-8d69-732d80385243-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 02:23:02 crc kubenswrapper[4824]: I1211 02:23:02.759803 4824 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Dec 11 02:23:02 crc kubenswrapper[4824]: I1211 02:23:02.825996 4824 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Dec 11 02:23:02 crc kubenswrapper[4824]: I1211 02:23:02.952030 4824 generic.go:334] "Generic (PLEG): container finished" podID="e7794eea-3d90-4bd6-8d69-732d80385243" containerID="4364c167b32c39ebe179b655d9e1d469552dee7e3454b999439b9e634cc9cf53" exitCode=0 Dec 11 02:23:02 crc kubenswrapper[4824]: I1211 02:23:02.952101 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e7794eea-3d90-4bd6-8d69-732d80385243","Type":"ContainerDied","Data":"4364c167b32c39ebe179b655d9e1d469552dee7e3454b999439b9e634cc9cf53"} Dec 11 02:23:02 crc kubenswrapper[4824]: I1211 02:23:02.952167 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e7794eea-3d90-4bd6-8d69-732d80385243","Type":"ContainerDied","Data":"0168d06ac466cff6709aa1e943886d2ba1cfb35d6cf024f295c68b7819325d02"} Dec 11 02:23:02 crc kubenswrapper[4824]: I1211 02:23:02.952166 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 11 02:23:02 crc kubenswrapper[4824]: I1211 02:23:02.952184 4824 scope.go:117] "RemoveContainer" containerID="4364c167b32c39ebe179b655d9e1d469552dee7e3454b999439b9e634cc9cf53" Dec 11 02:23:02 crc kubenswrapper[4824]: I1211 02:23:02.954695 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"21e824ae-7668-42ad-a039-b7795c13b449","Type":"ContainerStarted","Data":"be84ea003d2e8f4d8a5eb5a0f643a70724703733c9e3b3f7e94e072ac7a05140"} Dec 11 02:23:02 crc kubenswrapper[4824]: I1211 02:23:02.957302 4824 generic.go:334] "Generic (PLEG): container finished" podID="78553f90-b619-4ff6-84b4-05835483a29e" containerID="6f30f175ac348afce59696aa8ba26d931409e4ccddc9f33846b288313d383a58" exitCode=0 Dec 11 02:23:02 crc kubenswrapper[4824]: I1211 02:23:02.957432 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-7a4e-account-create-update-c7h9j" event={"ID":"78553f90-b619-4ff6-84b4-05835483a29e","Type":"ContainerDied","Data":"6f30f175ac348afce59696aa8ba26d931409e4ccddc9f33846b288313d383a58"} Dec 11 02:23:03 crc kubenswrapper[4824]: I1211 02:23:03.041294 4824 scope.go:117] "RemoveContainer" containerID="8411d991de0195412273e5c06a0d59951db008bb73091da57111623429573c22" Dec 11 02:23:03 crc kubenswrapper[4824]: I1211 02:23:03.045192 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 11 02:23:03 crc kubenswrapper[4824]: I1211 02:23:03.052647 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 11 02:23:03 crc kubenswrapper[4824]: I1211 02:23:03.072817 4824 scope.go:117] "RemoveContainer" containerID="4364c167b32c39ebe179b655d9e1d469552dee7e3454b999439b9e634cc9cf53" Dec 11 02:23:03 crc kubenswrapper[4824]: I1211 02:23:03.078091 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 11 02:23:03 crc kubenswrapper[4824]: E1211 02:23:03.078543 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7794eea-3d90-4bd6-8d69-732d80385243" containerName="glance-httpd" Dec 11 02:23:03 crc kubenswrapper[4824]: I1211 02:23:03.078559 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7794eea-3d90-4bd6-8d69-732d80385243" containerName="glance-httpd" Dec 11 02:23:03 crc kubenswrapper[4824]: E1211 02:23:03.078570 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7794eea-3d90-4bd6-8d69-732d80385243" containerName="glance-log" Dec 11 02:23:03 crc kubenswrapper[4824]: I1211 02:23:03.078576 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7794eea-3d90-4bd6-8d69-732d80385243" containerName="glance-log" Dec 11 02:23:03 crc kubenswrapper[4824]: I1211 02:23:03.078771 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7794eea-3d90-4bd6-8d69-732d80385243" containerName="glance-httpd" Dec 11 02:23:03 crc kubenswrapper[4824]: I1211 02:23:03.078790 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7794eea-3d90-4bd6-8d69-732d80385243" containerName="glance-log" Dec 11 02:23:03 crc kubenswrapper[4824]: E1211 02:23:03.078870 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4364c167b32c39ebe179b655d9e1d469552dee7e3454b999439b9e634cc9cf53\": container with ID starting with 4364c167b32c39ebe179b655d9e1d469552dee7e3454b999439b9e634cc9cf53 not found: ID does not exist" containerID="4364c167b32c39ebe179b655d9e1d469552dee7e3454b999439b9e634cc9cf53" Dec 11 02:23:03 crc kubenswrapper[4824]: I1211 02:23:03.078907 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4364c167b32c39ebe179b655d9e1d469552dee7e3454b999439b9e634cc9cf53"} err="failed to get container status \"4364c167b32c39ebe179b655d9e1d469552dee7e3454b999439b9e634cc9cf53\": rpc error: code = NotFound desc = could not find container \"4364c167b32c39ebe179b655d9e1d469552dee7e3454b999439b9e634cc9cf53\": container with ID starting with 4364c167b32c39ebe179b655d9e1d469552dee7e3454b999439b9e634cc9cf53 not found: ID does not exist" Dec 11 02:23:03 crc kubenswrapper[4824]: I1211 02:23:03.078932 4824 scope.go:117] "RemoveContainer" containerID="8411d991de0195412273e5c06a0d59951db008bb73091da57111623429573c22" Dec 11 02:23:03 crc kubenswrapper[4824]: I1211 02:23:03.079760 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 11 02:23:03 crc kubenswrapper[4824]: E1211 02:23:03.085961 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8411d991de0195412273e5c06a0d59951db008bb73091da57111623429573c22\": container with ID starting with 8411d991de0195412273e5c06a0d59951db008bb73091da57111623429573c22 not found: ID does not exist" containerID="8411d991de0195412273e5c06a0d59951db008bb73091da57111623429573c22" Dec 11 02:23:03 crc kubenswrapper[4824]: I1211 02:23:03.086001 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8411d991de0195412273e5c06a0d59951db008bb73091da57111623429573c22"} err="failed to get container status \"8411d991de0195412273e5c06a0d59951db008bb73091da57111623429573c22\": rpc error: code = NotFound desc = could not find container \"8411d991de0195412273e5c06a0d59951db008bb73091da57111623429573c22\": container with ID starting with 8411d991de0195412273e5c06a0d59951db008bb73091da57111623429573c22 not found: ID does not exist" Dec 11 02:23:03 crc kubenswrapper[4824]: I1211 02:23:03.086786 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 11 02:23:03 crc kubenswrapper[4824]: I1211 02:23:03.086932 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 11 02:23:03 crc kubenswrapper[4824]: I1211 02:23:03.087181 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 11 02:23:03 crc kubenswrapper[4824]: I1211 02:23:03.094407 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 11 02:23:03 crc kubenswrapper[4824]: I1211 02:23:03.233294 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6776fbd-a091-4715-9942-4dafac0630e2-config-data\") pod \"glance-default-internal-api-0\" (UID: \"e6776fbd-a091-4715-9942-4dafac0630e2\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:23:03 crc kubenswrapper[4824]: I1211 02:23:03.233517 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6776fbd-a091-4715-9942-4dafac0630e2-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"e6776fbd-a091-4715-9942-4dafac0630e2\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:23:03 crc kubenswrapper[4824]: I1211 02:23:03.233658 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e6776fbd-a091-4715-9942-4dafac0630e2-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"e6776fbd-a091-4715-9942-4dafac0630e2\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:23:03 crc kubenswrapper[4824]: I1211 02:23:03.233720 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e6776fbd-a091-4715-9942-4dafac0630e2-logs\") pod \"glance-default-internal-api-0\" (UID: \"e6776fbd-a091-4715-9942-4dafac0630e2\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:23:03 crc kubenswrapper[4824]: I1211 02:23:03.233748 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e6776fbd-a091-4715-9942-4dafac0630e2-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"e6776fbd-a091-4715-9942-4dafac0630e2\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:23:03 crc kubenswrapper[4824]: I1211 02:23:03.233870 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e6776fbd-a091-4715-9942-4dafac0630e2-scripts\") pod \"glance-default-internal-api-0\" (UID: \"e6776fbd-a091-4715-9942-4dafac0630e2\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:23:03 crc kubenswrapper[4824]: I1211 02:23:03.233895 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9bbl6\" (UniqueName: \"kubernetes.io/projected/e6776fbd-a091-4715-9942-4dafac0630e2-kube-api-access-9bbl6\") pod \"glance-default-internal-api-0\" (UID: \"e6776fbd-a091-4715-9942-4dafac0630e2\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:23:03 crc kubenswrapper[4824]: I1211 02:23:03.233990 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"e6776fbd-a091-4715-9942-4dafac0630e2\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:23:03 crc kubenswrapper[4824]: I1211 02:23:03.336688 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6776fbd-a091-4715-9942-4dafac0630e2-config-data\") pod \"glance-default-internal-api-0\" (UID: \"e6776fbd-a091-4715-9942-4dafac0630e2\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:23:03 crc kubenswrapper[4824]: I1211 02:23:03.337059 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6776fbd-a091-4715-9942-4dafac0630e2-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"e6776fbd-a091-4715-9942-4dafac0630e2\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:23:03 crc kubenswrapper[4824]: I1211 02:23:03.337141 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e6776fbd-a091-4715-9942-4dafac0630e2-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"e6776fbd-a091-4715-9942-4dafac0630e2\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:23:03 crc kubenswrapper[4824]: I1211 02:23:03.337168 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e6776fbd-a091-4715-9942-4dafac0630e2-logs\") pod \"glance-default-internal-api-0\" (UID: \"e6776fbd-a091-4715-9942-4dafac0630e2\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:23:03 crc kubenswrapper[4824]: I1211 02:23:03.337188 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e6776fbd-a091-4715-9942-4dafac0630e2-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"e6776fbd-a091-4715-9942-4dafac0630e2\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:23:03 crc kubenswrapper[4824]: I1211 02:23:03.337239 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e6776fbd-a091-4715-9942-4dafac0630e2-scripts\") pod \"glance-default-internal-api-0\" (UID: \"e6776fbd-a091-4715-9942-4dafac0630e2\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:23:03 crc kubenswrapper[4824]: I1211 02:23:03.337260 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9bbl6\" (UniqueName: \"kubernetes.io/projected/e6776fbd-a091-4715-9942-4dafac0630e2-kube-api-access-9bbl6\") pod \"glance-default-internal-api-0\" (UID: \"e6776fbd-a091-4715-9942-4dafac0630e2\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:23:03 crc kubenswrapper[4824]: I1211 02:23:03.337302 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"e6776fbd-a091-4715-9942-4dafac0630e2\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:23:03 crc kubenswrapper[4824]: I1211 02:23:03.338151 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e6776fbd-a091-4715-9942-4dafac0630e2-logs\") pod \"glance-default-internal-api-0\" (UID: \"e6776fbd-a091-4715-9942-4dafac0630e2\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:23:03 crc kubenswrapper[4824]: I1211 02:23:03.338298 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"e6776fbd-a091-4715-9942-4dafac0630e2\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-internal-api-0" Dec 11 02:23:03 crc kubenswrapper[4824]: I1211 02:23:03.338969 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e6776fbd-a091-4715-9942-4dafac0630e2-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"e6776fbd-a091-4715-9942-4dafac0630e2\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:23:03 crc kubenswrapper[4824]: I1211 02:23:03.342466 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6776fbd-a091-4715-9942-4dafac0630e2-config-data\") pod \"glance-default-internal-api-0\" (UID: \"e6776fbd-a091-4715-9942-4dafac0630e2\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:23:03 crc kubenswrapper[4824]: I1211 02:23:03.345935 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6776fbd-a091-4715-9942-4dafac0630e2-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"e6776fbd-a091-4715-9942-4dafac0630e2\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:23:03 crc kubenswrapper[4824]: I1211 02:23:03.348972 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e6776fbd-a091-4715-9942-4dafac0630e2-scripts\") pod \"glance-default-internal-api-0\" (UID: \"e6776fbd-a091-4715-9942-4dafac0630e2\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:23:03 crc kubenswrapper[4824]: I1211 02:23:03.350601 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e6776fbd-a091-4715-9942-4dafac0630e2-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"e6776fbd-a091-4715-9942-4dafac0630e2\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:23:03 crc kubenswrapper[4824]: I1211 02:23:03.355715 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-kpv29" Dec 11 02:23:03 crc kubenswrapper[4824]: I1211 02:23:03.369099 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9bbl6\" (UniqueName: \"kubernetes.io/projected/e6776fbd-a091-4715-9942-4dafac0630e2-kube-api-access-9bbl6\") pod \"glance-default-internal-api-0\" (UID: \"e6776fbd-a091-4715-9942-4dafac0630e2\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:23:03 crc kubenswrapper[4824]: I1211 02:23:03.381760 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"e6776fbd-a091-4715-9942-4dafac0630e2\") " pod="openstack/glance-default-internal-api-0" Dec 11 02:23:03 crc kubenswrapper[4824]: I1211 02:23:03.438090 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7zk6s\" (UniqueName: \"kubernetes.io/projected/33034069-6622-45c9-9045-44a5e01682bd-kube-api-access-7zk6s\") pod \"33034069-6622-45c9-9045-44a5e01682bd\" (UID: \"33034069-6622-45c9-9045-44a5e01682bd\") " Dec 11 02:23:03 crc kubenswrapper[4824]: I1211 02:23:03.438164 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/33034069-6622-45c9-9045-44a5e01682bd-operator-scripts\") pod \"33034069-6622-45c9-9045-44a5e01682bd\" (UID: \"33034069-6622-45c9-9045-44a5e01682bd\") " Dec 11 02:23:03 crc kubenswrapper[4824]: I1211 02:23:03.438934 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/33034069-6622-45c9-9045-44a5e01682bd-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "33034069-6622-45c9-9045-44a5e01682bd" (UID: "33034069-6622-45c9-9045-44a5e01682bd"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:23:03 crc kubenswrapper[4824]: I1211 02:23:03.441064 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/33034069-6622-45c9-9045-44a5e01682bd-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 02:23:03 crc kubenswrapper[4824]: I1211 02:23:03.442190 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33034069-6622-45c9-9045-44a5e01682bd-kube-api-access-7zk6s" (OuterVolumeSpecName: "kube-api-access-7zk6s") pod "33034069-6622-45c9-9045-44a5e01682bd" (UID: "33034069-6622-45c9-9045-44a5e01682bd"). InnerVolumeSpecName "kube-api-access-7zk6s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:23:03 crc kubenswrapper[4824]: I1211 02:23:03.449899 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 11 02:23:03 crc kubenswrapper[4824]: I1211 02:23:03.518381 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-2225-account-create-update-f9zht" Dec 11 02:23:03 crc kubenswrapper[4824]: I1211 02:23:03.543187 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7zk6s\" (UniqueName: \"kubernetes.io/projected/33034069-6622-45c9-9045-44a5e01682bd-kube-api-access-7zk6s\") on node \"crc\" DevicePath \"\"" Dec 11 02:23:03 crc kubenswrapper[4824]: I1211 02:23:03.643706 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/14d1e80f-7c53-4c15-a002-eacde62186bc-operator-scripts\") pod \"14d1e80f-7c53-4c15-a002-eacde62186bc\" (UID: \"14d1e80f-7c53-4c15-a002-eacde62186bc\") " Dec 11 02:23:03 crc kubenswrapper[4824]: I1211 02:23:03.644054 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-494lt\" (UniqueName: \"kubernetes.io/projected/14d1e80f-7c53-4c15-a002-eacde62186bc-kube-api-access-494lt\") pod \"14d1e80f-7c53-4c15-a002-eacde62186bc\" (UID: \"14d1e80f-7c53-4c15-a002-eacde62186bc\") " Dec 11 02:23:03 crc kubenswrapper[4824]: I1211 02:23:03.644810 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14d1e80f-7c53-4c15-a002-eacde62186bc-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "14d1e80f-7c53-4c15-a002-eacde62186bc" (UID: "14d1e80f-7c53-4c15-a002-eacde62186bc"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:23:03 crc kubenswrapper[4824]: I1211 02:23:03.653345 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14d1e80f-7c53-4c15-a002-eacde62186bc-kube-api-access-494lt" (OuterVolumeSpecName: "kube-api-access-494lt") pod "14d1e80f-7c53-4c15-a002-eacde62186bc" (UID: "14d1e80f-7c53-4c15-a002-eacde62186bc"). InnerVolumeSpecName "kube-api-access-494lt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:23:03 crc kubenswrapper[4824]: I1211 02:23:03.699553 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-tqtd7" Dec 11 02:23:03 crc kubenswrapper[4824]: I1211 02:23:03.727576 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-kchr8" Dec 11 02:23:03 crc kubenswrapper[4824]: I1211 02:23:03.753260 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/14d1e80f-7c53-4c15-a002-eacde62186bc-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 02:23:03 crc kubenswrapper[4824]: I1211 02:23:03.753289 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-494lt\" (UniqueName: \"kubernetes.io/projected/14d1e80f-7c53-4c15-a002-eacde62186bc-kube-api-access-494lt\") on node \"crc\" DevicePath \"\"" Dec 11 02:23:03 crc kubenswrapper[4824]: I1211 02:23:03.755807 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-32c8-account-create-update-stjs4" Dec 11 02:23:03 crc kubenswrapper[4824]: I1211 02:23:03.854690 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vm67l\" (UniqueName: \"kubernetes.io/projected/eb82f620-fb10-4f76-b2eb-adde800579bf-kube-api-access-vm67l\") pod \"eb82f620-fb10-4f76-b2eb-adde800579bf\" (UID: \"eb82f620-fb10-4f76-b2eb-adde800579bf\") " Dec 11 02:23:03 crc kubenswrapper[4824]: I1211 02:23:03.854742 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sn9sl\" (UniqueName: \"kubernetes.io/projected/295fe4d0-5910-49d6-b8ac-56ff445666f9-kube-api-access-sn9sl\") pod \"295fe4d0-5910-49d6-b8ac-56ff445666f9\" (UID: \"295fe4d0-5910-49d6-b8ac-56ff445666f9\") " Dec 11 02:23:03 crc kubenswrapper[4824]: I1211 02:23:03.854769 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eb82f620-fb10-4f76-b2eb-adde800579bf-operator-scripts\") pod \"eb82f620-fb10-4f76-b2eb-adde800579bf\" (UID: \"eb82f620-fb10-4f76-b2eb-adde800579bf\") " Dec 11 02:23:03 crc kubenswrapper[4824]: I1211 02:23:03.854805 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/295fe4d0-5910-49d6-b8ac-56ff445666f9-operator-scripts\") pod \"295fe4d0-5910-49d6-b8ac-56ff445666f9\" (UID: \"295fe4d0-5910-49d6-b8ac-56ff445666f9\") " Dec 11 02:23:03 crc kubenswrapper[4824]: I1211 02:23:03.854863 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bpm9h\" (UniqueName: \"kubernetes.io/projected/f0fefd9f-c55d-4bb0-8d9c-b3f7c42ac4d4-kube-api-access-bpm9h\") pod \"f0fefd9f-c55d-4bb0-8d9c-b3f7c42ac4d4\" (UID: \"f0fefd9f-c55d-4bb0-8d9c-b3f7c42ac4d4\") " Dec 11 02:23:03 crc kubenswrapper[4824]: I1211 02:23:03.855151 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f0fefd9f-c55d-4bb0-8d9c-b3f7c42ac4d4-operator-scripts\") pod \"f0fefd9f-c55d-4bb0-8d9c-b3f7c42ac4d4\" (UID: \"f0fefd9f-c55d-4bb0-8d9c-b3f7c42ac4d4\") " Dec 11 02:23:03 crc kubenswrapper[4824]: I1211 02:23:03.859741 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb82f620-fb10-4f76-b2eb-adde800579bf-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "eb82f620-fb10-4f76-b2eb-adde800579bf" (UID: "eb82f620-fb10-4f76-b2eb-adde800579bf"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:23:03 crc kubenswrapper[4824]: I1211 02:23:03.860607 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f0fefd9f-c55d-4bb0-8d9c-b3f7c42ac4d4-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f0fefd9f-c55d-4bb0-8d9c-b3f7c42ac4d4" (UID: "f0fefd9f-c55d-4bb0-8d9c-b3f7c42ac4d4"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:23:03 crc kubenswrapper[4824]: I1211 02:23:03.861127 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/295fe4d0-5910-49d6-b8ac-56ff445666f9-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "295fe4d0-5910-49d6-b8ac-56ff445666f9" (UID: "295fe4d0-5910-49d6-b8ac-56ff445666f9"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:23:03 crc kubenswrapper[4824]: I1211 02:23:03.861282 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0fefd9f-c55d-4bb0-8d9c-b3f7c42ac4d4-kube-api-access-bpm9h" (OuterVolumeSpecName: "kube-api-access-bpm9h") pod "f0fefd9f-c55d-4bb0-8d9c-b3f7c42ac4d4" (UID: "f0fefd9f-c55d-4bb0-8d9c-b3f7c42ac4d4"). InnerVolumeSpecName "kube-api-access-bpm9h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:23:03 crc kubenswrapper[4824]: I1211 02:23:03.864258 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/295fe4d0-5910-49d6-b8ac-56ff445666f9-kube-api-access-sn9sl" (OuterVolumeSpecName: "kube-api-access-sn9sl") pod "295fe4d0-5910-49d6-b8ac-56ff445666f9" (UID: "295fe4d0-5910-49d6-b8ac-56ff445666f9"). InnerVolumeSpecName "kube-api-access-sn9sl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:23:03 crc kubenswrapper[4824]: I1211 02:23:03.865830 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb82f620-fb10-4f76-b2eb-adde800579bf-kube-api-access-vm67l" (OuterVolumeSpecName: "kube-api-access-vm67l") pod "eb82f620-fb10-4f76-b2eb-adde800579bf" (UID: "eb82f620-fb10-4f76-b2eb-adde800579bf"). InnerVolumeSpecName "kube-api-access-vm67l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:23:03 crc kubenswrapper[4824]: I1211 02:23:03.958141 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vm67l\" (UniqueName: \"kubernetes.io/projected/eb82f620-fb10-4f76-b2eb-adde800579bf-kube-api-access-vm67l\") on node \"crc\" DevicePath \"\"" Dec 11 02:23:03 crc kubenswrapper[4824]: I1211 02:23:03.958177 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sn9sl\" (UniqueName: \"kubernetes.io/projected/295fe4d0-5910-49d6-b8ac-56ff445666f9-kube-api-access-sn9sl\") on node \"crc\" DevicePath \"\"" Dec 11 02:23:03 crc kubenswrapper[4824]: I1211 02:23:03.958188 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eb82f620-fb10-4f76-b2eb-adde800579bf-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 02:23:03 crc kubenswrapper[4824]: I1211 02:23:03.958197 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/295fe4d0-5910-49d6-b8ac-56ff445666f9-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 02:23:03 crc kubenswrapper[4824]: I1211 02:23:03.958205 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bpm9h\" (UniqueName: \"kubernetes.io/projected/f0fefd9f-c55d-4bb0-8d9c-b3f7c42ac4d4-kube-api-access-bpm9h\") on node \"crc\" DevicePath \"\"" Dec 11 02:23:03 crc kubenswrapper[4824]: I1211 02:23:03.958213 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f0fefd9f-c55d-4bb0-8d9c-b3f7c42ac4d4-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 02:23:03 crc kubenswrapper[4824]: I1211 02:23:03.980229 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-tqtd7" Dec 11 02:23:03 crc kubenswrapper[4824]: I1211 02:23:03.980996 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-tqtd7" event={"ID":"295fe4d0-5910-49d6-b8ac-56ff445666f9","Type":"ContainerDied","Data":"802dc4619fc2d3c4bf0328050f821b682a37918b3beb30d96dfb44df879a7286"} Dec 11 02:23:03 crc kubenswrapper[4824]: I1211 02:23:03.981040 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="802dc4619fc2d3c4bf0328050f821b682a37918b3beb30d96dfb44df879a7286" Dec 11 02:23:03 crc kubenswrapper[4824]: I1211 02:23:03.989071 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-kchr8" event={"ID":"eb82f620-fb10-4f76-b2eb-adde800579bf","Type":"ContainerDied","Data":"735dc448330803a4639aaa351528f2bdd7b9ce2e202ee3cf99da602cf3c199b1"} Dec 11 02:23:03 crc kubenswrapper[4824]: I1211 02:23:03.989141 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="735dc448330803a4639aaa351528f2bdd7b9ce2e202ee3cf99da602cf3c199b1" Dec 11 02:23:03 crc kubenswrapper[4824]: I1211 02:23:03.989214 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-kchr8" Dec 11 02:23:04 crc kubenswrapper[4824]: I1211 02:23:04.002322 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ac4513fb-74e7-4bf2-8766-ef96066f6c13","Type":"ContainerStarted","Data":"89142fec9d57da4263f55c1f4f360928a586a3ff2fdefbea2211607dcc9299cb"} Dec 11 02:23:04 crc kubenswrapper[4824]: I1211 02:23:04.002363 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ac4513fb-74e7-4bf2-8766-ef96066f6c13","Type":"ContainerStarted","Data":"bcc193c61becc42985695f86d0aedcd6a5a29cd82c5d441abcf57888640e1235"} Dec 11 02:23:04 crc kubenswrapper[4824]: I1211 02:23:04.010560 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"21e824ae-7668-42ad-a039-b7795c13b449","Type":"ContainerStarted","Data":"7cae66953fc8d23c4b105dbcd073ffc20b212d26fde6175d619b470cdaf772f4"} Dec 11 02:23:04 crc kubenswrapper[4824]: I1211 02:23:04.012192 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-32c8-account-create-update-stjs4" event={"ID":"f0fefd9f-c55d-4bb0-8d9c-b3f7c42ac4d4","Type":"ContainerDied","Data":"7c35c0f066068d53d320d966e076e1125251b7ea3a138e89eb217b9c4de37779"} Dec 11 02:23:04 crc kubenswrapper[4824]: I1211 02:23:04.012217 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7c35c0f066068d53d320d966e076e1125251b7ea3a138e89eb217b9c4de37779" Dec 11 02:23:04 crc kubenswrapper[4824]: I1211 02:23:04.012266 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-32c8-account-create-update-stjs4" Dec 11 02:23:04 crc kubenswrapper[4824]: I1211 02:23:04.021442 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-2225-account-create-update-f9zht" event={"ID":"14d1e80f-7c53-4c15-a002-eacde62186bc","Type":"ContainerDied","Data":"8e9ea3d66f7f3114147086eaa93f181380d18267781a12b525b94a5e5a5b4612"} Dec 11 02:23:04 crc kubenswrapper[4824]: I1211 02:23:04.021468 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8e9ea3d66f7f3114147086eaa93f181380d18267781a12b525b94a5e5a5b4612" Dec 11 02:23:04 crc kubenswrapper[4824]: I1211 02:23:04.021487 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-2225-account-create-update-f9zht" Dec 11 02:23:04 crc kubenswrapper[4824]: I1211 02:23:04.025883 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-kpv29" Dec 11 02:23:04 crc kubenswrapper[4824]: I1211 02:23:04.036587 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-kpv29" event={"ID":"33034069-6622-45c9-9045-44a5e01682bd","Type":"ContainerDied","Data":"f1cbb8d7110d74e008c329d173d639a142b99bb56dd37e3b1fd6e2f6437dfc41"} Dec 11 02:23:04 crc kubenswrapper[4824]: I1211 02:23:04.036622 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f1cbb8d7110d74e008c329d173d639a142b99bb56dd37e3b1fd6e2f6437dfc41" Dec 11 02:23:04 crc kubenswrapper[4824]: I1211 02:23:04.127779 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 11 02:23:04 crc kubenswrapper[4824]: I1211 02:23:04.433440 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-7a4e-account-create-update-c7h9j" Dec 11 02:23:04 crc kubenswrapper[4824]: I1211 02:23:04.577795 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2dchb\" (UniqueName: \"kubernetes.io/projected/78553f90-b619-4ff6-84b4-05835483a29e-kube-api-access-2dchb\") pod \"78553f90-b619-4ff6-84b4-05835483a29e\" (UID: \"78553f90-b619-4ff6-84b4-05835483a29e\") " Dec 11 02:23:04 crc kubenswrapper[4824]: I1211 02:23:04.577996 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/78553f90-b619-4ff6-84b4-05835483a29e-operator-scripts\") pod \"78553f90-b619-4ff6-84b4-05835483a29e\" (UID: \"78553f90-b619-4ff6-84b4-05835483a29e\") " Dec 11 02:23:04 crc kubenswrapper[4824]: I1211 02:23:04.579107 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/78553f90-b619-4ff6-84b4-05835483a29e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "78553f90-b619-4ff6-84b4-05835483a29e" (UID: "78553f90-b619-4ff6-84b4-05835483a29e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:23:04 crc kubenswrapper[4824]: I1211 02:23:04.582838 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78553f90-b619-4ff6-84b4-05835483a29e-kube-api-access-2dchb" (OuterVolumeSpecName: "kube-api-access-2dchb") pod "78553f90-b619-4ff6-84b4-05835483a29e" (UID: "78553f90-b619-4ff6-84b4-05835483a29e"). InnerVolumeSpecName "kube-api-access-2dchb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:23:04 crc kubenswrapper[4824]: I1211 02:23:04.646224 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7794eea-3d90-4bd6-8d69-732d80385243" path="/var/lib/kubelet/pods/e7794eea-3d90-4bd6-8d69-732d80385243/volumes" Dec 11 02:23:04 crc kubenswrapper[4824]: I1211 02:23:04.680327 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/78553f90-b619-4ff6-84b4-05835483a29e-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 02:23:04 crc kubenswrapper[4824]: I1211 02:23:04.680365 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2dchb\" (UniqueName: \"kubernetes.io/projected/78553f90-b619-4ff6-84b4-05835483a29e-kube-api-access-2dchb\") on node \"crc\" DevicePath \"\"" Dec 11 02:23:05 crc kubenswrapper[4824]: I1211 02:23:05.037514 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"21e824ae-7668-42ad-a039-b7795c13b449","Type":"ContainerStarted","Data":"4f692b5f18b6ad468c6348b5cac9f5a4e880abcd6b639834dcade08d9753af20"} Dec 11 02:23:05 crc kubenswrapper[4824]: I1211 02:23:05.037904 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="21e824ae-7668-42ad-a039-b7795c13b449" containerName="ceilometer-central-agent" containerID="cri-o://1de53b66a0c686e91e815f096e88be9dbe58faf50bdda3a0583df3ba0fe5527e" gracePeriod=30 Dec 11 02:23:05 crc kubenswrapper[4824]: I1211 02:23:05.038173 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 11 02:23:05 crc kubenswrapper[4824]: I1211 02:23:05.038246 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="21e824ae-7668-42ad-a039-b7795c13b449" containerName="proxy-httpd" containerID="cri-o://4f692b5f18b6ad468c6348b5cac9f5a4e880abcd6b639834dcade08d9753af20" gracePeriod=30 Dec 11 02:23:05 crc kubenswrapper[4824]: I1211 02:23:05.038297 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="21e824ae-7668-42ad-a039-b7795c13b449" containerName="sg-core" containerID="cri-o://7cae66953fc8d23c4b105dbcd073ffc20b212d26fde6175d619b470cdaf772f4" gracePeriod=30 Dec 11 02:23:05 crc kubenswrapper[4824]: I1211 02:23:05.038339 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="21e824ae-7668-42ad-a039-b7795c13b449" containerName="ceilometer-notification-agent" containerID="cri-o://be84ea003d2e8f4d8a5eb5a0f643a70724703733c9e3b3f7e94e072ac7a05140" gracePeriod=30 Dec 11 02:23:05 crc kubenswrapper[4824]: I1211 02:23:05.047223 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e6776fbd-a091-4715-9942-4dafac0630e2","Type":"ContainerStarted","Data":"2c9641f21252c01beb7403736a4ba87c8705676f37fe29c6e7ab2b57f3a0b5e3"} Dec 11 02:23:05 crc kubenswrapper[4824]: I1211 02:23:05.047258 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e6776fbd-a091-4715-9942-4dafac0630e2","Type":"ContainerStarted","Data":"1a701e67a90f0bd9a1106e5463b7e445b8a8dc5d16d5ae754fc47f00adf68e2c"} Dec 11 02:23:05 crc kubenswrapper[4824]: I1211 02:23:05.057509 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-7a4e-account-create-update-c7h9j" Dec 11 02:23:05 crc kubenswrapper[4824]: I1211 02:23:05.057505 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-7a4e-account-create-update-c7h9j" event={"ID":"78553f90-b619-4ff6-84b4-05835483a29e","Type":"ContainerDied","Data":"5c342c01573ca4dc6466ca473d8d16cf7b8eff05738dce828bc268fdedb297e8"} Dec 11 02:23:05 crc kubenswrapper[4824]: I1211 02:23:05.057639 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5c342c01573ca4dc6466ca473d8d16cf7b8eff05738dce828bc268fdedb297e8" Dec 11 02:23:05 crc kubenswrapper[4824]: I1211 02:23:05.065975 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.340194716 podStartE2EDuration="6.065957302s" podCreationTimestamp="2025-12-11 02:22:59 +0000 UTC" firstStartedPulling="2025-12-11 02:23:00.841431113 +0000 UTC m=+1322.530468492" lastFinishedPulling="2025-12-11 02:23:04.567193699 +0000 UTC m=+1326.256231078" observedRunningTime="2025-12-11 02:23:05.059737945 +0000 UTC m=+1326.748775324" watchObservedRunningTime="2025-12-11 02:23:05.065957302 +0000 UTC m=+1326.754994681" Dec 11 02:23:05 crc kubenswrapper[4824]: I1211 02:23:05.079067 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ac4513fb-74e7-4bf2-8766-ef96066f6c13","Type":"ContainerStarted","Data":"823cd6369314d635ab9d0807311b4bc674564b9f3022af266c4c7b305c4f5710"} Dec 11 02:23:05 crc kubenswrapper[4824]: I1211 02:23:05.107393 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.107377851 podStartE2EDuration="3.107377851s" podCreationTimestamp="2025-12-11 02:23:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:23:05.103025703 +0000 UTC m=+1326.792063082" watchObservedRunningTime="2025-12-11 02:23:05.107377851 +0000 UTC m=+1326.796415230" Dec 11 02:23:05 crc kubenswrapper[4824]: I1211 02:23:05.441022 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-lx54d"] Dec 11 02:23:05 crc kubenswrapper[4824]: E1211 02:23:05.441384 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78553f90-b619-4ff6-84b4-05835483a29e" containerName="mariadb-account-create-update" Dec 11 02:23:05 crc kubenswrapper[4824]: I1211 02:23:05.441401 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="78553f90-b619-4ff6-84b4-05835483a29e" containerName="mariadb-account-create-update" Dec 11 02:23:05 crc kubenswrapper[4824]: E1211 02:23:05.441409 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33034069-6622-45c9-9045-44a5e01682bd" containerName="mariadb-database-create" Dec 11 02:23:05 crc kubenswrapper[4824]: I1211 02:23:05.441415 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="33034069-6622-45c9-9045-44a5e01682bd" containerName="mariadb-database-create" Dec 11 02:23:05 crc kubenswrapper[4824]: E1211 02:23:05.441425 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb82f620-fb10-4f76-b2eb-adde800579bf" containerName="mariadb-database-create" Dec 11 02:23:05 crc kubenswrapper[4824]: I1211 02:23:05.441432 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb82f620-fb10-4f76-b2eb-adde800579bf" containerName="mariadb-database-create" Dec 11 02:23:05 crc kubenswrapper[4824]: E1211 02:23:05.441441 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14d1e80f-7c53-4c15-a002-eacde62186bc" containerName="mariadb-account-create-update" Dec 11 02:23:05 crc kubenswrapper[4824]: I1211 02:23:05.441446 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="14d1e80f-7c53-4c15-a002-eacde62186bc" containerName="mariadb-account-create-update" Dec 11 02:23:05 crc kubenswrapper[4824]: E1211 02:23:05.441482 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="295fe4d0-5910-49d6-b8ac-56ff445666f9" containerName="mariadb-database-create" Dec 11 02:23:05 crc kubenswrapper[4824]: I1211 02:23:05.441489 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="295fe4d0-5910-49d6-b8ac-56ff445666f9" containerName="mariadb-database-create" Dec 11 02:23:05 crc kubenswrapper[4824]: E1211 02:23:05.441504 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0fefd9f-c55d-4bb0-8d9c-b3f7c42ac4d4" containerName="mariadb-account-create-update" Dec 11 02:23:05 crc kubenswrapper[4824]: I1211 02:23:05.441510 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0fefd9f-c55d-4bb0-8d9c-b3f7c42ac4d4" containerName="mariadb-account-create-update" Dec 11 02:23:05 crc kubenswrapper[4824]: I1211 02:23:05.441671 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb82f620-fb10-4f76-b2eb-adde800579bf" containerName="mariadb-database-create" Dec 11 02:23:05 crc kubenswrapper[4824]: I1211 02:23:05.441690 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="14d1e80f-7c53-4c15-a002-eacde62186bc" containerName="mariadb-account-create-update" Dec 11 02:23:05 crc kubenswrapper[4824]: I1211 02:23:05.441702 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="295fe4d0-5910-49d6-b8ac-56ff445666f9" containerName="mariadb-database-create" Dec 11 02:23:05 crc kubenswrapper[4824]: I1211 02:23:05.441714 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0fefd9f-c55d-4bb0-8d9c-b3f7c42ac4d4" containerName="mariadb-account-create-update" Dec 11 02:23:05 crc kubenswrapper[4824]: I1211 02:23:05.441729 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="78553f90-b619-4ff6-84b4-05835483a29e" containerName="mariadb-account-create-update" Dec 11 02:23:05 crc kubenswrapper[4824]: I1211 02:23:05.441740 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="33034069-6622-45c9-9045-44a5e01682bd" containerName="mariadb-database-create" Dec 11 02:23:05 crc kubenswrapper[4824]: I1211 02:23:05.442313 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-lx54d" Dec 11 02:23:05 crc kubenswrapper[4824]: I1211 02:23:05.445404 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-pqff8" Dec 11 02:23:05 crc kubenswrapper[4824]: I1211 02:23:05.445807 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Dec 11 02:23:05 crc kubenswrapper[4824]: I1211 02:23:05.447616 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 11 02:23:05 crc kubenswrapper[4824]: I1211 02:23:05.459802 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-lx54d"] Dec 11 02:23:05 crc kubenswrapper[4824]: I1211 02:23:05.594761 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10c51806-9b95-4b93-825f-45cd566834c2-config-data\") pod \"nova-cell0-conductor-db-sync-lx54d\" (UID: \"10c51806-9b95-4b93-825f-45cd566834c2\") " pod="openstack/nova-cell0-conductor-db-sync-lx54d" Dec 11 02:23:05 crc kubenswrapper[4824]: I1211 02:23:05.595106 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-892d5\" (UniqueName: \"kubernetes.io/projected/10c51806-9b95-4b93-825f-45cd566834c2-kube-api-access-892d5\") pod \"nova-cell0-conductor-db-sync-lx54d\" (UID: \"10c51806-9b95-4b93-825f-45cd566834c2\") " pod="openstack/nova-cell0-conductor-db-sync-lx54d" Dec 11 02:23:05 crc kubenswrapper[4824]: I1211 02:23:05.595648 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10c51806-9b95-4b93-825f-45cd566834c2-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-lx54d\" (UID: \"10c51806-9b95-4b93-825f-45cd566834c2\") " pod="openstack/nova-cell0-conductor-db-sync-lx54d" Dec 11 02:23:05 crc kubenswrapper[4824]: I1211 02:23:05.595755 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/10c51806-9b95-4b93-825f-45cd566834c2-scripts\") pod \"nova-cell0-conductor-db-sync-lx54d\" (UID: \"10c51806-9b95-4b93-825f-45cd566834c2\") " pod="openstack/nova-cell0-conductor-db-sync-lx54d" Dec 11 02:23:05 crc kubenswrapper[4824]: I1211 02:23:05.698240 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/10c51806-9b95-4b93-825f-45cd566834c2-scripts\") pod \"nova-cell0-conductor-db-sync-lx54d\" (UID: \"10c51806-9b95-4b93-825f-45cd566834c2\") " pod="openstack/nova-cell0-conductor-db-sync-lx54d" Dec 11 02:23:05 crc kubenswrapper[4824]: I1211 02:23:05.698517 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10c51806-9b95-4b93-825f-45cd566834c2-config-data\") pod \"nova-cell0-conductor-db-sync-lx54d\" (UID: \"10c51806-9b95-4b93-825f-45cd566834c2\") " pod="openstack/nova-cell0-conductor-db-sync-lx54d" Dec 11 02:23:05 crc kubenswrapper[4824]: I1211 02:23:05.698656 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-892d5\" (UniqueName: \"kubernetes.io/projected/10c51806-9b95-4b93-825f-45cd566834c2-kube-api-access-892d5\") pod \"nova-cell0-conductor-db-sync-lx54d\" (UID: \"10c51806-9b95-4b93-825f-45cd566834c2\") " pod="openstack/nova-cell0-conductor-db-sync-lx54d" Dec 11 02:23:05 crc kubenswrapper[4824]: I1211 02:23:05.698862 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10c51806-9b95-4b93-825f-45cd566834c2-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-lx54d\" (UID: \"10c51806-9b95-4b93-825f-45cd566834c2\") " pod="openstack/nova-cell0-conductor-db-sync-lx54d" Dec 11 02:23:05 crc kubenswrapper[4824]: I1211 02:23:05.705159 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/10c51806-9b95-4b93-825f-45cd566834c2-scripts\") pod \"nova-cell0-conductor-db-sync-lx54d\" (UID: \"10c51806-9b95-4b93-825f-45cd566834c2\") " pod="openstack/nova-cell0-conductor-db-sync-lx54d" Dec 11 02:23:05 crc kubenswrapper[4824]: I1211 02:23:05.705203 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10c51806-9b95-4b93-825f-45cd566834c2-config-data\") pod \"nova-cell0-conductor-db-sync-lx54d\" (UID: \"10c51806-9b95-4b93-825f-45cd566834c2\") " pod="openstack/nova-cell0-conductor-db-sync-lx54d" Dec 11 02:23:05 crc kubenswrapper[4824]: I1211 02:23:05.715698 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10c51806-9b95-4b93-825f-45cd566834c2-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-lx54d\" (UID: \"10c51806-9b95-4b93-825f-45cd566834c2\") " pod="openstack/nova-cell0-conductor-db-sync-lx54d" Dec 11 02:23:05 crc kubenswrapper[4824]: I1211 02:23:05.728545 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-892d5\" (UniqueName: \"kubernetes.io/projected/10c51806-9b95-4b93-825f-45cd566834c2-kube-api-access-892d5\") pod \"nova-cell0-conductor-db-sync-lx54d\" (UID: \"10c51806-9b95-4b93-825f-45cd566834c2\") " pod="openstack/nova-cell0-conductor-db-sync-lx54d" Dec 11 02:23:05 crc kubenswrapper[4824]: I1211 02:23:05.759906 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-lx54d" Dec 11 02:23:06 crc kubenswrapper[4824]: I1211 02:23:06.100263 4824 generic.go:334] "Generic (PLEG): container finished" podID="21e824ae-7668-42ad-a039-b7795c13b449" containerID="4f692b5f18b6ad468c6348b5cac9f5a4e880abcd6b639834dcade08d9753af20" exitCode=0 Dec 11 02:23:06 crc kubenswrapper[4824]: I1211 02:23:06.100545 4824 generic.go:334] "Generic (PLEG): container finished" podID="21e824ae-7668-42ad-a039-b7795c13b449" containerID="7cae66953fc8d23c4b105dbcd073ffc20b212d26fde6175d619b470cdaf772f4" exitCode=2 Dec 11 02:23:06 crc kubenswrapper[4824]: I1211 02:23:06.100555 4824 generic.go:334] "Generic (PLEG): container finished" podID="21e824ae-7668-42ad-a039-b7795c13b449" containerID="be84ea003d2e8f4d8a5eb5a0f643a70724703733c9e3b3f7e94e072ac7a05140" exitCode=0 Dec 11 02:23:06 crc kubenswrapper[4824]: I1211 02:23:06.100562 4824 generic.go:334] "Generic (PLEG): container finished" podID="21e824ae-7668-42ad-a039-b7795c13b449" containerID="1de53b66a0c686e91e815f096e88be9dbe58faf50bdda3a0583df3ba0fe5527e" exitCode=0 Dec 11 02:23:06 crc kubenswrapper[4824]: I1211 02:23:06.100613 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"21e824ae-7668-42ad-a039-b7795c13b449","Type":"ContainerDied","Data":"4f692b5f18b6ad468c6348b5cac9f5a4e880abcd6b639834dcade08d9753af20"} Dec 11 02:23:06 crc kubenswrapper[4824]: I1211 02:23:06.100650 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"21e824ae-7668-42ad-a039-b7795c13b449","Type":"ContainerDied","Data":"7cae66953fc8d23c4b105dbcd073ffc20b212d26fde6175d619b470cdaf772f4"} Dec 11 02:23:06 crc kubenswrapper[4824]: I1211 02:23:06.100662 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"21e824ae-7668-42ad-a039-b7795c13b449","Type":"ContainerDied","Data":"be84ea003d2e8f4d8a5eb5a0f643a70724703733c9e3b3f7e94e072ac7a05140"} Dec 11 02:23:06 crc kubenswrapper[4824]: I1211 02:23:06.100671 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"21e824ae-7668-42ad-a039-b7795c13b449","Type":"ContainerDied","Data":"1de53b66a0c686e91e815f096e88be9dbe58faf50bdda3a0583df3ba0fe5527e"} Dec 11 02:23:06 crc kubenswrapper[4824]: I1211 02:23:06.112150 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e6776fbd-a091-4715-9942-4dafac0630e2","Type":"ContainerStarted","Data":"27159fd9609b71fb895745c04c6d9190563a6f576a09f3b1f62fc72e280382a8"} Dec 11 02:23:06 crc kubenswrapper[4824]: I1211 02:23:06.163339 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.163292323 podStartE2EDuration="3.163292323s" podCreationTimestamp="2025-12-11 02:23:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:23:06.156495813 +0000 UTC m=+1327.845533232" watchObservedRunningTime="2025-12-11 02:23:06.163292323 +0000 UTC m=+1327.852329722" Dec 11 02:23:06 crc kubenswrapper[4824]: I1211 02:23:06.261810 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-lx54d"] Dec 11 02:23:06 crc kubenswrapper[4824]: W1211 02:23:06.261830 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod10c51806_9b95_4b93_825f_45cd566834c2.slice/crio-9b2a7797d1b2f50818a7e3bc018c320c9dfd2a8abcc1d7517fc10ce263470904 WatchSource:0}: Error finding container 9b2a7797d1b2f50818a7e3bc018c320c9dfd2a8abcc1d7517fc10ce263470904: Status 404 returned error can't find the container with id 9b2a7797d1b2f50818a7e3bc018c320c9dfd2a8abcc1d7517fc10ce263470904 Dec 11 02:23:06 crc kubenswrapper[4824]: I1211 02:23:06.331879 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 11 02:23:06 crc kubenswrapper[4824]: I1211 02:23:06.414141 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/21e824ae-7668-42ad-a039-b7795c13b449-log-httpd\") pod \"21e824ae-7668-42ad-a039-b7795c13b449\" (UID: \"21e824ae-7668-42ad-a039-b7795c13b449\") " Dec 11 02:23:06 crc kubenswrapper[4824]: I1211 02:23:06.414200 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/21e824ae-7668-42ad-a039-b7795c13b449-sg-core-conf-yaml\") pod \"21e824ae-7668-42ad-a039-b7795c13b449\" (UID: \"21e824ae-7668-42ad-a039-b7795c13b449\") " Dec 11 02:23:06 crc kubenswrapper[4824]: I1211 02:23:06.414278 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4xzbw\" (UniqueName: \"kubernetes.io/projected/21e824ae-7668-42ad-a039-b7795c13b449-kube-api-access-4xzbw\") pod \"21e824ae-7668-42ad-a039-b7795c13b449\" (UID: \"21e824ae-7668-42ad-a039-b7795c13b449\") " Dec 11 02:23:06 crc kubenswrapper[4824]: I1211 02:23:06.414311 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/21e824ae-7668-42ad-a039-b7795c13b449-run-httpd\") pod \"21e824ae-7668-42ad-a039-b7795c13b449\" (UID: \"21e824ae-7668-42ad-a039-b7795c13b449\") " Dec 11 02:23:06 crc kubenswrapper[4824]: I1211 02:23:06.414345 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21e824ae-7668-42ad-a039-b7795c13b449-combined-ca-bundle\") pod \"21e824ae-7668-42ad-a039-b7795c13b449\" (UID: \"21e824ae-7668-42ad-a039-b7795c13b449\") " Dec 11 02:23:06 crc kubenswrapper[4824]: I1211 02:23:06.414812 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/21e824ae-7668-42ad-a039-b7795c13b449-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "21e824ae-7668-42ad-a039-b7795c13b449" (UID: "21e824ae-7668-42ad-a039-b7795c13b449"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:23:06 crc kubenswrapper[4824]: I1211 02:23:06.414867 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/21e824ae-7668-42ad-a039-b7795c13b449-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "21e824ae-7668-42ad-a039-b7795c13b449" (UID: "21e824ae-7668-42ad-a039-b7795c13b449"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:23:06 crc kubenswrapper[4824]: I1211 02:23:06.415196 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/21e824ae-7668-42ad-a039-b7795c13b449-scripts\") pod \"21e824ae-7668-42ad-a039-b7795c13b449\" (UID: \"21e824ae-7668-42ad-a039-b7795c13b449\") " Dec 11 02:23:06 crc kubenswrapper[4824]: I1211 02:23:06.415265 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21e824ae-7668-42ad-a039-b7795c13b449-config-data\") pod \"21e824ae-7668-42ad-a039-b7795c13b449\" (UID: \"21e824ae-7668-42ad-a039-b7795c13b449\") " Dec 11 02:23:06 crc kubenswrapper[4824]: I1211 02:23:06.415656 4824 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/21e824ae-7668-42ad-a039-b7795c13b449-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 11 02:23:06 crc kubenswrapper[4824]: I1211 02:23:06.415674 4824 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/21e824ae-7668-42ad-a039-b7795c13b449-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 11 02:23:06 crc kubenswrapper[4824]: I1211 02:23:06.438245 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21e824ae-7668-42ad-a039-b7795c13b449-kube-api-access-4xzbw" (OuterVolumeSpecName: "kube-api-access-4xzbw") pod "21e824ae-7668-42ad-a039-b7795c13b449" (UID: "21e824ae-7668-42ad-a039-b7795c13b449"). InnerVolumeSpecName "kube-api-access-4xzbw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:23:06 crc kubenswrapper[4824]: I1211 02:23:06.445860 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21e824ae-7668-42ad-a039-b7795c13b449-scripts" (OuterVolumeSpecName: "scripts") pod "21e824ae-7668-42ad-a039-b7795c13b449" (UID: "21e824ae-7668-42ad-a039-b7795c13b449"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:23:06 crc kubenswrapper[4824]: I1211 02:23:06.451496 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21e824ae-7668-42ad-a039-b7795c13b449-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "21e824ae-7668-42ad-a039-b7795c13b449" (UID: "21e824ae-7668-42ad-a039-b7795c13b449"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:23:06 crc kubenswrapper[4824]: I1211 02:23:06.492691 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21e824ae-7668-42ad-a039-b7795c13b449-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "21e824ae-7668-42ad-a039-b7795c13b449" (UID: "21e824ae-7668-42ad-a039-b7795c13b449"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:23:06 crc kubenswrapper[4824]: I1211 02:23:06.517534 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/21e824ae-7668-42ad-a039-b7795c13b449-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 02:23:06 crc kubenswrapper[4824]: I1211 02:23:06.517576 4824 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/21e824ae-7668-42ad-a039-b7795c13b449-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 11 02:23:06 crc kubenswrapper[4824]: I1211 02:23:06.517593 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4xzbw\" (UniqueName: \"kubernetes.io/projected/21e824ae-7668-42ad-a039-b7795c13b449-kube-api-access-4xzbw\") on node \"crc\" DevicePath \"\"" Dec 11 02:23:06 crc kubenswrapper[4824]: I1211 02:23:06.517607 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21e824ae-7668-42ad-a039-b7795c13b449-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 02:23:06 crc kubenswrapper[4824]: I1211 02:23:06.551447 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21e824ae-7668-42ad-a039-b7795c13b449-config-data" (OuterVolumeSpecName: "config-data") pod "21e824ae-7668-42ad-a039-b7795c13b449" (UID: "21e824ae-7668-42ad-a039-b7795c13b449"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:23:06 crc kubenswrapper[4824]: I1211 02:23:06.619144 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21e824ae-7668-42ad-a039-b7795c13b449-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 02:23:07 crc kubenswrapper[4824]: I1211 02:23:07.119883 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-lx54d" event={"ID":"10c51806-9b95-4b93-825f-45cd566834c2","Type":"ContainerStarted","Data":"9b2a7797d1b2f50818a7e3bc018c320c9dfd2a8abcc1d7517fc10ce263470904"} Dec 11 02:23:07 crc kubenswrapper[4824]: I1211 02:23:07.126029 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"21e824ae-7668-42ad-a039-b7795c13b449","Type":"ContainerDied","Data":"daba60c06319fb10332c69c2556d0a520fc7052df83989937b08cb73639c1dc7"} Dec 11 02:23:07 crc kubenswrapper[4824]: I1211 02:23:07.126061 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 11 02:23:07 crc kubenswrapper[4824]: I1211 02:23:07.126102 4824 scope.go:117] "RemoveContainer" containerID="4f692b5f18b6ad468c6348b5cac9f5a4e880abcd6b639834dcade08d9753af20" Dec 11 02:23:07 crc kubenswrapper[4824]: I1211 02:23:07.151807 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 11 02:23:07 crc kubenswrapper[4824]: I1211 02:23:07.154865 4824 scope.go:117] "RemoveContainer" containerID="7cae66953fc8d23c4b105dbcd073ffc20b212d26fde6175d619b470cdaf772f4" Dec 11 02:23:07 crc kubenswrapper[4824]: I1211 02:23:07.163913 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 11 02:23:07 crc kubenswrapper[4824]: I1211 02:23:07.176684 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 11 02:23:07 crc kubenswrapper[4824]: E1211 02:23:07.177072 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21e824ae-7668-42ad-a039-b7795c13b449" containerName="ceilometer-central-agent" Dec 11 02:23:07 crc kubenswrapper[4824]: I1211 02:23:07.177093 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="21e824ae-7668-42ad-a039-b7795c13b449" containerName="ceilometer-central-agent" Dec 11 02:23:07 crc kubenswrapper[4824]: E1211 02:23:07.177116 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21e824ae-7668-42ad-a039-b7795c13b449" containerName="sg-core" Dec 11 02:23:07 crc kubenswrapper[4824]: I1211 02:23:07.177134 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="21e824ae-7668-42ad-a039-b7795c13b449" containerName="sg-core" Dec 11 02:23:07 crc kubenswrapper[4824]: E1211 02:23:07.177148 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21e824ae-7668-42ad-a039-b7795c13b449" containerName="proxy-httpd" Dec 11 02:23:07 crc kubenswrapper[4824]: I1211 02:23:07.177154 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="21e824ae-7668-42ad-a039-b7795c13b449" containerName="proxy-httpd" Dec 11 02:23:07 crc kubenswrapper[4824]: E1211 02:23:07.177175 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21e824ae-7668-42ad-a039-b7795c13b449" containerName="ceilometer-notification-agent" Dec 11 02:23:07 crc kubenswrapper[4824]: I1211 02:23:07.177181 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="21e824ae-7668-42ad-a039-b7795c13b449" containerName="ceilometer-notification-agent" Dec 11 02:23:07 crc kubenswrapper[4824]: I1211 02:23:07.177337 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="21e824ae-7668-42ad-a039-b7795c13b449" containerName="ceilometer-notification-agent" Dec 11 02:23:07 crc kubenswrapper[4824]: I1211 02:23:07.177351 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="21e824ae-7668-42ad-a039-b7795c13b449" containerName="ceilometer-central-agent" Dec 11 02:23:07 crc kubenswrapper[4824]: I1211 02:23:07.177367 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="21e824ae-7668-42ad-a039-b7795c13b449" containerName="sg-core" Dec 11 02:23:07 crc kubenswrapper[4824]: I1211 02:23:07.177380 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="21e824ae-7668-42ad-a039-b7795c13b449" containerName="proxy-httpd" Dec 11 02:23:07 crc kubenswrapper[4824]: I1211 02:23:07.178932 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 11 02:23:07 crc kubenswrapper[4824]: I1211 02:23:07.182413 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 11 02:23:07 crc kubenswrapper[4824]: I1211 02:23:07.182441 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 11 02:23:07 crc kubenswrapper[4824]: I1211 02:23:07.182616 4824 scope.go:117] "RemoveContainer" containerID="be84ea003d2e8f4d8a5eb5a0f643a70724703733c9e3b3f7e94e072ac7a05140" Dec 11 02:23:07 crc kubenswrapper[4824]: I1211 02:23:07.199891 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 11 02:23:07 crc kubenswrapper[4824]: I1211 02:23:07.216248 4824 scope.go:117] "RemoveContainer" containerID="1de53b66a0c686e91e815f096e88be9dbe58faf50bdda3a0583df3ba0fe5527e" Dec 11 02:23:07 crc kubenswrapper[4824]: I1211 02:23:07.229926 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c131ad74-c568-4770-a558-0a35252557e6-scripts\") pod \"ceilometer-0\" (UID: \"c131ad74-c568-4770-a558-0a35252557e6\") " pod="openstack/ceilometer-0" Dec 11 02:23:07 crc kubenswrapper[4824]: I1211 02:23:07.229983 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-687ts\" (UniqueName: \"kubernetes.io/projected/c131ad74-c568-4770-a558-0a35252557e6-kube-api-access-687ts\") pod \"ceilometer-0\" (UID: \"c131ad74-c568-4770-a558-0a35252557e6\") " pod="openstack/ceilometer-0" Dec 11 02:23:07 crc kubenswrapper[4824]: I1211 02:23:07.230043 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c131ad74-c568-4770-a558-0a35252557e6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c131ad74-c568-4770-a558-0a35252557e6\") " pod="openstack/ceilometer-0" Dec 11 02:23:07 crc kubenswrapper[4824]: I1211 02:23:07.230085 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c131ad74-c568-4770-a558-0a35252557e6-log-httpd\") pod \"ceilometer-0\" (UID: \"c131ad74-c568-4770-a558-0a35252557e6\") " pod="openstack/ceilometer-0" Dec 11 02:23:07 crc kubenswrapper[4824]: I1211 02:23:07.230119 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c131ad74-c568-4770-a558-0a35252557e6-config-data\") pod \"ceilometer-0\" (UID: \"c131ad74-c568-4770-a558-0a35252557e6\") " pod="openstack/ceilometer-0" Dec 11 02:23:07 crc kubenswrapper[4824]: I1211 02:23:07.230165 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c131ad74-c568-4770-a558-0a35252557e6-run-httpd\") pod \"ceilometer-0\" (UID: \"c131ad74-c568-4770-a558-0a35252557e6\") " pod="openstack/ceilometer-0" Dec 11 02:23:07 crc kubenswrapper[4824]: I1211 02:23:07.230201 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c131ad74-c568-4770-a558-0a35252557e6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c131ad74-c568-4770-a558-0a35252557e6\") " pod="openstack/ceilometer-0" Dec 11 02:23:07 crc kubenswrapper[4824]: I1211 02:23:07.252452 4824 patch_prober.go:28] interesting pod/machine-config-daemon-gx6xt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 02:23:07 crc kubenswrapper[4824]: I1211 02:23:07.252490 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 02:23:07 crc kubenswrapper[4824]: I1211 02:23:07.331638 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c131ad74-c568-4770-a558-0a35252557e6-log-httpd\") pod \"ceilometer-0\" (UID: \"c131ad74-c568-4770-a558-0a35252557e6\") " pod="openstack/ceilometer-0" Dec 11 02:23:07 crc kubenswrapper[4824]: I1211 02:23:07.331688 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c131ad74-c568-4770-a558-0a35252557e6-config-data\") pod \"ceilometer-0\" (UID: \"c131ad74-c568-4770-a558-0a35252557e6\") " pod="openstack/ceilometer-0" Dec 11 02:23:07 crc kubenswrapper[4824]: I1211 02:23:07.331728 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c131ad74-c568-4770-a558-0a35252557e6-run-httpd\") pod \"ceilometer-0\" (UID: \"c131ad74-c568-4770-a558-0a35252557e6\") " pod="openstack/ceilometer-0" Dec 11 02:23:07 crc kubenswrapper[4824]: I1211 02:23:07.331759 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c131ad74-c568-4770-a558-0a35252557e6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c131ad74-c568-4770-a558-0a35252557e6\") " pod="openstack/ceilometer-0" Dec 11 02:23:07 crc kubenswrapper[4824]: I1211 02:23:07.331808 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c131ad74-c568-4770-a558-0a35252557e6-scripts\") pod \"ceilometer-0\" (UID: \"c131ad74-c568-4770-a558-0a35252557e6\") " pod="openstack/ceilometer-0" Dec 11 02:23:07 crc kubenswrapper[4824]: I1211 02:23:07.331832 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-687ts\" (UniqueName: \"kubernetes.io/projected/c131ad74-c568-4770-a558-0a35252557e6-kube-api-access-687ts\") pod \"ceilometer-0\" (UID: \"c131ad74-c568-4770-a558-0a35252557e6\") " pod="openstack/ceilometer-0" Dec 11 02:23:07 crc kubenswrapper[4824]: I1211 02:23:07.331868 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c131ad74-c568-4770-a558-0a35252557e6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c131ad74-c568-4770-a558-0a35252557e6\") " pod="openstack/ceilometer-0" Dec 11 02:23:07 crc kubenswrapper[4824]: I1211 02:23:07.332521 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c131ad74-c568-4770-a558-0a35252557e6-run-httpd\") pod \"ceilometer-0\" (UID: \"c131ad74-c568-4770-a558-0a35252557e6\") " pod="openstack/ceilometer-0" Dec 11 02:23:07 crc kubenswrapper[4824]: I1211 02:23:07.332597 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c131ad74-c568-4770-a558-0a35252557e6-log-httpd\") pod \"ceilometer-0\" (UID: \"c131ad74-c568-4770-a558-0a35252557e6\") " pod="openstack/ceilometer-0" Dec 11 02:23:07 crc kubenswrapper[4824]: I1211 02:23:07.337921 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c131ad74-c568-4770-a558-0a35252557e6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c131ad74-c568-4770-a558-0a35252557e6\") " pod="openstack/ceilometer-0" Dec 11 02:23:07 crc kubenswrapper[4824]: I1211 02:23:07.337924 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c131ad74-c568-4770-a558-0a35252557e6-scripts\") pod \"ceilometer-0\" (UID: \"c131ad74-c568-4770-a558-0a35252557e6\") " pod="openstack/ceilometer-0" Dec 11 02:23:07 crc kubenswrapper[4824]: I1211 02:23:07.338432 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c131ad74-c568-4770-a558-0a35252557e6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c131ad74-c568-4770-a558-0a35252557e6\") " pod="openstack/ceilometer-0" Dec 11 02:23:07 crc kubenswrapper[4824]: I1211 02:23:07.340211 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c131ad74-c568-4770-a558-0a35252557e6-config-data\") pod \"ceilometer-0\" (UID: \"c131ad74-c568-4770-a558-0a35252557e6\") " pod="openstack/ceilometer-0" Dec 11 02:23:07 crc kubenswrapper[4824]: I1211 02:23:07.370867 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-687ts\" (UniqueName: \"kubernetes.io/projected/c131ad74-c568-4770-a558-0a35252557e6-kube-api-access-687ts\") pod \"ceilometer-0\" (UID: \"c131ad74-c568-4770-a558-0a35252557e6\") " pod="openstack/ceilometer-0" Dec 11 02:23:07 crc kubenswrapper[4824]: I1211 02:23:07.513309 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 11 02:23:08 crc kubenswrapper[4824]: I1211 02:23:08.031513 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 11 02:23:08 crc kubenswrapper[4824]: W1211 02:23:08.042525 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc131ad74_c568_4770_a558_0a35252557e6.slice/crio-7b564d1ee84d7bd8a6b6ed546e913669b6ed1f72c609cbedfd783e5addff9181 WatchSource:0}: Error finding container 7b564d1ee84d7bd8a6b6ed546e913669b6ed1f72c609cbedfd783e5addff9181: Status 404 returned error can't find the container with id 7b564d1ee84d7bd8a6b6ed546e913669b6ed1f72c609cbedfd783e5addff9181 Dec 11 02:23:08 crc kubenswrapper[4824]: I1211 02:23:08.137674 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c131ad74-c568-4770-a558-0a35252557e6","Type":"ContainerStarted","Data":"7b564d1ee84d7bd8a6b6ed546e913669b6ed1f72c609cbedfd783e5addff9181"} Dec 11 02:23:08 crc kubenswrapper[4824]: I1211 02:23:08.641866 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="21e824ae-7668-42ad-a039-b7795c13b449" path="/var/lib/kubelet/pods/21e824ae-7668-42ad-a039-b7795c13b449/volumes" Dec 11 02:23:09 crc kubenswrapper[4824]: I1211 02:23:09.146089 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c131ad74-c568-4770-a558-0a35252557e6","Type":"ContainerStarted","Data":"df55cfd286feefba6723f8bc0e3485489567103fbed49e48a3647d56c8fa47ae"} Dec 11 02:23:12 crc kubenswrapper[4824]: I1211 02:23:12.458328 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 11 02:23:12 crc kubenswrapper[4824]: I1211 02:23:12.458899 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 11 02:23:12 crc kubenswrapper[4824]: I1211 02:23:12.485576 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 11 02:23:12 crc kubenswrapper[4824]: I1211 02:23:12.499737 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 11 02:23:13 crc kubenswrapper[4824]: I1211 02:23:13.190050 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 11 02:23:13 crc kubenswrapper[4824]: I1211 02:23:13.190094 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 11 02:23:13 crc kubenswrapper[4824]: I1211 02:23:13.450830 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 11 02:23:13 crc kubenswrapper[4824]: I1211 02:23:13.450902 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 11 02:23:13 crc kubenswrapper[4824]: I1211 02:23:13.489019 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 11 02:23:13 crc kubenswrapper[4824]: I1211 02:23:13.515460 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 11 02:23:14 crc kubenswrapper[4824]: I1211 02:23:14.198641 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 11 02:23:14 crc kubenswrapper[4824]: I1211 02:23:14.198951 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 11 02:23:15 crc kubenswrapper[4824]: I1211 02:23:15.081552 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 11 02:23:15 crc kubenswrapper[4824]: I1211 02:23:15.127563 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 11 02:23:15 crc kubenswrapper[4824]: I1211 02:23:15.219496 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c131ad74-c568-4770-a558-0a35252557e6","Type":"ContainerStarted","Data":"2a39c8b1f09eeaba55381a4c808bebc5d8b0c4d6cda41187acbb6fdcba4771fc"} Dec 11 02:23:15 crc kubenswrapper[4824]: I1211 02:23:15.238324 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-lx54d" event={"ID":"10c51806-9b95-4b93-825f-45cd566834c2","Type":"ContainerStarted","Data":"eec6b03b5b66e40b5d93beb65d557f9e3afa4db04c99efeb00feb123c145b50b"} Dec 11 02:23:15 crc kubenswrapper[4824]: I1211 02:23:15.270020 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-lx54d" podStartSLOduration=2.043443771 podStartE2EDuration="10.270004262s" podCreationTimestamp="2025-12-11 02:23:05 +0000 UTC" firstStartedPulling="2025-12-11 02:23:06.263648343 +0000 UTC m=+1327.952685722" lastFinishedPulling="2025-12-11 02:23:14.490208834 +0000 UTC m=+1336.179246213" observedRunningTime="2025-12-11 02:23:15.257348335 +0000 UTC m=+1336.946385714" watchObservedRunningTime="2025-12-11 02:23:15.270004262 +0000 UTC m=+1336.959041631" Dec 11 02:23:16 crc kubenswrapper[4824]: I1211 02:23:16.246067 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c131ad74-c568-4770-a558-0a35252557e6","Type":"ContainerStarted","Data":"12e57fca44f30e39e3f058f91352760efc1560be1845b816b2f2743586687407"} Dec 11 02:23:16 crc kubenswrapper[4824]: I1211 02:23:16.246204 4824 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 11 02:23:16 crc kubenswrapper[4824]: I1211 02:23:16.246519 4824 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 11 02:23:16 crc kubenswrapper[4824]: I1211 02:23:16.517310 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 11 02:23:16 crc kubenswrapper[4824]: I1211 02:23:16.519183 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 11 02:23:17 crc kubenswrapper[4824]: I1211 02:23:17.257434 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c131ad74-c568-4770-a558-0a35252557e6","Type":"ContainerStarted","Data":"a0f3f1a45bf60afa01aac315cc2d3c681e0350b8600e067ac080016761e832c2"} Dec 11 02:23:17 crc kubenswrapper[4824]: I1211 02:23:17.274360 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.335877492 podStartE2EDuration="10.274343307s" podCreationTimestamp="2025-12-11 02:23:07 +0000 UTC" firstStartedPulling="2025-12-11 02:23:08.045105101 +0000 UTC m=+1329.734142480" lastFinishedPulling="2025-12-11 02:23:16.983570886 +0000 UTC m=+1338.672608295" observedRunningTime="2025-12-11 02:23:17.273583077 +0000 UTC m=+1338.962620456" watchObservedRunningTime="2025-12-11 02:23:17.274343307 +0000 UTC m=+1338.963380686" Dec 11 02:23:18 crc kubenswrapper[4824]: I1211 02:23:18.264521 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 11 02:23:25 crc kubenswrapper[4824]: I1211 02:23:25.906399 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 11 02:23:25 crc kubenswrapper[4824]: I1211 02:23:25.907287 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c131ad74-c568-4770-a558-0a35252557e6" containerName="proxy-httpd" containerID="cri-o://a0f3f1a45bf60afa01aac315cc2d3c681e0350b8600e067ac080016761e832c2" gracePeriod=30 Dec 11 02:23:25 crc kubenswrapper[4824]: I1211 02:23:25.907295 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c131ad74-c568-4770-a558-0a35252557e6" containerName="sg-core" containerID="cri-o://12e57fca44f30e39e3f058f91352760efc1560be1845b816b2f2743586687407" gracePeriod=30 Dec 11 02:23:25 crc kubenswrapper[4824]: I1211 02:23:25.907489 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c131ad74-c568-4770-a558-0a35252557e6" containerName="ceilometer-central-agent" containerID="cri-o://df55cfd286feefba6723f8bc0e3485489567103fbed49e48a3647d56c8fa47ae" gracePeriod=30 Dec 11 02:23:25 crc kubenswrapper[4824]: I1211 02:23:25.907504 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c131ad74-c568-4770-a558-0a35252557e6" containerName="ceilometer-notification-agent" containerID="cri-o://2a39c8b1f09eeaba55381a4c808bebc5d8b0c4d6cda41187acbb6fdcba4771fc" gracePeriod=30 Dec 11 02:23:26 crc kubenswrapper[4824]: I1211 02:23:26.341816 4824 generic.go:334] "Generic (PLEG): container finished" podID="c131ad74-c568-4770-a558-0a35252557e6" containerID="a0f3f1a45bf60afa01aac315cc2d3c681e0350b8600e067ac080016761e832c2" exitCode=0 Dec 11 02:23:26 crc kubenswrapper[4824]: I1211 02:23:26.342331 4824 generic.go:334] "Generic (PLEG): container finished" podID="c131ad74-c568-4770-a558-0a35252557e6" containerID="12e57fca44f30e39e3f058f91352760efc1560be1845b816b2f2743586687407" exitCode=2 Dec 11 02:23:26 crc kubenswrapper[4824]: I1211 02:23:26.342358 4824 generic.go:334] "Generic (PLEG): container finished" podID="c131ad74-c568-4770-a558-0a35252557e6" containerID="df55cfd286feefba6723f8bc0e3485489567103fbed49e48a3647d56c8fa47ae" exitCode=0 Dec 11 02:23:26 crc kubenswrapper[4824]: I1211 02:23:26.342302 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c131ad74-c568-4770-a558-0a35252557e6","Type":"ContainerDied","Data":"a0f3f1a45bf60afa01aac315cc2d3c681e0350b8600e067ac080016761e832c2"} Dec 11 02:23:26 crc kubenswrapper[4824]: I1211 02:23:26.342763 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c131ad74-c568-4770-a558-0a35252557e6","Type":"ContainerDied","Data":"12e57fca44f30e39e3f058f91352760efc1560be1845b816b2f2743586687407"} Dec 11 02:23:26 crc kubenswrapper[4824]: I1211 02:23:26.342859 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c131ad74-c568-4770-a558-0a35252557e6","Type":"ContainerDied","Data":"df55cfd286feefba6723f8bc0e3485489567103fbed49e48a3647d56c8fa47ae"} Dec 11 02:23:26 crc kubenswrapper[4824]: I1211 02:23:26.344078 4824 generic.go:334] "Generic (PLEG): container finished" podID="10c51806-9b95-4b93-825f-45cd566834c2" containerID="eec6b03b5b66e40b5d93beb65d557f9e3afa4db04c99efeb00feb123c145b50b" exitCode=0 Dec 11 02:23:26 crc kubenswrapper[4824]: I1211 02:23:26.344207 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-lx54d" event={"ID":"10c51806-9b95-4b93-825f-45cd566834c2","Type":"ContainerDied","Data":"eec6b03b5b66e40b5d93beb65d557f9e3afa4db04c99efeb00feb123c145b50b"} Dec 11 02:23:26 crc kubenswrapper[4824]: I1211 02:23:26.811049 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 11 02:23:26 crc kubenswrapper[4824]: I1211 02:23:26.950557 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c131ad74-c568-4770-a558-0a35252557e6-sg-core-conf-yaml\") pod \"c131ad74-c568-4770-a558-0a35252557e6\" (UID: \"c131ad74-c568-4770-a558-0a35252557e6\") " Dec 11 02:23:26 crc kubenswrapper[4824]: I1211 02:23:26.950655 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c131ad74-c568-4770-a558-0a35252557e6-run-httpd\") pod \"c131ad74-c568-4770-a558-0a35252557e6\" (UID: \"c131ad74-c568-4770-a558-0a35252557e6\") " Dec 11 02:23:26 crc kubenswrapper[4824]: I1211 02:23:26.950691 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c131ad74-c568-4770-a558-0a35252557e6-log-httpd\") pod \"c131ad74-c568-4770-a558-0a35252557e6\" (UID: \"c131ad74-c568-4770-a558-0a35252557e6\") " Dec 11 02:23:26 crc kubenswrapper[4824]: I1211 02:23:26.950826 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c131ad74-c568-4770-a558-0a35252557e6-combined-ca-bundle\") pod \"c131ad74-c568-4770-a558-0a35252557e6\" (UID: \"c131ad74-c568-4770-a558-0a35252557e6\") " Dec 11 02:23:26 crc kubenswrapper[4824]: I1211 02:23:26.950867 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c131ad74-c568-4770-a558-0a35252557e6-config-data\") pod \"c131ad74-c568-4770-a558-0a35252557e6\" (UID: \"c131ad74-c568-4770-a558-0a35252557e6\") " Dec 11 02:23:26 crc kubenswrapper[4824]: I1211 02:23:26.950925 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-687ts\" (UniqueName: \"kubernetes.io/projected/c131ad74-c568-4770-a558-0a35252557e6-kube-api-access-687ts\") pod \"c131ad74-c568-4770-a558-0a35252557e6\" (UID: \"c131ad74-c568-4770-a558-0a35252557e6\") " Dec 11 02:23:26 crc kubenswrapper[4824]: I1211 02:23:26.950993 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c131ad74-c568-4770-a558-0a35252557e6-scripts\") pod \"c131ad74-c568-4770-a558-0a35252557e6\" (UID: \"c131ad74-c568-4770-a558-0a35252557e6\") " Dec 11 02:23:26 crc kubenswrapper[4824]: I1211 02:23:26.951546 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c131ad74-c568-4770-a558-0a35252557e6-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "c131ad74-c568-4770-a558-0a35252557e6" (UID: "c131ad74-c568-4770-a558-0a35252557e6"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:23:26 crc kubenswrapper[4824]: I1211 02:23:26.951632 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c131ad74-c568-4770-a558-0a35252557e6-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "c131ad74-c568-4770-a558-0a35252557e6" (UID: "c131ad74-c568-4770-a558-0a35252557e6"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:23:26 crc kubenswrapper[4824]: I1211 02:23:26.956186 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c131ad74-c568-4770-a558-0a35252557e6-kube-api-access-687ts" (OuterVolumeSpecName: "kube-api-access-687ts") pod "c131ad74-c568-4770-a558-0a35252557e6" (UID: "c131ad74-c568-4770-a558-0a35252557e6"). InnerVolumeSpecName "kube-api-access-687ts". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:23:26 crc kubenswrapper[4824]: I1211 02:23:26.957505 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c131ad74-c568-4770-a558-0a35252557e6-scripts" (OuterVolumeSpecName: "scripts") pod "c131ad74-c568-4770-a558-0a35252557e6" (UID: "c131ad74-c568-4770-a558-0a35252557e6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:23:26 crc kubenswrapper[4824]: I1211 02:23:26.998180 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c131ad74-c568-4770-a558-0a35252557e6-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "c131ad74-c568-4770-a558-0a35252557e6" (UID: "c131ad74-c568-4770-a558-0a35252557e6"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:23:27 crc kubenswrapper[4824]: I1211 02:23:27.051726 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c131ad74-c568-4770-a558-0a35252557e6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c131ad74-c568-4770-a558-0a35252557e6" (UID: "c131ad74-c568-4770-a558-0a35252557e6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:23:27 crc kubenswrapper[4824]: I1211 02:23:27.052743 4824 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c131ad74-c568-4770-a558-0a35252557e6-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 11 02:23:27 crc kubenswrapper[4824]: I1211 02:23:27.052768 4824 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c131ad74-c568-4770-a558-0a35252557e6-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 11 02:23:27 crc kubenswrapper[4824]: I1211 02:23:27.052780 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c131ad74-c568-4770-a558-0a35252557e6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 02:23:27 crc kubenswrapper[4824]: I1211 02:23:27.052790 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-687ts\" (UniqueName: \"kubernetes.io/projected/c131ad74-c568-4770-a558-0a35252557e6-kube-api-access-687ts\") on node \"crc\" DevicePath \"\"" Dec 11 02:23:27 crc kubenswrapper[4824]: I1211 02:23:27.052799 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c131ad74-c568-4770-a558-0a35252557e6-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 02:23:27 crc kubenswrapper[4824]: I1211 02:23:27.052807 4824 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c131ad74-c568-4770-a558-0a35252557e6-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 11 02:23:27 crc kubenswrapper[4824]: I1211 02:23:27.092833 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c131ad74-c568-4770-a558-0a35252557e6-config-data" (OuterVolumeSpecName: "config-data") pod "c131ad74-c568-4770-a558-0a35252557e6" (UID: "c131ad74-c568-4770-a558-0a35252557e6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:23:27 crc kubenswrapper[4824]: I1211 02:23:27.154212 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c131ad74-c568-4770-a558-0a35252557e6-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 02:23:27 crc kubenswrapper[4824]: I1211 02:23:27.365362 4824 generic.go:334] "Generic (PLEG): container finished" podID="c131ad74-c568-4770-a558-0a35252557e6" containerID="2a39c8b1f09eeaba55381a4c808bebc5d8b0c4d6cda41187acbb6fdcba4771fc" exitCode=0 Dec 11 02:23:27 crc kubenswrapper[4824]: I1211 02:23:27.368187 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c131ad74-c568-4770-a558-0a35252557e6","Type":"ContainerDied","Data":"2a39c8b1f09eeaba55381a4c808bebc5d8b0c4d6cda41187acbb6fdcba4771fc"} Dec 11 02:23:27 crc kubenswrapper[4824]: I1211 02:23:27.368273 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 11 02:23:27 crc kubenswrapper[4824]: I1211 02:23:27.368311 4824 scope.go:117] "RemoveContainer" containerID="a0f3f1a45bf60afa01aac315cc2d3c681e0350b8600e067ac080016761e832c2" Dec 11 02:23:27 crc kubenswrapper[4824]: I1211 02:23:27.368284 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c131ad74-c568-4770-a558-0a35252557e6","Type":"ContainerDied","Data":"7b564d1ee84d7bd8a6b6ed546e913669b6ed1f72c609cbedfd783e5addff9181"} Dec 11 02:23:27 crc kubenswrapper[4824]: I1211 02:23:27.416560 4824 scope.go:117] "RemoveContainer" containerID="12e57fca44f30e39e3f058f91352760efc1560be1845b816b2f2743586687407" Dec 11 02:23:27 crc kubenswrapper[4824]: I1211 02:23:27.446763 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 11 02:23:27 crc kubenswrapper[4824]: I1211 02:23:27.460392 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 11 02:23:27 crc kubenswrapper[4824]: I1211 02:23:27.462520 4824 scope.go:117] "RemoveContainer" containerID="2a39c8b1f09eeaba55381a4c808bebc5d8b0c4d6cda41187acbb6fdcba4771fc" Dec 11 02:23:27 crc kubenswrapper[4824]: I1211 02:23:27.473758 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 11 02:23:27 crc kubenswrapper[4824]: E1211 02:23:27.474234 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c131ad74-c568-4770-a558-0a35252557e6" containerName="sg-core" Dec 11 02:23:27 crc kubenswrapper[4824]: I1211 02:23:27.474252 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="c131ad74-c568-4770-a558-0a35252557e6" containerName="sg-core" Dec 11 02:23:27 crc kubenswrapper[4824]: E1211 02:23:27.474293 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c131ad74-c568-4770-a558-0a35252557e6" containerName="ceilometer-central-agent" Dec 11 02:23:27 crc kubenswrapper[4824]: I1211 02:23:27.474301 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="c131ad74-c568-4770-a558-0a35252557e6" containerName="ceilometer-central-agent" Dec 11 02:23:27 crc kubenswrapper[4824]: E1211 02:23:27.474309 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c131ad74-c568-4770-a558-0a35252557e6" containerName="proxy-httpd" Dec 11 02:23:27 crc kubenswrapper[4824]: I1211 02:23:27.474315 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="c131ad74-c568-4770-a558-0a35252557e6" containerName="proxy-httpd" Dec 11 02:23:27 crc kubenswrapper[4824]: E1211 02:23:27.474344 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c131ad74-c568-4770-a558-0a35252557e6" containerName="ceilometer-notification-agent" Dec 11 02:23:27 crc kubenswrapper[4824]: I1211 02:23:27.474368 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="c131ad74-c568-4770-a558-0a35252557e6" containerName="ceilometer-notification-agent" Dec 11 02:23:27 crc kubenswrapper[4824]: I1211 02:23:27.474559 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="c131ad74-c568-4770-a558-0a35252557e6" containerName="sg-core" Dec 11 02:23:27 crc kubenswrapper[4824]: I1211 02:23:27.474575 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="c131ad74-c568-4770-a558-0a35252557e6" containerName="proxy-httpd" Dec 11 02:23:27 crc kubenswrapper[4824]: I1211 02:23:27.474609 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="c131ad74-c568-4770-a558-0a35252557e6" containerName="ceilometer-central-agent" Dec 11 02:23:27 crc kubenswrapper[4824]: I1211 02:23:27.474629 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="c131ad74-c568-4770-a558-0a35252557e6" containerName="ceilometer-notification-agent" Dec 11 02:23:27 crc kubenswrapper[4824]: I1211 02:23:27.477724 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 11 02:23:27 crc kubenswrapper[4824]: I1211 02:23:27.482012 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 11 02:23:27 crc kubenswrapper[4824]: I1211 02:23:27.483892 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 11 02:23:27 crc kubenswrapper[4824]: I1211 02:23:27.511855 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 11 02:23:27 crc kubenswrapper[4824]: I1211 02:23:27.519603 4824 scope.go:117] "RemoveContainer" containerID="df55cfd286feefba6723f8bc0e3485489567103fbed49e48a3647d56c8fa47ae" Dec 11 02:23:27 crc kubenswrapper[4824]: I1211 02:23:27.557583 4824 scope.go:117] "RemoveContainer" containerID="a0f3f1a45bf60afa01aac315cc2d3c681e0350b8600e067ac080016761e832c2" Dec 11 02:23:27 crc kubenswrapper[4824]: E1211 02:23:27.558798 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a0f3f1a45bf60afa01aac315cc2d3c681e0350b8600e067ac080016761e832c2\": container with ID starting with a0f3f1a45bf60afa01aac315cc2d3c681e0350b8600e067ac080016761e832c2 not found: ID does not exist" containerID="a0f3f1a45bf60afa01aac315cc2d3c681e0350b8600e067ac080016761e832c2" Dec 11 02:23:27 crc kubenswrapper[4824]: I1211 02:23:27.558869 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0f3f1a45bf60afa01aac315cc2d3c681e0350b8600e067ac080016761e832c2"} err="failed to get container status \"a0f3f1a45bf60afa01aac315cc2d3c681e0350b8600e067ac080016761e832c2\": rpc error: code = NotFound desc = could not find container \"a0f3f1a45bf60afa01aac315cc2d3c681e0350b8600e067ac080016761e832c2\": container with ID starting with a0f3f1a45bf60afa01aac315cc2d3c681e0350b8600e067ac080016761e832c2 not found: ID does not exist" Dec 11 02:23:27 crc kubenswrapper[4824]: I1211 02:23:27.558927 4824 scope.go:117] "RemoveContainer" containerID="12e57fca44f30e39e3f058f91352760efc1560be1845b816b2f2743586687407" Dec 11 02:23:27 crc kubenswrapper[4824]: E1211 02:23:27.559346 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"12e57fca44f30e39e3f058f91352760efc1560be1845b816b2f2743586687407\": container with ID starting with 12e57fca44f30e39e3f058f91352760efc1560be1845b816b2f2743586687407 not found: ID does not exist" containerID="12e57fca44f30e39e3f058f91352760efc1560be1845b816b2f2743586687407" Dec 11 02:23:27 crc kubenswrapper[4824]: I1211 02:23:27.559383 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"12e57fca44f30e39e3f058f91352760efc1560be1845b816b2f2743586687407"} err="failed to get container status \"12e57fca44f30e39e3f058f91352760efc1560be1845b816b2f2743586687407\": rpc error: code = NotFound desc = could not find container \"12e57fca44f30e39e3f058f91352760efc1560be1845b816b2f2743586687407\": container with ID starting with 12e57fca44f30e39e3f058f91352760efc1560be1845b816b2f2743586687407 not found: ID does not exist" Dec 11 02:23:27 crc kubenswrapper[4824]: I1211 02:23:27.559409 4824 scope.go:117] "RemoveContainer" containerID="2a39c8b1f09eeaba55381a4c808bebc5d8b0c4d6cda41187acbb6fdcba4771fc" Dec 11 02:23:27 crc kubenswrapper[4824]: E1211 02:23:27.559752 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a39c8b1f09eeaba55381a4c808bebc5d8b0c4d6cda41187acbb6fdcba4771fc\": container with ID starting with 2a39c8b1f09eeaba55381a4c808bebc5d8b0c4d6cda41187acbb6fdcba4771fc not found: ID does not exist" containerID="2a39c8b1f09eeaba55381a4c808bebc5d8b0c4d6cda41187acbb6fdcba4771fc" Dec 11 02:23:27 crc kubenswrapper[4824]: I1211 02:23:27.559792 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a39c8b1f09eeaba55381a4c808bebc5d8b0c4d6cda41187acbb6fdcba4771fc"} err="failed to get container status \"2a39c8b1f09eeaba55381a4c808bebc5d8b0c4d6cda41187acbb6fdcba4771fc\": rpc error: code = NotFound desc = could not find container \"2a39c8b1f09eeaba55381a4c808bebc5d8b0c4d6cda41187acbb6fdcba4771fc\": container with ID starting with 2a39c8b1f09eeaba55381a4c808bebc5d8b0c4d6cda41187acbb6fdcba4771fc not found: ID does not exist" Dec 11 02:23:27 crc kubenswrapper[4824]: I1211 02:23:27.559839 4824 scope.go:117] "RemoveContainer" containerID="df55cfd286feefba6723f8bc0e3485489567103fbed49e48a3647d56c8fa47ae" Dec 11 02:23:27 crc kubenswrapper[4824]: E1211 02:23:27.560451 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df55cfd286feefba6723f8bc0e3485489567103fbed49e48a3647d56c8fa47ae\": container with ID starting with df55cfd286feefba6723f8bc0e3485489567103fbed49e48a3647d56c8fa47ae not found: ID does not exist" containerID="df55cfd286feefba6723f8bc0e3485489567103fbed49e48a3647d56c8fa47ae" Dec 11 02:23:27 crc kubenswrapper[4824]: I1211 02:23:27.560482 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df55cfd286feefba6723f8bc0e3485489567103fbed49e48a3647d56c8fa47ae"} err="failed to get container status \"df55cfd286feefba6723f8bc0e3485489567103fbed49e48a3647d56c8fa47ae\": rpc error: code = NotFound desc = could not find container \"df55cfd286feefba6723f8bc0e3485489567103fbed49e48a3647d56c8fa47ae\": container with ID starting with df55cfd286feefba6723f8bc0e3485489567103fbed49e48a3647d56c8fa47ae not found: ID does not exist" Dec 11 02:23:27 crc kubenswrapper[4824]: I1211 02:23:27.576013 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2hfrt\" (UniqueName: \"kubernetes.io/projected/c4182e68-340b-43ce-b458-0cac6fa9f7fd-kube-api-access-2hfrt\") pod \"ceilometer-0\" (UID: \"c4182e68-340b-43ce-b458-0cac6fa9f7fd\") " pod="openstack/ceilometer-0" Dec 11 02:23:27 crc kubenswrapper[4824]: I1211 02:23:27.576079 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4182e68-340b-43ce-b458-0cac6fa9f7fd-config-data\") pod \"ceilometer-0\" (UID: \"c4182e68-340b-43ce-b458-0cac6fa9f7fd\") " pod="openstack/ceilometer-0" Dec 11 02:23:27 crc kubenswrapper[4824]: I1211 02:23:27.576229 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c4182e68-340b-43ce-b458-0cac6fa9f7fd-scripts\") pod \"ceilometer-0\" (UID: \"c4182e68-340b-43ce-b458-0cac6fa9f7fd\") " pod="openstack/ceilometer-0" Dec 11 02:23:27 crc kubenswrapper[4824]: I1211 02:23:27.576303 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c4182e68-340b-43ce-b458-0cac6fa9f7fd-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c4182e68-340b-43ce-b458-0cac6fa9f7fd\") " pod="openstack/ceilometer-0" Dec 11 02:23:27 crc kubenswrapper[4824]: I1211 02:23:27.576325 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4182e68-340b-43ce-b458-0cac6fa9f7fd-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c4182e68-340b-43ce-b458-0cac6fa9f7fd\") " pod="openstack/ceilometer-0" Dec 11 02:23:27 crc kubenswrapper[4824]: I1211 02:23:27.576365 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c4182e68-340b-43ce-b458-0cac6fa9f7fd-run-httpd\") pod \"ceilometer-0\" (UID: \"c4182e68-340b-43ce-b458-0cac6fa9f7fd\") " pod="openstack/ceilometer-0" Dec 11 02:23:27 crc kubenswrapper[4824]: I1211 02:23:27.576460 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c4182e68-340b-43ce-b458-0cac6fa9f7fd-log-httpd\") pod \"ceilometer-0\" (UID: \"c4182e68-340b-43ce-b458-0cac6fa9f7fd\") " pod="openstack/ceilometer-0" Dec 11 02:23:27 crc kubenswrapper[4824]: I1211 02:23:27.678873 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c4182e68-340b-43ce-b458-0cac6fa9f7fd-log-httpd\") pod \"ceilometer-0\" (UID: \"c4182e68-340b-43ce-b458-0cac6fa9f7fd\") " pod="openstack/ceilometer-0" Dec 11 02:23:27 crc kubenswrapper[4824]: I1211 02:23:27.679269 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2hfrt\" (UniqueName: \"kubernetes.io/projected/c4182e68-340b-43ce-b458-0cac6fa9f7fd-kube-api-access-2hfrt\") pod \"ceilometer-0\" (UID: \"c4182e68-340b-43ce-b458-0cac6fa9f7fd\") " pod="openstack/ceilometer-0" Dec 11 02:23:27 crc kubenswrapper[4824]: I1211 02:23:27.679304 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4182e68-340b-43ce-b458-0cac6fa9f7fd-config-data\") pod \"ceilometer-0\" (UID: \"c4182e68-340b-43ce-b458-0cac6fa9f7fd\") " pod="openstack/ceilometer-0" Dec 11 02:23:27 crc kubenswrapper[4824]: I1211 02:23:27.679333 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c4182e68-340b-43ce-b458-0cac6fa9f7fd-scripts\") pod \"ceilometer-0\" (UID: \"c4182e68-340b-43ce-b458-0cac6fa9f7fd\") " pod="openstack/ceilometer-0" Dec 11 02:23:27 crc kubenswrapper[4824]: I1211 02:23:27.679370 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c4182e68-340b-43ce-b458-0cac6fa9f7fd-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c4182e68-340b-43ce-b458-0cac6fa9f7fd\") " pod="openstack/ceilometer-0" Dec 11 02:23:27 crc kubenswrapper[4824]: I1211 02:23:27.679394 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4182e68-340b-43ce-b458-0cac6fa9f7fd-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c4182e68-340b-43ce-b458-0cac6fa9f7fd\") " pod="openstack/ceilometer-0" Dec 11 02:23:27 crc kubenswrapper[4824]: I1211 02:23:27.679417 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c4182e68-340b-43ce-b458-0cac6fa9f7fd-run-httpd\") pod \"ceilometer-0\" (UID: \"c4182e68-340b-43ce-b458-0cac6fa9f7fd\") " pod="openstack/ceilometer-0" Dec 11 02:23:27 crc kubenswrapper[4824]: I1211 02:23:27.679905 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c4182e68-340b-43ce-b458-0cac6fa9f7fd-run-httpd\") pod \"ceilometer-0\" (UID: \"c4182e68-340b-43ce-b458-0cac6fa9f7fd\") " pod="openstack/ceilometer-0" Dec 11 02:23:27 crc kubenswrapper[4824]: I1211 02:23:27.680155 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c4182e68-340b-43ce-b458-0cac6fa9f7fd-log-httpd\") pod \"ceilometer-0\" (UID: \"c4182e68-340b-43ce-b458-0cac6fa9f7fd\") " pod="openstack/ceilometer-0" Dec 11 02:23:27 crc kubenswrapper[4824]: I1211 02:23:27.684461 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4182e68-340b-43ce-b458-0cac6fa9f7fd-config-data\") pod \"ceilometer-0\" (UID: \"c4182e68-340b-43ce-b458-0cac6fa9f7fd\") " pod="openstack/ceilometer-0" Dec 11 02:23:27 crc kubenswrapper[4824]: I1211 02:23:27.688153 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4182e68-340b-43ce-b458-0cac6fa9f7fd-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c4182e68-340b-43ce-b458-0cac6fa9f7fd\") " pod="openstack/ceilometer-0" Dec 11 02:23:27 crc kubenswrapper[4824]: I1211 02:23:27.691033 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c4182e68-340b-43ce-b458-0cac6fa9f7fd-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c4182e68-340b-43ce-b458-0cac6fa9f7fd\") " pod="openstack/ceilometer-0" Dec 11 02:23:27 crc kubenswrapper[4824]: I1211 02:23:27.696517 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c4182e68-340b-43ce-b458-0cac6fa9f7fd-scripts\") pod \"ceilometer-0\" (UID: \"c4182e68-340b-43ce-b458-0cac6fa9f7fd\") " pod="openstack/ceilometer-0" Dec 11 02:23:27 crc kubenswrapper[4824]: I1211 02:23:27.699140 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2hfrt\" (UniqueName: \"kubernetes.io/projected/c4182e68-340b-43ce-b458-0cac6fa9f7fd-kube-api-access-2hfrt\") pod \"ceilometer-0\" (UID: \"c4182e68-340b-43ce-b458-0cac6fa9f7fd\") " pod="openstack/ceilometer-0" Dec 11 02:23:27 crc kubenswrapper[4824]: I1211 02:23:27.786233 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-lx54d" Dec 11 02:23:27 crc kubenswrapper[4824]: I1211 02:23:27.798743 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 11 02:23:27 crc kubenswrapper[4824]: I1211 02:23:27.883126 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10c51806-9b95-4b93-825f-45cd566834c2-combined-ca-bundle\") pod \"10c51806-9b95-4b93-825f-45cd566834c2\" (UID: \"10c51806-9b95-4b93-825f-45cd566834c2\") " Dec 11 02:23:27 crc kubenswrapper[4824]: I1211 02:23:27.883195 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10c51806-9b95-4b93-825f-45cd566834c2-config-data\") pod \"10c51806-9b95-4b93-825f-45cd566834c2\" (UID: \"10c51806-9b95-4b93-825f-45cd566834c2\") " Dec 11 02:23:27 crc kubenswrapper[4824]: I1211 02:23:27.883946 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-892d5\" (UniqueName: \"kubernetes.io/projected/10c51806-9b95-4b93-825f-45cd566834c2-kube-api-access-892d5\") pod \"10c51806-9b95-4b93-825f-45cd566834c2\" (UID: \"10c51806-9b95-4b93-825f-45cd566834c2\") " Dec 11 02:23:27 crc kubenswrapper[4824]: I1211 02:23:27.884084 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/10c51806-9b95-4b93-825f-45cd566834c2-scripts\") pod \"10c51806-9b95-4b93-825f-45cd566834c2\" (UID: \"10c51806-9b95-4b93-825f-45cd566834c2\") " Dec 11 02:23:27 crc kubenswrapper[4824]: I1211 02:23:27.889056 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10c51806-9b95-4b93-825f-45cd566834c2-kube-api-access-892d5" (OuterVolumeSpecName: "kube-api-access-892d5") pod "10c51806-9b95-4b93-825f-45cd566834c2" (UID: "10c51806-9b95-4b93-825f-45cd566834c2"). InnerVolumeSpecName "kube-api-access-892d5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:23:27 crc kubenswrapper[4824]: I1211 02:23:27.893546 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10c51806-9b95-4b93-825f-45cd566834c2-scripts" (OuterVolumeSpecName: "scripts") pod "10c51806-9b95-4b93-825f-45cd566834c2" (UID: "10c51806-9b95-4b93-825f-45cd566834c2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:23:27 crc kubenswrapper[4824]: I1211 02:23:27.917647 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10c51806-9b95-4b93-825f-45cd566834c2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "10c51806-9b95-4b93-825f-45cd566834c2" (UID: "10c51806-9b95-4b93-825f-45cd566834c2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:23:27 crc kubenswrapper[4824]: I1211 02:23:27.919330 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10c51806-9b95-4b93-825f-45cd566834c2-config-data" (OuterVolumeSpecName: "config-data") pod "10c51806-9b95-4b93-825f-45cd566834c2" (UID: "10c51806-9b95-4b93-825f-45cd566834c2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:23:27 crc kubenswrapper[4824]: I1211 02:23:27.990001 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/10c51806-9b95-4b93-825f-45cd566834c2-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 02:23:27 crc kubenswrapper[4824]: I1211 02:23:27.990032 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10c51806-9b95-4b93-825f-45cd566834c2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 02:23:27 crc kubenswrapper[4824]: I1211 02:23:27.990044 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10c51806-9b95-4b93-825f-45cd566834c2-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 02:23:27 crc kubenswrapper[4824]: I1211 02:23:27.990053 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-892d5\" (UniqueName: \"kubernetes.io/projected/10c51806-9b95-4b93-825f-45cd566834c2-kube-api-access-892d5\") on node \"crc\" DevicePath \"\"" Dec 11 02:23:28 crc kubenswrapper[4824]: I1211 02:23:28.260398 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 11 02:23:28 crc kubenswrapper[4824]: I1211 02:23:28.377743 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c4182e68-340b-43ce-b458-0cac6fa9f7fd","Type":"ContainerStarted","Data":"5cf0bae8996e79eb913896800a52c69909d05bf2273733a60573cb0c3cac9b46"} Dec 11 02:23:28 crc kubenswrapper[4824]: I1211 02:23:28.379674 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-lx54d" event={"ID":"10c51806-9b95-4b93-825f-45cd566834c2","Type":"ContainerDied","Data":"9b2a7797d1b2f50818a7e3bc018c320c9dfd2a8abcc1d7517fc10ce263470904"} Dec 11 02:23:28 crc kubenswrapper[4824]: I1211 02:23:28.379716 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9b2a7797d1b2f50818a7e3bc018c320c9dfd2a8abcc1d7517fc10ce263470904" Dec 11 02:23:28 crc kubenswrapper[4824]: I1211 02:23:28.379764 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-lx54d" Dec 11 02:23:28 crc kubenswrapper[4824]: I1211 02:23:28.462207 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 11 02:23:28 crc kubenswrapper[4824]: E1211 02:23:28.462552 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10c51806-9b95-4b93-825f-45cd566834c2" containerName="nova-cell0-conductor-db-sync" Dec 11 02:23:28 crc kubenswrapper[4824]: I1211 02:23:28.462568 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="10c51806-9b95-4b93-825f-45cd566834c2" containerName="nova-cell0-conductor-db-sync" Dec 11 02:23:28 crc kubenswrapper[4824]: I1211 02:23:28.462753 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="10c51806-9b95-4b93-825f-45cd566834c2" containerName="nova-cell0-conductor-db-sync" Dec 11 02:23:28 crc kubenswrapper[4824]: I1211 02:23:28.463325 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 11 02:23:28 crc kubenswrapper[4824]: I1211 02:23:28.465789 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-pqff8" Dec 11 02:23:28 crc kubenswrapper[4824]: I1211 02:23:28.465910 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 11 02:23:28 crc kubenswrapper[4824]: I1211 02:23:28.473852 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 11 02:23:28 crc kubenswrapper[4824]: I1211 02:23:28.608103 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-srpdd\" (UniqueName: \"kubernetes.io/projected/1443d344-6dc0-4aab-96fd-522ae53f31cc-kube-api-access-srpdd\") pod \"nova-cell0-conductor-0\" (UID: \"1443d344-6dc0-4aab-96fd-522ae53f31cc\") " pod="openstack/nova-cell0-conductor-0" Dec 11 02:23:28 crc kubenswrapper[4824]: I1211 02:23:28.608209 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1443d344-6dc0-4aab-96fd-522ae53f31cc-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"1443d344-6dc0-4aab-96fd-522ae53f31cc\") " pod="openstack/nova-cell0-conductor-0" Dec 11 02:23:28 crc kubenswrapper[4824]: I1211 02:23:28.608314 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1443d344-6dc0-4aab-96fd-522ae53f31cc-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"1443d344-6dc0-4aab-96fd-522ae53f31cc\") " pod="openstack/nova-cell0-conductor-0" Dec 11 02:23:28 crc kubenswrapper[4824]: I1211 02:23:28.648184 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c131ad74-c568-4770-a558-0a35252557e6" path="/var/lib/kubelet/pods/c131ad74-c568-4770-a558-0a35252557e6/volumes" Dec 11 02:23:28 crc kubenswrapper[4824]: I1211 02:23:28.709650 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1443d344-6dc0-4aab-96fd-522ae53f31cc-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"1443d344-6dc0-4aab-96fd-522ae53f31cc\") " pod="openstack/nova-cell0-conductor-0" Dec 11 02:23:28 crc kubenswrapper[4824]: I1211 02:23:28.709799 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1443d344-6dc0-4aab-96fd-522ae53f31cc-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"1443d344-6dc0-4aab-96fd-522ae53f31cc\") " pod="openstack/nova-cell0-conductor-0" Dec 11 02:23:28 crc kubenswrapper[4824]: I1211 02:23:28.709899 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-srpdd\" (UniqueName: \"kubernetes.io/projected/1443d344-6dc0-4aab-96fd-522ae53f31cc-kube-api-access-srpdd\") pod \"nova-cell0-conductor-0\" (UID: \"1443d344-6dc0-4aab-96fd-522ae53f31cc\") " pod="openstack/nova-cell0-conductor-0" Dec 11 02:23:28 crc kubenswrapper[4824]: I1211 02:23:28.715395 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1443d344-6dc0-4aab-96fd-522ae53f31cc-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"1443d344-6dc0-4aab-96fd-522ae53f31cc\") " pod="openstack/nova-cell0-conductor-0" Dec 11 02:23:28 crc kubenswrapper[4824]: I1211 02:23:28.716647 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1443d344-6dc0-4aab-96fd-522ae53f31cc-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"1443d344-6dc0-4aab-96fd-522ae53f31cc\") " pod="openstack/nova-cell0-conductor-0" Dec 11 02:23:28 crc kubenswrapper[4824]: I1211 02:23:28.733604 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-srpdd\" (UniqueName: \"kubernetes.io/projected/1443d344-6dc0-4aab-96fd-522ae53f31cc-kube-api-access-srpdd\") pod \"nova-cell0-conductor-0\" (UID: \"1443d344-6dc0-4aab-96fd-522ae53f31cc\") " pod="openstack/nova-cell0-conductor-0" Dec 11 02:23:28 crc kubenswrapper[4824]: I1211 02:23:28.778201 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 11 02:23:29 crc kubenswrapper[4824]: I1211 02:23:29.316910 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 11 02:23:29 crc kubenswrapper[4824]: I1211 02:23:29.394343 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c4182e68-340b-43ce-b458-0cac6fa9f7fd","Type":"ContainerStarted","Data":"1f88f5ae1bf5cbbcf93220c49f7f08aa3db2f1f5e7c46ed76b5a5919ea60fd63"} Dec 11 02:23:29 crc kubenswrapper[4824]: I1211 02:23:29.399346 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"1443d344-6dc0-4aab-96fd-522ae53f31cc","Type":"ContainerStarted","Data":"381bd0056773fc031f910b9f7eccc54d4f7426f61ecf0e7cfd5707a099b21fb6"} Dec 11 02:23:30 crc kubenswrapper[4824]: I1211 02:23:30.412257 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c4182e68-340b-43ce-b458-0cac6fa9f7fd","Type":"ContainerStarted","Data":"ae6988e689b7a435f7dff39aa726ebdb43772d5aad1cd52cf4fb6b62c24ae678"} Dec 11 02:23:30 crc kubenswrapper[4824]: I1211 02:23:30.414432 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"1443d344-6dc0-4aab-96fd-522ae53f31cc","Type":"ContainerStarted","Data":"604d52fea92ba3e2a4761ed94ad2c318b89039c218cb402f93ffb367a0e81dfb"} Dec 11 02:23:30 crc kubenswrapper[4824]: I1211 02:23:30.414710 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 11 02:23:30 crc kubenswrapper[4824]: I1211 02:23:30.486482 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.486461983 podStartE2EDuration="2.486461983s" podCreationTimestamp="2025-12-11 02:23:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:23:30.478247776 +0000 UTC m=+1352.167285185" watchObservedRunningTime="2025-12-11 02:23:30.486461983 +0000 UTC m=+1352.175499362" Dec 11 02:23:31 crc kubenswrapper[4824]: I1211 02:23:31.425310 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c4182e68-340b-43ce-b458-0cac6fa9f7fd","Type":"ContainerStarted","Data":"16899e5bc595474bd65445e7d16197f13d56f2ae311171be359b4486c19514f6"} Dec 11 02:23:33 crc kubenswrapper[4824]: I1211 02:23:33.449782 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c4182e68-340b-43ce-b458-0cac6fa9f7fd","Type":"ContainerStarted","Data":"381b142df922b142dcfbebe34bef199920c42f16e2dba9b9cc688f157ba4e937"} Dec 11 02:23:33 crc kubenswrapper[4824]: I1211 02:23:33.450555 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 11 02:23:33 crc kubenswrapper[4824]: I1211 02:23:33.476333 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.032703192 podStartE2EDuration="6.47631005s" podCreationTimestamp="2025-12-11 02:23:27 +0000 UTC" firstStartedPulling="2025-12-11 02:23:28.260314279 +0000 UTC m=+1349.949351658" lastFinishedPulling="2025-12-11 02:23:32.703921147 +0000 UTC m=+1354.392958516" observedRunningTime="2025-12-11 02:23:33.473329026 +0000 UTC m=+1355.162366405" watchObservedRunningTime="2025-12-11 02:23:33.47631005 +0000 UTC m=+1355.165347459" Dec 11 02:23:37 crc kubenswrapper[4824]: I1211 02:23:37.250964 4824 patch_prober.go:28] interesting pod/machine-config-daemon-gx6xt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 02:23:37 crc kubenswrapper[4824]: I1211 02:23:37.251219 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 02:23:37 crc kubenswrapper[4824]: I1211 02:23:37.251252 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" Dec 11 02:23:37 crc kubenswrapper[4824]: I1211 02:23:37.251656 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d40ff17e029de5775ac2507f904eb2259da25d15d3aa3e3a0e4c811c4717820d"} pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 11 02:23:37 crc kubenswrapper[4824]: I1211 02:23:37.251705 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" containerName="machine-config-daemon" containerID="cri-o://d40ff17e029de5775ac2507f904eb2259da25d15d3aa3e3a0e4c811c4717820d" gracePeriod=600 Dec 11 02:23:37 crc kubenswrapper[4824]: I1211 02:23:37.489298 4824 generic.go:334] "Generic (PLEG): container finished" podID="44b156e4-64a4-4d45-aa5e-9b10a862faed" containerID="d40ff17e029de5775ac2507f904eb2259da25d15d3aa3e3a0e4c811c4717820d" exitCode=0 Dec 11 02:23:37 crc kubenswrapper[4824]: I1211 02:23:37.489366 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" event={"ID":"44b156e4-64a4-4d45-aa5e-9b10a862faed","Type":"ContainerDied","Data":"d40ff17e029de5775ac2507f904eb2259da25d15d3aa3e3a0e4c811c4717820d"} Dec 11 02:23:37 crc kubenswrapper[4824]: I1211 02:23:37.489589 4824 scope.go:117] "RemoveContainer" containerID="d1afa0a444a8a2d00d6d09bdd1c078096ee2bfd1a3af1878373a7a899d9f6039" Dec 11 02:23:38 crc kubenswrapper[4824]: I1211 02:23:38.530835 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" event={"ID":"44b156e4-64a4-4d45-aa5e-9b10a862faed","Type":"ContainerStarted","Data":"1d7148bf050b9a450d49eb09a5933cb6d8dd6490c731a4e84f5c37876b3c58a6"} Dec 11 02:23:38 crc kubenswrapper[4824]: I1211 02:23:38.825887 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 11 02:23:39 crc kubenswrapper[4824]: I1211 02:23:39.410998 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-lzc98"] Dec 11 02:23:39 crc kubenswrapper[4824]: I1211 02:23:39.413577 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-lzc98" Dec 11 02:23:39 crc kubenswrapper[4824]: I1211 02:23:39.417343 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Dec 11 02:23:39 crc kubenswrapper[4824]: I1211 02:23:39.417377 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Dec 11 02:23:39 crc kubenswrapper[4824]: I1211 02:23:39.446242 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-lzc98"] Dec 11 02:23:39 crc kubenswrapper[4824]: I1211 02:23:39.576365 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95v4g\" (UniqueName: \"kubernetes.io/projected/f973a712-53f1-4d52-8767-c65f665dcb41-kube-api-access-95v4g\") pod \"nova-cell0-cell-mapping-lzc98\" (UID: \"f973a712-53f1-4d52-8767-c65f665dcb41\") " pod="openstack/nova-cell0-cell-mapping-lzc98" Dec 11 02:23:39 crc kubenswrapper[4824]: I1211 02:23:39.577513 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f973a712-53f1-4d52-8767-c65f665dcb41-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-lzc98\" (UID: \"f973a712-53f1-4d52-8767-c65f665dcb41\") " pod="openstack/nova-cell0-cell-mapping-lzc98" Dec 11 02:23:39 crc kubenswrapper[4824]: I1211 02:23:39.577604 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f973a712-53f1-4d52-8767-c65f665dcb41-config-data\") pod \"nova-cell0-cell-mapping-lzc98\" (UID: \"f973a712-53f1-4d52-8767-c65f665dcb41\") " pod="openstack/nova-cell0-cell-mapping-lzc98" Dec 11 02:23:39 crc kubenswrapper[4824]: I1211 02:23:39.577655 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f973a712-53f1-4d52-8767-c65f665dcb41-scripts\") pod \"nova-cell0-cell-mapping-lzc98\" (UID: \"f973a712-53f1-4d52-8767-c65f665dcb41\") " pod="openstack/nova-cell0-cell-mapping-lzc98" Dec 11 02:23:39 crc kubenswrapper[4824]: I1211 02:23:39.602465 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 11 02:23:39 crc kubenswrapper[4824]: I1211 02:23:39.603648 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 11 02:23:39 crc kubenswrapper[4824]: I1211 02:23:39.605968 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 11 02:23:39 crc kubenswrapper[4824]: I1211 02:23:39.635013 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 11 02:23:39 crc kubenswrapper[4824]: I1211 02:23:39.644150 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 11 02:23:39 crc kubenswrapper[4824]: I1211 02:23:39.645300 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 11 02:23:39 crc kubenswrapper[4824]: I1211 02:23:39.648404 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 11 02:23:39 crc kubenswrapper[4824]: I1211 02:23:39.680607 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f973a712-53f1-4d52-8767-c65f665dcb41-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-lzc98\" (UID: \"f973a712-53f1-4d52-8767-c65f665dcb41\") " pod="openstack/nova-cell0-cell-mapping-lzc98" Dec 11 02:23:39 crc kubenswrapper[4824]: I1211 02:23:39.680673 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f850bc8-5b69-478a-b8d4-1bd17f34b4b2-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"8f850bc8-5b69-478a-b8d4-1bd17f34b4b2\") " pod="openstack/nova-cell1-novncproxy-0" Dec 11 02:23:39 crc kubenswrapper[4824]: I1211 02:23:39.680712 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f850bc8-5b69-478a-b8d4-1bd17f34b4b2-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"8f850bc8-5b69-478a-b8d4-1bd17f34b4b2\") " pod="openstack/nova-cell1-novncproxy-0" Dec 11 02:23:39 crc kubenswrapper[4824]: I1211 02:23:39.680740 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dgrpt\" (UniqueName: \"kubernetes.io/projected/8f850bc8-5b69-478a-b8d4-1bd17f34b4b2-kube-api-access-dgrpt\") pod \"nova-cell1-novncproxy-0\" (UID: \"8f850bc8-5b69-478a-b8d4-1bd17f34b4b2\") " pod="openstack/nova-cell1-novncproxy-0" Dec 11 02:23:39 crc kubenswrapper[4824]: I1211 02:23:39.680790 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f973a712-53f1-4d52-8767-c65f665dcb41-config-data\") pod \"nova-cell0-cell-mapping-lzc98\" (UID: \"f973a712-53f1-4d52-8767-c65f665dcb41\") " pod="openstack/nova-cell0-cell-mapping-lzc98" Dec 11 02:23:39 crc kubenswrapper[4824]: I1211 02:23:39.680831 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f973a712-53f1-4d52-8767-c65f665dcb41-scripts\") pod \"nova-cell0-cell-mapping-lzc98\" (UID: \"f973a712-53f1-4d52-8767-c65f665dcb41\") " pod="openstack/nova-cell0-cell-mapping-lzc98" Dec 11 02:23:39 crc kubenswrapper[4824]: I1211 02:23:39.680863 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95v4g\" (UniqueName: \"kubernetes.io/projected/f973a712-53f1-4d52-8767-c65f665dcb41-kube-api-access-95v4g\") pod \"nova-cell0-cell-mapping-lzc98\" (UID: \"f973a712-53f1-4d52-8767-c65f665dcb41\") " pod="openstack/nova-cell0-cell-mapping-lzc98" Dec 11 02:23:39 crc kubenswrapper[4824]: I1211 02:23:39.689689 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 11 02:23:39 crc kubenswrapper[4824]: I1211 02:23:39.697684 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f973a712-53f1-4d52-8767-c65f665dcb41-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-lzc98\" (UID: \"f973a712-53f1-4d52-8767-c65f665dcb41\") " pod="openstack/nova-cell0-cell-mapping-lzc98" Dec 11 02:23:39 crc kubenswrapper[4824]: I1211 02:23:39.703485 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f973a712-53f1-4d52-8767-c65f665dcb41-scripts\") pod \"nova-cell0-cell-mapping-lzc98\" (UID: \"f973a712-53f1-4d52-8767-c65f665dcb41\") " pod="openstack/nova-cell0-cell-mapping-lzc98" Dec 11 02:23:39 crc kubenswrapper[4824]: I1211 02:23:39.710148 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f973a712-53f1-4d52-8767-c65f665dcb41-config-data\") pod \"nova-cell0-cell-mapping-lzc98\" (UID: \"f973a712-53f1-4d52-8767-c65f665dcb41\") " pod="openstack/nova-cell0-cell-mapping-lzc98" Dec 11 02:23:39 crc kubenswrapper[4824]: I1211 02:23:39.732835 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95v4g\" (UniqueName: \"kubernetes.io/projected/f973a712-53f1-4d52-8767-c65f665dcb41-kube-api-access-95v4g\") pod \"nova-cell0-cell-mapping-lzc98\" (UID: \"f973a712-53f1-4d52-8767-c65f665dcb41\") " pod="openstack/nova-cell0-cell-mapping-lzc98" Dec 11 02:23:39 crc kubenswrapper[4824]: I1211 02:23:39.741339 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-lzc98" Dec 11 02:23:39 crc kubenswrapper[4824]: I1211 02:23:39.741749 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 11 02:23:39 crc kubenswrapper[4824]: I1211 02:23:39.746760 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 11 02:23:39 crc kubenswrapper[4824]: I1211 02:23:39.756380 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 11 02:23:39 crc kubenswrapper[4824]: I1211 02:23:39.764748 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 11 02:23:39 crc kubenswrapper[4824]: I1211 02:23:39.784899 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sdxcl\" (UniqueName: \"kubernetes.io/projected/d3f62926-29c4-4fd2-8e1b-926a53823b65-kube-api-access-sdxcl\") pod \"nova-scheduler-0\" (UID: \"d3f62926-29c4-4fd2-8e1b-926a53823b65\") " pod="openstack/nova-scheduler-0" Dec 11 02:23:39 crc kubenswrapper[4824]: I1211 02:23:39.785068 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f850bc8-5b69-478a-b8d4-1bd17f34b4b2-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"8f850bc8-5b69-478a-b8d4-1bd17f34b4b2\") " pod="openstack/nova-cell1-novncproxy-0" Dec 11 02:23:39 crc kubenswrapper[4824]: I1211 02:23:39.785138 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f850bc8-5b69-478a-b8d4-1bd17f34b4b2-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"8f850bc8-5b69-478a-b8d4-1bd17f34b4b2\") " pod="openstack/nova-cell1-novncproxy-0" Dec 11 02:23:39 crc kubenswrapper[4824]: I1211 02:23:39.785170 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3f62926-29c4-4fd2-8e1b-926a53823b65-config-data\") pod \"nova-scheduler-0\" (UID: \"d3f62926-29c4-4fd2-8e1b-926a53823b65\") " pod="openstack/nova-scheduler-0" Dec 11 02:23:39 crc kubenswrapper[4824]: I1211 02:23:39.785199 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dgrpt\" (UniqueName: \"kubernetes.io/projected/8f850bc8-5b69-478a-b8d4-1bd17f34b4b2-kube-api-access-dgrpt\") pod \"nova-cell1-novncproxy-0\" (UID: \"8f850bc8-5b69-478a-b8d4-1bd17f34b4b2\") " pod="openstack/nova-cell1-novncproxy-0" Dec 11 02:23:39 crc kubenswrapper[4824]: I1211 02:23:39.785227 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3f62926-29c4-4fd2-8e1b-926a53823b65-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d3f62926-29c4-4fd2-8e1b-926a53823b65\") " pod="openstack/nova-scheduler-0" Dec 11 02:23:39 crc kubenswrapper[4824]: I1211 02:23:39.796761 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f850bc8-5b69-478a-b8d4-1bd17f34b4b2-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"8f850bc8-5b69-478a-b8d4-1bd17f34b4b2\") " pod="openstack/nova-cell1-novncproxy-0" Dec 11 02:23:39 crc kubenswrapper[4824]: I1211 02:23:39.796768 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f850bc8-5b69-478a-b8d4-1bd17f34b4b2-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"8f850bc8-5b69-478a-b8d4-1bd17f34b4b2\") " pod="openstack/nova-cell1-novncproxy-0" Dec 11 02:23:39 crc kubenswrapper[4824]: I1211 02:23:39.825391 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 11 02:23:39 crc kubenswrapper[4824]: I1211 02:23:39.827077 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 11 02:23:39 crc kubenswrapper[4824]: I1211 02:23:39.830794 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 11 02:23:39 crc kubenswrapper[4824]: I1211 02:23:39.848101 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dgrpt\" (UniqueName: \"kubernetes.io/projected/8f850bc8-5b69-478a-b8d4-1bd17f34b4b2-kube-api-access-dgrpt\") pod \"nova-cell1-novncproxy-0\" (UID: \"8f850bc8-5b69-478a-b8d4-1bd17f34b4b2\") " pod="openstack/nova-cell1-novncproxy-0" Dec 11 02:23:39 crc kubenswrapper[4824]: I1211 02:23:39.869878 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 11 02:23:39 crc kubenswrapper[4824]: I1211 02:23:39.893255 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3f62926-29c4-4fd2-8e1b-926a53823b65-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d3f62926-29c4-4fd2-8e1b-926a53823b65\") " pod="openstack/nova-scheduler-0" Dec 11 02:23:39 crc kubenswrapper[4824]: I1211 02:23:39.893314 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7af70e6-c4c5-477f-a5af-876490e4ddf7-config-data\") pod \"nova-api-0\" (UID: \"f7af70e6-c4c5-477f-a5af-876490e4ddf7\") " pod="openstack/nova-api-0" Dec 11 02:23:39 crc kubenswrapper[4824]: I1211 02:23:39.893337 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f7af70e6-c4c5-477f-a5af-876490e4ddf7-logs\") pod \"nova-api-0\" (UID: \"f7af70e6-c4c5-477f-a5af-876490e4ddf7\") " pod="openstack/nova-api-0" Dec 11 02:23:39 crc kubenswrapper[4824]: I1211 02:23:39.893393 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sdxcl\" (UniqueName: \"kubernetes.io/projected/d3f62926-29c4-4fd2-8e1b-926a53823b65-kube-api-access-sdxcl\") pod \"nova-scheduler-0\" (UID: \"d3f62926-29c4-4fd2-8e1b-926a53823b65\") " pod="openstack/nova-scheduler-0" Dec 11 02:23:39 crc kubenswrapper[4824]: I1211 02:23:39.893474 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rfm97\" (UniqueName: \"kubernetes.io/projected/f7af70e6-c4c5-477f-a5af-876490e4ddf7-kube-api-access-rfm97\") pod \"nova-api-0\" (UID: \"f7af70e6-c4c5-477f-a5af-876490e4ddf7\") " pod="openstack/nova-api-0" Dec 11 02:23:39 crc kubenswrapper[4824]: I1211 02:23:39.893494 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7af70e6-c4c5-477f-a5af-876490e4ddf7-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f7af70e6-c4c5-477f-a5af-876490e4ddf7\") " pod="openstack/nova-api-0" Dec 11 02:23:39 crc kubenswrapper[4824]: I1211 02:23:39.893519 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3f62926-29c4-4fd2-8e1b-926a53823b65-config-data\") pod \"nova-scheduler-0\" (UID: \"d3f62926-29c4-4fd2-8e1b-926a53823b65\") " pod="openstack/nova-scheduler-0" Dec 11 02:23:39 crc kubenswrapper[4824]: I1211 02:23:39.903648 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3f62926-29c4-4fd2-8e1b-926a53823b65-config-data\") pod \"nova-scheduler-0\" (UID: \"d3f62926-29c4-4fd2-8e1b-926a53823b65\") " pod="openstack/nova-scheduler-0" Dec 11 02:23:39 crc kubenswrapper[4824]: I1211 02:23:39.913352 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3f62926-29c4-4fd2-8e1b-926a53823b65-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d3f62926-29c4-4fd2-8e1b-926a53823b65\") " pod="openstack/nova-scheduler-0" Dec 11 02:23:39 crc kubenswrapper[4824]: I1211 02:23:39.928951 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 11 02:23:39 crc kubenswrapper[4824]: I1211 02:23:39.930011 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sdxcl\" (UniqueName: \"kubernetes.io/projected/d3f62926-29c4-4fd2-8e1b-926a53823b65-kube-api-access-sdxcl\") pod \"nova-scheduler-0\" (UID: \"d3f62926-29c4-4fd2-8e1b-926a53823b65\") " pod="openstack/nova-scheduler-0" Dec 11 02:23:39 crc kubenswrapper[4824]: I1211 02:23:39.951275 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-bjsxw"] Dec 11 02:23:39 crc kubenswrapper[4824]: I1211 02:23:39.954329 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-bjsxw" Dec 11 02:23:39 crc kubenswrapper[4824]: I1211 02:23:39.981881 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-bjsxw"] Dec 11 02:23:39 crc kubenswrapper[4824]: I1211 02:23:39.992834 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 11 02:23:39 crc kubenswrapper[4824]: I1211 02:23:39.995208 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb277b6d-502e-4bcc-801c-0767aac6d4af-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"cb277b6d-502e-4bcc-801c-0767aac6d4af\") " pod="openstack/nova-metadata-0" Dec 11 02:23:39 crc kubenswrapper[4824]: I1211 02:23:39.995237 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb277b6d-502e-4bcc-801c-0767aac6d4af-config-data\") pod \"nova-metadata-0\" (UID: \"cb277b6d-502e-4bcc-801c-0767aac6d4af\") " pod="openstack/nova-metadata-0" Dec 11 02:23:39 crc kubenswrapper[4824]: I1211 02:23:39.995254 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cb277b6d-502e-4bcc-801c-0767aac6d4af-logs\") pod \"nova-metadata-0\" (UID: \"cb277b6d-502e-4bcc-801c-0767aac6d4af\") " pod="openstack/nova-metadata-0" Dec 11 02:23:39 crc kubenswrapper[4824]: I1211 02:23:39.995333 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rfm97\" (UniqueName: \"kubernetes.io/projected/f7af70e6-c4c5-477f-a5af-876490e4ddf7-kube-api-access-rfm97\") pod \"nova-api-0\" (UID: \"f7af70e6-c4c5-477f-a5af-876490e4ddf7\") " pod="openstack/nova-api-0" Dec 11 02:23:39 crc kubenswrapper[4824]: I1211 02:23:39.995361 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7af70e6-c4c5-477f-a5af-876490e4ddf7-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f7af70e6-c4c5-477f-a5af-876490e4ddf7\") " pod="openstack/nova-api-0" Dec 11 02:23:39 crc kubenswrapper[4824]: I1211 02:23:39.995407 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7af70e6-c4c5-477f-a5af-876490e4ddf7-config-data\") pod \"nova-api-0\" (UID: \"f7af70e6-c4c5-477f-a5af-876490e4ddf7\") " pod="openstack/nova-api-0" Dec 11 02:23:39 crc kubenswrapper[4824]: I1211 02:23:39.995492 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f7af70e6-c4c5-477f-a5af-876490e4ddf7-logs\") pod \"nova-api-0\" (UID: \"f7af70e6-c4c5-477f-a5af-876490e4ddf7\") " pod="openstack/nova-api-0" Dec 11 02:23:39 crc kubenswrapper[4824]: I1211 02:23:39.995554 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4lxz\" (UniqueName: \"kubernetes.io/projected/cb277b6d-502e-4bcc-801c-0767aac6d4af-kube-api-access-z4lxz\") pod \"nova-metadata-0\" (UID: \"cb277b6d-502e-4bcc-801c-0767aac6d4af\") " pod="openstack/nova-metadata-0" Dec 11 02:23:39 crc kubenswrapper[4824]: I1211 02:23:39.998671 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f7af70e6-c4c5-477f-a5af-876490e4ddf7-logs\") pod \"nova-api-0\" (UID: \"f7af70e6-c4c5-477f-a5af-876490e4ddf7\") " pod="openstack/nova-api-0" Dec 11 02:23:39 crc kubenswrapper[4824]: I1211 02:23:39.999502 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7af70e6-c4c5-477f-a5af-876490e4ddf7-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f7af70e6-c4c5-477f-a5af-876490e4ddf7\") " pod="openstack/nova-api-0" Dec 11 02:23:40 crc kubenswrapper[4824]: I1211 02:23:40.000918 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7af70e6-c4c5-477f-a5af-876490e4ddf7-config-data\") pod \"nova-api-0\" (UID: \"f7af70e6-c4c5-477f-a5af-876490e4ddf7\") " pod="openstack/nova-api-0" Dec 11 02:23:40 crc kubenswrapper[4824]: I1211 02:23:40.013165 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rfm97\" (UniqueName: \"kubernetes.io/projected/f7af70e6-c4c5-477f-a5af-876490e4ddf7-kube-api-access-rfm97\") pod \"nova-api-0\" (UID: \"f7af70e6-c4c5-477f-a5af-876490e4ddf7\") " pod="openstack/nova-api-0" Dec 11 02:23:40 crc kubenswrapper[4824]: I1211 02:23:40.099945 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aa930a8a-6891-4790-9c51-2165f75270c5-dns-svc\") pod \"dnsmasq-dns-757b4f8459-bjsxw\" (UID: \"aa930a8a-6891-4790-9c51-2165f75270c5\") " pod="openstack/dnsmasq-dns-757b4f8459-bjsxw" Dec 11 02:23:40 crc kubenswrapper[4824]: I1211 02:23:40.100341 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aa930a8a-6891-4790-9c51-2165f75270c5-dns-swift-storage-0\") pod \"dnsmasq-dns-757b4f8459-bjsxw\" (UID: \"aa930a8a-6891-4790-9c51-2165f75270c5\") " pod="openstack/dnsmasq-dns-757b4f8459-bjsxw" Dec 11 02:23:40 crc kubenswrapper[4824]: I1211 02:23:40.100780 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aa930a8a-6891-4790-9c51-2165f75270c5-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4f8459-bjsxw\" (UID: \"aa930a8a-6891-4790-9c51-2165f75270c5\") " pod="openstack/dnsmasq-dns-757b4f8459-bjsxw" Dec 11 02:23:40 crc kubenswrapper[4824]: I1211 02:23:40.100806 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aa930a8a-6891-4790-9c51-2165f75270c5-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4f8459-bjsxw\" (UID: \"aa930a8a-6891-4790-9c51-2165f75270c5\") " pod="openstack/dnsmasq-dns-757b4f8459-bjsxw" Dec 11 02:23:40 crc kubenswrapper[4824]: I1211 02:23:40.100887 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d8tw9\" (UniqueName: \"kubernetes.io/projected/aa930a8a-6891-4790-9c51-2165f75270c5-kube-api-access-d8tw9\") pod \"dnsmasq-dns-757b4f8459-bjsxw\" (UID: \"aa930a8a-6891-4790-9c51-2165f75270c5\") " pod="openstack/dnsmasq-dns-757b4f8459-bjsxw" Dec 11 02:23:40 crc kubenswrapper[4824]: I1211 02:23:40.101175 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4lxz\" (UniqueName: \"kubernetes.io/projected/cb277b6d-502e-4bcc-801c-0767aac6d4af-kube-api-access-z4lxz\") pod \"nova-metadata-0\" (UID: \"cb277b6d-502e-4bcc-801c-0767aac6d4af\") " pod="openstack/nova-metadata-0" Dec 11 02:23:40 crc kubenswrapper[4824]: I1211 02:23:40.101262 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb277b6d-502e-4bcc-801c-0767aac6d4af-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"cb277b6d-502e-4bcc-801c-0767aac6d4af\") " pod="openstack/nova-metadata-0" Dec 11 02:23:40 crc kubenswrapper[4824]: I1211 02:23:40.101291 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb277b6d-502e-4bcc-801c-0767aac6d4af-config-data\") pod \"nova-metadata-0\" (UID: \"cb277b6d-502e-4bcc-801c-0767aac6d4af\") " pod="openstack/nova-metadata-0" Dec 11 02:23:40 crc kubenswrapper[4824]: I1211 02:23:40.101310 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cb277b6d-502e-4bcc-801c-0767aac6d4af-logs\") pod \"nova-metadata-0\" (UID: \"cb277b6d-502e-4bcc-801c-0767aac6d4af\") " pod="openstack/nova-metadata-0" Dec 11 02:23:40 crc kubenswrapper[4824]: I1211 02:23:40.101498 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa930a8a-6891-4790-9c51-2165f75270c5-config\") pod \"dnsmasq-dns-757b4f8459-bjsxw\" (UID: \"aa930a8a-6891-4790-9c51-2165f75270c5\") " pod="openstack/dnsmasq-dns-757b4f8459-bjsxw" Dec 11 02:23:40 crc kubenswrapper[4824]: I1211 02:23:40.101899 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cb277b6d-502e-4bcc-801c-0767aac6d4af-logs\") pod \"nova-metadata-0\" (UID: \"cb277b6d-502e-4bcc-801c-0767aac6d4af\") " pod="openstack/nova-metadata-0" Dec 11 02:23:40 crc kubenswrapper[4824]: I1211 02:23:40.105796 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb277b6d-502e-4bcc-801c-0767aac6d4af-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"cb277b6d-502e-4bcc-801c-0767aac6d4af\") " pod="openstack/nova-metadata-0" Dec 11 02:23:40 crc kubenswrapper[4824]: I1211 02:23:40.119197 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb277b6d-502e-4bcc-801c-0767aac6d4af-config-data\") pod \"nova-metadata-0\" (UID: \"cb277b6d-502e-4bcc-801c-0767aac6d4af\") " pod="openstack/nova-metadata-0" Dec 11 02:23:40 crc kubenswrapper[4824]: I1211 02:23:40.120314 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4lxz\" (UniqueName: \"kubernetes.io/projected/cb277b6d-502e-4bcc-801c-0767aac6d4af-kube-api-access-z4lxz\") pod \"nova-metadata-0\" (UID: \"cb277b6d-502e-4bcc-801c-0767aac6d4af\") " pod="openstack/nova-metadata-0" Dec 11 02:23:40 crc kubenswrapper[4824]: I1211 02:23:40.203055 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa930a8a-6891-4790-9c51-2165f75270c5-config\") pod \"dnsmasq-dns-757b4f8459-bjsxw\" (UID: \"aa930a8a-6891-4790-9c51-2165f75270c5\") " pod="openstack/dnsmasq-dns-757b4f8459-bjsxw" Dec 11 02:23:40 crc kubenswrapper[4824]: I1211 02:23:40.203105 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aa930a8a-6891-4790-9c51-2165f75270c5-dns-svc\") pod \"dnsmasq-dns-757b4f8459-bjsxw\" (UID: \"aa930a8a-6891-4790-9c51-2165f75270c5\") " pod="openstack/dnsmasq-dns-757b4f8459-bjsxw" Dec 11 02:23:40 crc kubenswrapper[4824]: I1211 02:23:40.203166 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aa930a8a-6891-4790-9c51-2165f75270c5-dns-swift-storage-0\") pod \"dnsmasq-dns-757b4f8459-bjsxw\" (UID: \"aa930a8a-6891-4790-9c51-2165f75270c5\") " pod="openstack/dnsmasq-dns-757b4f8459-bjsxw" Dec 11 02:23:40 crc kubenswrapper[4824]: I1211 02:23:40.203194 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aa930a8a-6891-4790-9c51-2165f75270c5-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4f8459-bjsxw\" (UID: \"aa930a8a-6891-4790-9c51-2165f75270c5\") " pod="openstack/dnsmasq-dns-757b4f8459-bjsxw" Dec 11 02:23:40 crc kubenswrapper[4824]: I1211 02:23:40.203210 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aa930a8a-6891-4790-9c51-2165f75270c5-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4f8459-bjsxw\" (UID: \"aa930a8a-6891-4790-9c51-2165f75270c5\") " pod="openstack/dnsmasq-dns-757b4f8459-bjsxw" Dec 11 02:23:40 crc kubenswrapper[4824]: I1211 02:23:40.203253 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d8tw9\" (UniqueName: \"kubernetes.io/projected/aa930a8a-6891-4790-9c51-2165f75270c5-kube-api-access-d8tw9\") pod \"dnsmasq-dns-757b4f8459-bjsxw\" (UID: \"aa930a8a-6891-4790-9c51-2165f75270c5\") " pod="openstack/dnsmasq-dns-757b4f8459-bjsxw" Dec 11 02:23:40 crc kubenswrapper[4824]: I1211 02:23:40.204313 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa930a8a-6891-4790-9c51-2165f75270c5-config\") pod \"dnsmasq-dns-757b4f8459-bjsxw\" (UID: \"aa930a8a-6891-4790-9c51-2165f75270c5\") " pod="openstack/dnsmasq-dns-757b4f8459-bjsxw" Dec 11 02:23:40 crc kubenswrapper[4824]: I1211 02:23:40.204781 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aa930a8a-6891-4790-9c51-2165f75270c5-dns-svc\") pod \"dnsmasq-dns-757b4f8459-bjsxw\" (UID: \"aa930a8a-6891-4790-9c51-2165f75270c5\") " pod="openstack/dnsmasq-dns-757b4f8459-bjsxw" Dec 11 02:23:40 crc kubenswrapper[4824]: I1211 02:23:40.205279 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aa930a8a-6891-4790-9c51-2165f75270c5-dns-swift-storage-0\") pod \"dnsmasq-dns-757b4f8459-bjsxw\" (UID: \"aa930a8a-6891-4790-9c51-2165f75270c5\") " pod="openstack/dnsmasq-dns-757b4f8459-bjsxw" Dec 11 02:23:40 crc kubenswrapper[4824]: I1211 02:23:40.205750 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aa930a8a-6891-4790-9c51-2165f75270c5-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4f8459-bjsxw\" (UID: \"aa930a8a-6891-4790-9c51-2165f75270c5\") " pod="openstack/dnsmasq-dns-757b4f8459-bjsxw" Dec 11 02:23:40 crc kubenswrapper[4824]: I1211 02:23:40.206510 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aa930a8a-6891-4790-9c51-2165f75270c5-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4f8459-bjsxw\" (UID: \"aa930a8a-6891-4790-9c51-2165f75270c5\") " pod="openstack/dnsmasq-dns-757b4f8459-bjsxw" Dec 11 02:23:40 crc kubenswrapper[4824]: I1211 02:23:40.220946 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d8tw9\" (UniqueName: \"kubernetes.io/projected/aa930a8a-6891-4790-9c51-2165f75270c5-kube-api-access-d8tw9\") pod \"dnsmasq-dns-757b4f8459-bjsxw\" (UID: \"aa930a8a-6891-4790-9c51-2165f75270c5\") " pod="openstack/dnsmasq-dns-757b4f8459-bjsxw" Dec 11 02:23:40 crc kubenswrapper[4824]: I1211 02:23:40.274672 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 11 02:23:40 crc kubenswrapper[4824]: I1211 02:23:40.276877 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 11 02:23:40 crc kubenswrapper[4824]: I1211 02:23:40.295670 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-bjsxw" Dec 11 02:23:40 crc kubenswrapper[4824]: I1211 02:23:40.435546 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-lzc98"] Dec 11 02:23:40 crc kubenswrapper[4824]: I1211 02:23:40.453135 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-2k725"] Dec 11 02:23:40 crc kubenswrapper[4824]: I1211 02:23:40.463914 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-2k725" Dec 11 02:23:40 crc kubenswrapper[4824]: I1211 02:23:40.480295 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Dec 11 02:23:40 crc kubenswrapper[4824]: I1211 02:23:40.480634 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 11 02:23:40 crc kubenswrapper[4824]: I1211 02:23:40.490004 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-2k725"] Dec 11 02:23:40 crc kubenswrapper[4824]: I1211 02:23:40.540997 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 11 02:23:40 crc kubenswrapper[4824]: I1211 02:23:40.562702 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 11 02:23:40 crc kubenswrapper[4824]: I1211 02:23:40.570830 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-lzc98" event={"ID":"f973a712-53f1-4d52-8767-c65f665dcb41","Type":"ContainerStarted","Data":"f81a14ed30c75a3d4a846e5b0083aa49375c740fe4ad2072b6a09eb45923e29a"} Dec 11 02:23:40 crc kubenswrapper[4824]: I1211 02:23:40.649284 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a30aa0d2-55dc-4375-b72d-953a3c57043f-config-data\") pod \"nova-cell1-conductor-db-sync-2k725\" (UID: \"a30aa0d2-55dc-4375-b72d-953a3c57043f\") " pod="openstack/nova-cell1-conductor-db-sync-2k725" Dec 11 02:23:40 crc kubenswrapper[4824]: I1211 02:23:40.649321 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ztrqr\" (UniqueName: \"kubernetes.io/projected/a30aa0d2-55dc-4375-b72d-953a3c57043f-kube-api-access-ztrqr\") pod \"nova-cell1-conductor-db-sync-2k725\" (UID: \"a30aa0d2-55dc-4375-b72d-953a3c57043f\") " pod="openstack/nova-cell1-conductor-db-sync-2k725" Dec 11 02:23:40 crc kubenswrapper[4824]: I1211 02:23:40.649350 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a30aa0d2-55dc-4375-b72d-953a3c57043f-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-2k725\" (UID: \"a30aa0d2-55dc-4375-b72d-953a3c57043f\") " pod="openstack/nova-cell1-conductor-db-sync-2k725" Dec 11 02:23:40 crc kubenswrapper[4824]: I1211 02:23:40.649370 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a30aa0d2-55dc-4375-b72d-953a3c57043f-scripts\") pod \"nova-cell1-conductor-db-sync-2k725\" (UID: \"a30aa0d2-55dc-4375-b72d-953a3c57043f\") " pod="openstack/nova-cell1-conductor-db-sync-2k725" Dec 11 02:23:40 crc kubenswrapper[4824]: I1211 02:23:40.751183 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a30aa0d2-55dc-4375-b72d-953a3c57043f-config-data\") pod \"nova-cell1-conductor-db-sync-2k725\" (UID: \"a30aa0d2-55dc-4375-b72d-953a3c57043f\") " pod="openstack/nova-cell1-conductor-db-sync-2k725" Dec 11 02:23:40 crc kubenswrapper[4824]: I1211 02:23:40.751219 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ztrqr\" (UniqueName: \"kubernetes.io/projected/a30aa0d2-55dc-4375-b72d-953a3c57043f-kube-api-access-ztrqr\") pod \"nova-cell1-conductor-db-sync-2k725\" (UID: \"a30aa0d2-55dc-4375-b72d-953a3c57043f\") " pod="openstack/nova-cell1-conductor-db-sync-2k725" Dec 11 02:23:40 crc kubenswrapper[4824]: I1211 02:23:40.751263 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a30aa0d2-55dc-4375-b72d-953a3c57043f-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-2k725\" (UID: \"a30aa0d2-55dc-4375-b72d-953a3c57043f\") " pod="openstack/nova-cell1-conductor-db-sync-2k725" Dec 11 02:23:40 crc kubenswrapper[4824]: I1211 02:23:40.751282 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a30aa0d2-55dc-4375-b72d-953a3c57043f-scripts\") pod \"nova-cell1-conductor-db-sync-2k725\" (UID: \"a30aa0d2-55dc-4375-b72d-953a3c57043f\") " pod="openstack/nova-cell1-conductor-db-sync-2k725" Dec 11 02:23:40 crc kubenswrapper[4824]: I1211 02:23:40.757363 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a30aa0d2-55dc-4375-b72d-953a3c57043f-config-data\") pod \"nova-cell1-conductor-db-sync-2k725\" (UID: \"a30aa0d2-55dc-4375-b72d-953a3c57043f\") " pod="openstack/nova-cell1-conductor-db-sync-2k725" Dec 11 02:23:40 crc kubenswrapper[4824]: I1211 02:23:40.757974 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a30aa0d2-55dc-4375-b72d-953a3c57043f-scripts\") pod \"nova-cell1-conductor-db-sync-2k725\" (UID: \"a30aa0d2-55dc-4375-b72d-953a3c57043f\") " pod="openstack/nova-cell1-conductor-db-sync-2k725" Dec 11 02:23:40 crc kubenswrapper[4824]: I1211 02:23:40.761691 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a30aa0d2-55dc-4375-b72d-953a3c57043f-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-2k725\" (UID: \"a30aa0d2-55dc-4375-b72d-953a3c57043f\") " pod="openstack/nova-cell1-conductor-db-sync-2k725" Dec 11 02:23:40 crc kubenswrapper[4824]: I1211 02:23:40.810904 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ztrqr\" (UniqueName: \"kubernetes.io/projected/a30aa0d2-55dc-4375-b72d-953a3c57043f-kube-api-access-ztrqr\") pod \"nova-cell1-conductor-db-sync-2k725\" (UID: \"a30aa0d2-55dc-4375-b72d-953a3c57043f\") " pod="openstack/nova-cell1-conductor-db-sync-2k725" Dec 11 02:23:40 crc kubenswrapper[4824]: I1211 02:23:40.813535 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-2k725" Dec 11 02:23:41 crc kubenswrapper[4824]: I1211 02:23:41.002369 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 11 02:23:41 crc kubenswrapper[4824]: I1211 02:23:41.079173 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 11 02:23:41 crc kubenswrapper[4824]: I1211 02:23:41.119718 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-bjsxw"] Dec 11 02:23:41 crc kubenswrapper[4824]: I1211 02:23:41.484838 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-2k725"] Dec 11 02:23:41 crc kubenswrapper[4824]: W1211 02:23:41.499654 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda30aa0d2_55dc_4375_b72d_953a3c57043f.slice/crio-03f614b0bff13fef3b7ca08a4af0dfbb5c5cf7c6de1d57ae53f583b51b920184 WatchSource:0}: Error finding container 03f614b0bff13fef3b7ca08a4af0dfbb5c5cf7c6de1d57ae53f583b51b920184: Status 404 returned error can't find the container with id 03f614b0bff13fef3b7ca08a4af0dfbb5c5cf7c6de1d57ae53f583b51b920184 Dec 11 02:23:41 crc kubenswrapper[4824]: I1211 02:23:41.582681 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d3f62926-29c4-4fd2-8e1b-926a53823b65","Type":"ContainerStarted","Data":"f4c38321c7e01a8231ee7c762c1b2e29c9b7f44fac81d8bab4a5cedc47bda76b"} Dec 11 02:23:41 crc kubenswrapper[4824]: I1211 02:23:41.584456 4824 generic.go:334] "Generic (PLEG): container finished" podID="aa930a8a-6891-4790-9c51-2165f75270c5" containerID="044c445997a4ae9b4eb965768ef2443c7b5408a6b12599b224f8d29a366069b2" exitCode=0 Dec 11 02:23:41 crc kubenswrapper[4824]: I1211 02:23:41.584516 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-bjsxw" event={"ID":"aa930a8a-6891-4790-9c51-2165f75270c5","Type":"ContainerDied","Data":"044c445997a4ae9b4eb965768ef2443c7b5408a6b12599b224f8d29a366069b2"} Dec 11 02:23:41 crc kubenswrapper[4824]: I1211 02:23:41.584541 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-bjsxw" event={"ID":"aa930a8a-6891-4790-9c51-2165f75270c5","Type":"ContainerStarted","Data":"108296ee90d9a63c87e2b07b29f3723b29c9ff3a59aa496123c325c0d21bd524"} Dec 11 02:23:41 crc kubenswrapper[4824]: I1211 02:23:41.612317 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f7af70e6-c4c5-477f-a5af-876490e4ddf7","Type":"ContainerStarted","Data":"260830c63558a99219a76a0c71bcae9a5b8cf7a3b8d34cc6e40391f3f2bff41b"} Dec 11 02:23:41 crc kubenswrapper[4824]: I1211 02:23:41.616780 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-2k725" event={"ID":"a30aa0d2-55dc-4375-b72d-953a3c57043f","Type":"ContainerStarted","Data":"03f614b0bff13fef3b7ca08a4af0dfbb5c5cf7c6de1d57ae53f583b51b920184"} Dec 11 02:23:41 crc kubenswrapper[4824]: I1211 02:23:41.618968 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-lzc98" event={"ID":"f973a712-53f1-4d52-8767-c65f665dcb41","Type":"ContainerStarted","Data":"adc63ff1b02af6a4a9cc9409bb1319f46d832f8180fd67da9d5a27531bffdcea"} Dec 11 02:23:41 crc kubenswrapper[4824]: I1211 02:23:41.623832 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"cb277b6d-502e-4bcc-801c-0767aac6d4af","Type":"ContainerStarted","Data":"8d22c28feaf6e212f660907ae75b743cdbf553bb1c335718cdc60a69563ea360"} Dec 11 02:23:41 crc kubenswrapper[4824]: I1211 02:23:41.627774 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"8f850bc8-5b69-478a-b8d4-1bd17f34b4b2","Type":"ContainerStarted","Data":"3ae43ba0ce737bb48adf1d1b9861b7e276443787156939a8d9ef7f8dd9a0ca5c"} Dec 11 02:23:41 crc kubenswrapper[4824]: I1211 02:23:41.647769 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-lzc98" podStartSLOduration=2.6477510669999997 podStartE2EDuration="2.647751067s" podCreationTimestamp="2025-12-11 02:23:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:23:41.63312251 +0000 UTC m=+1363.322159889" watchObservedRunningTime="2025-12-11 02:23:41.647751067 +0000 UTC m=+1363.336788446" Dec 11 02:23:42 crc kubenswrapper[4824]: I1211 02:23:42.658816 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-bjsxw" event={"ID":"aa930a8a-6891-4790-9c51-2165f75270c5","Type":"ContainerStarted","Data":"11952cf39c9a35bfbe8a40322a6d5e89939669557153ae881d9eddb635214b4a"} Dec 11 02:23:42 crc kubenswrapper[4824]: I1211 02:23:42.659401 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-757b4f8459-bjsxw" Dec 11 02:23:42 crc kubenswrapper[4824]: I1211 02:23:42.680454 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-2k725" event={"ID":"a30aa0d2-55dc-4375-b72d-953a3c57043f","Type":"ContainerStarted","Data":"46278df8294fe1a2c45a468015a14a6cd20406e47e9a25c003822d43a2f84332"} Dec 11 02:23:42 crc kubenswrapper[4824]: I1211 02:23:42.684354 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-757b4f8459-bjsxw" podStartSLOduration=3.684337043 podStartE2EDuration="3.684337043s" podCreationTimestamp="2025-12-11 02:23:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:23:42.67781383 +0000 UTC m=+1364.366851219" watchObservedRunningTime="2025-12-11 02:23:42.684337043 +0000 UTC m=+1364.373374422" Dec 11 02:23:43 crc kubenswrapper[4824]: I1211 02:23:43.130377 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-2k725" podStartSLOduration=3.130362152 podStartE2EDuration="3.130362152s" podCreationTimestamp="2025-12-11 02:23:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:23:42.731015475 +0000 UTC m=+1364.420052924" watchObservedRunningTime="2025-12-11 02:23:43.130362152 +0000 UTC m=+1364.819399521" Dec 11 02:23:43 crc kubenswrapper[4824]: I1211 02:23:43.141249 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 11 02:23:43 crc kubenswrapper[4824]: I1211 02:23:43.153392 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 11 02:23:45 crc kubenswrapper[4824]: I1211 02:23:45.721649 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"8f850bc8-5b69-478a-b8d4-1bd17f34b4b2","Type":"ContainerStarted","Data":"f9a6eef8a9b612afb5ddf1eb1a6fbf0a6f1270d7de14586e824ce8b0e4c638df"} Dec 11 02:23:45 crc kubenswrapper[4824]: I1211 02:23:45.722383 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="8f850bc8-5b69-478a-b8d4-1bd17f34b4b2" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://f9a6eef8a9b612afb5ddf1eb1a6fbf0a6f1270d7de14586e824ce8b0e4c638df" gracePeriod=30 Dec 11 02:23:45 crc kubenswrapper[4824]: I1211 02:23:45.727685 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d3f62926-29c4-4fd2-8e1b-926a53823b65","Type":"ContainerStarted","Data":"ae5e31f6a12cd22e1a504146c800e694804f108905cb4e2c880cb196fa831a42"} Dec 11 02:23:45 crc kubenswrapper[4824]: I1211 02:23:45.730058 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f7af70e6-c4c5-477f-a5af-876490e4ddf7","Type":"ContainerStarted","Data":"4ed05f464df1d6511a61ef661dce449fda7495a35fa8f436707fada58aaee475"} Dec 11 02:23:45 crc kubenswrapper[4824]: I1211 02:23:45.730228 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f7af70e6-c4c5-477f-a5af-876490e4ddf7","Type":"ContainerStarted","Data":"1115cb51cefd19b75e6688302803085b6377a4ad2a750518905c401166fd9d32"} Dec 11 02:23:45 crc kubenswrapper[4824]: I1211 02:23:45.732280 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"cb277b6d-502e-4bcc-801c-0767aac6d4af","Type":"ContainerStarted","Data":"7ffc2a2e7a90bc6a7e90cd703324799658f15e3a108226b58bc2330715e80e0a"} Dec 11 02:23:45 crc kubenswrapper[4824]: I1211 02:23:45.732416 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"cb277b6d-502e-4bcc-801c-0767aac6d4af","Type":"ContainerStarted","Data":"1a63c990c98815d08d5309f5eeceb10060e1716d5e2d0bdf46e0f285c8406733"} Dec 11 02:23:45 crc kubenswrapper[4824]: I1211 02:23:45.732484 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="cb277b6d-502e-4bcc-801c-0767aac6d4af" containerName="nova-metadata-log" containerID="cri-o://1a63c990c98815d08d5309f5eeceb10060e1716d5e2d0bdf46e0f285c8406733" gracePeriod=30 Dec 11 02:23:45 crc kubenswrapper[4824]: I1211 02:23:45.732590 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="cb277b6d-502e-4bcc-801c-0767aac6d4af" containerName="nova-metadata-metadata" containerID="cri-o://7ffc2a2e7a90bc6a7e90cd703324799658f15e3a108226b58bc2330715e80e0a" gracePeriod=30 Dec 11 02:23:45 crc kubenswrapper[4824]: I1211 02:23:45.756541 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.841935003 podStartE2EDuration="6.756517639s" podCreationTimestamp="2025-12-11 02:23:39 +0000 UTC" firstStartedPulling="2025-12-11 02:23:40.569419962 +0000 UTC m=+1362.258457341" lastFinishedPulling="2025-12-11 02:23:44.484002588 +0000 UTC m=+1366.173039977" observedRunningTime="2025-12-11 02:23:45.746793554 +0000 UTC m=+1367.435830943" watchObservedRunningTime="2025-12-11 02:23:45.756517639 +0000 UTC m=+1367.445555028" Dec 11 02:23:45 crc kubenswrapper[4824]: I1211 02:23:45.770899 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.864959881 podStartE2EDuration="6.770878739s" podCreationTimestamp="2025-12-11 02:23:39 +0000 UTC" firstStartedPulling="2025-12-11 02:23:40.577839024 +0000 UTC m=+1362.266876393" lastFinishedPulling="2025-12-11 02:23:44.483757832 +0000 UTC m=+1366.172795251" observedRunningTime="2025-12-11 02:23:45.760367015 +0000 UTC m=+1367.449404404" watchObservedRunningTime="2025-12-11 02:23:45.770878739 +0000 UTC m=+1367.459916128" Dec 11 02:23:45 crc kubenswrapper[4824]: I1211 02:23:45.787377 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.302302472 podStartE2EDuration="6.787360463s" podCreationTimestamp="2025-12-11 02:23:39 +0000 UTC" firstStartedPulling="2025-12-11 02:23:41.022349585 +0000 UTC m=+1362.711386964" lastFinishedPulling="2025-12-11 02:23:44.507407536 +0000 UTC m=+1366.196444955" observedRunningTime="2025-12-11 02:23:45.780670445 +0000 UTC m=+1367.469707844" watchObservedRunningTime="2025-12-11 02:23:45.787360463 +0000 UTC m=+1367.476397842" Dec 11 02:23:45 crc kubenswrapper[4824]: I1211 02:23:45.808043 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.448114141 podStartE2EDuration="6.808027522s" podCreationTimestamp="2025-12-11 02:23:39 +0000 UTC" firstStartedPulling="2025-12-11 02:23:41.124394356 +0000 UTC m=+1362.813431735" lastFinishedPulling="2025-12-11 02:23:44.484307727 +0000 UTC m=+1366.173345116" observedRunningTime="2025-12-11 02:23:45.800919994 +0000 UTC m=+1367.489957383" watchObservedRunningTime="2025-12-11 02:23:45.808027522 +0000 UTC m=+1367.497064901" Dec 11 02:23:46 crc kubenswrapper[4824]: I1211 02:23:46.352974 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 11 02:23:46 crc kubenswrapper[4824]: I1211 02:23:46.489528 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb277b6d-502e-4bcc-801c-0767aac6d4af-combined-ca-bundle\") pod \"cb277b6d-502e-4bcc-801c-0767aac6d4af\" (UID: \"cb277b6d-502e-4bcc-801c-0767aac6d4af\") " Dec 11 02:23:46 crc kubenswrapper[4824]: I1211 02:23:46.489594 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cb277b6d-502e-4bcc-801c-0767aac6d4af-logs\") pod \"cb277b6d-502e-4bcc-801c-0767aac6d4af\" (UID: \"cb277b6d-502e-4bcc-801c-0767aac6d4af\") " Dec 11 02:23:46 crc kubenswrapper[4824]: I1211 02:23:46.489675 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z4lxz\" (UniqueName: \"kubernetes.io/projected/cb277b6d-502e-4bcc-801c-0767aac6d4af-kube-api-access-z4lxz\") pod \"cb277b6d-502e-4bcc-801c-0767aac6d4af\" (UID: \"cb277b6d-502e-4bcc-801c-0767aac6d4af\") " Dec 11 02:23:46 crc kubenswrapper[4824]: I1211 02:23:46.489788 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb277b6d-502e-4bcc-801c-0767aac6d4af-config-data\") pod \"cb277b6d-502e-4bcc-801c-0767aac6d4af\" (UID: \"cb277b6d-502e-4bcc-801c-0767aac6d4af\") " Dec 11 02:23:46 crc kubenswrapper[4824]: I1211 02:23:46.490820 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cb277b6d-502e-4bcc-801c-0767aac6d4af-logs" (OuterVolumeSpecName: "logs") pod "cb277b6d-502e-4bcc-801c-0767aac6d4af" (UID: "cb277b6d-502e-4bcc-801c-0767aac6d4af"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:23:46 crc kubenswrapper[4824]: I1211 02:23:46.514902 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb277b6d-502e-4bcc-801c-0767aac6d4af-kube-api-access-z4lxz" (OuterVolumeSpecName: "kube-api-access-z4lxz") pod "cb277b6d-502e-4bcc-801c-0767aac6d4af" (UID: "cb277b6d-502e-4bcc-801c-0767aac6d4af"). InnerVolumeSpecName "kube-api-access-z4lxz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:23:46 crc kubenswrapper[4824]: I1211 02:23:46.538019 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb277b6d-502e-4bcc-801c-0767aac6d4af-config-data" (OuterVolumeSpecName: "config-data") pod "cb277b6d-502e-4bcc-801c-0767aac6d4af" (UID: "cb277b6d-502e-4bcc-801c-0767aac6d4af"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:23:46 crc kubenswrapper[4824]: I1211 02:23:46.564684 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb277b6d-502e-4bcc-801c-0767aac6d4af-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cb277b6d-502e-4bcc-801c-0767aac6d4af" (UID: "cb277b6d-502e-4bcc-801c-0767aac6d4af"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:23:46 crc kubenswrapper[4824]: I1211 02:23:46.591561 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb277b6d-502e-4bcc-801c-0767aac6d4af-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 02:23:46 crc kubenswrapper[4824]: I1211 02:23:46.591591 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cb277b6d-502e-4bcc-801c-0767aac6d4af-logs\") on node \"crc\" DevicePath \"\"" Dec 11 02:23:46 crc kubenswrapper[4824]: I1211 02:23:46.591601 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z4lxz\" (UniqueName: \"kubernetes.io/projected/cb277b6d-502e-4bcc-801c-0767aac6d4af-kube-api-access-z4lxz\") on node \"crc\" DevicePath \"\"" Dec 11 02:23:46 crc kubenswrapper[4824]: I1211 02:23:46.591612 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb277b6d-502e-4bcc-801c-0767aac6d4af-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 02:23:46 crc kubenswrapper[4824]: I1211 02:23:46.741058 4824 generic.go:334] "Generic (PLEG): container finished" podID="cb277b6d-502e-4bcc-801c-0767aac6d4af" containerID="7ffc2a2e7a90bc6a7e90cd703324799658f15e3a108226b58bc2330715e80e0a" exitCode=0 Dec 11 02:23:46 crc kubenswrapper[4824]: I1211 02:23:46.741090 4824 generic.go:334] "Generic (PLEG): container finished" podID="cb277b6d-502e-4bcc-801c-0767aac6d4af" containerID="1a63c990c98815d08d5309f5eeceb10060e1716d5e2d0bdf46e0f285c8406733" exitCode=143 Dec 11 02:23:46 crc kubenswrapper[4824]: I1211 02:23:46.741146 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 11 02:23:46 crc kubenswrapper[4824]: I1211 02:23:46.741189 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"cb277b6d-502e-4bcc-801c-0767aac6d4af","Type":"ContainerDied","Data":"7ffc2a2e7a90bc6a7e90cd703324799658f15e3a108226b58bc2330715e80e0a"} Dec 11 02:23:46 crc kubenswrapper[4824]: I1211 02:23:46.741214 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"cb277b6d-502e-4bcc-801c-0767aac6d4af","Type":"ContainerDied","Data":"1a63c990c98815d08d5309f5eeceb10060e1716d5e2d0bdf46e0f285c8406733"} Dec 11 02:23:46 crc kubenswrapper[4824]: I1211 02:23:46.741223 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"cb277b6d-502e-4bcc-801c-0767aac6d4af","Type":"ContainerDied","Data":"8d22c28feaf6e212f660907ae75b743cdbf553bb1c335718cdc60a69563ea360"} Dec 11 02:23:46 crc kubenswrapper[4824]: I1211 02:23:46.741238 4824 scope.go:117] "RemoveContainer" containerID="7ffc2a2e7a90bc6a7e90cd703324799658f15e3a108226b58bc2330715e80e0a" Dec 11 02:23:46 crc kubenswrapper[4824]: I1211 02:23:46.768482 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 11 02:23:46 crc kubenswrapper[4824]: I1211 02:23:46.771883 4824 scope.go:117] "RemoveContainer" containerID="1a63c990c98815d08d5309f5eeceb10060e1716d5e2d0bdf46e0f285c8406733" Dec 11 02:23:46 crc kubenswrapper[4824]: I1211 02:23:46.791555 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 11 02:23:46 crc kubenswrapper[4824]: I1211 02:23:46.792573 4824 scope.go:117] "RemoveContainer" containerID="7ffc2a2e7a90bc6a7e90cd703324799658f15e3a108226b58bc2330715e80e0a" Dec 11 02:23:46 crc kubenswrapper[4824]: E1211 02:23:46.797449 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ffc2a2e7a90bc6a7e90cd703324799658f15e3a108226b58bc2330715e80e0a\": container with ID starting with 7ffc2a2e7a90bc6a7e90cd703324799658f15e3a108226b58bc2330715e80e0a not found: ID does not exist" containerID="7ffc2a2e7a90bc6a7e90cd703324799658f15e3a108226b58bc2330715e80e0a" Dec 11 02:23:46 crc kubenswrapper[4824]: I1211 02:23:46.797502 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ffc2a2e7a90bc6a7e90cd703324799658f15e3a108226b58bc2330715e80e0a"} err="failed to get container status \"7ffc2a2e7a90bc6a7e90cd703324799658f15e3a108226b58bc2330715e80e0a\": rpc error: code = NotFound desc = could not find container \"7ffc2a2e7a90bc6a7e90cd703324799658f15e3a108226b58bc2330715e80e0a\": container with ID starting with 7ffc2a2e7a90bc6a7e90cd703324799658f15e3a108226b58bc2330715e80e0a not found: ID does not exist" Dec 11 02:23:46 crc kubenswrapper[4824]: I1211 02:23:46.797539 4824 scope.go:117] "RemoveContainer" containerID="1a63c990c98815d08d5309f5eeceb10060e1716d5e2d0bdf46e0f285c8406733" Dec 11 02:23:46 crc kubenswrapper[4824]: E1211 02:23:46.800072 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a63c990c98815d08d5309f5eeceb10060e1716d5e2d0bdf46e0f285c8406733\": container with ID starting with 1a63c990c98815d08d5309f5eeceb10060e1716d5e2d0bdf46e0f285c8406733 not found: ID does not exist" containerID="1a63c990c98815d08d5309f5eeceb10060e1716d5e2d0bdf46e0f285c8406733" Dec 11 02:23:46 crc kubenswrapper[4824]: I1211 02:23:46.800106 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a63c990c98815d08d5309f5eeceb10060e1716d5e2d0bdf46e0f285c8406733"} err="failed to get container status \"1a63c990c98815d08d5309f5eeceb10060e1716d5e2d0bdf46e0f285c8406733\": rpc error: code = NotFound desc = could not find container \"1a63c990c98815d08d5309f5eeceb10060e1716d5e2d0bdf46e0f285c8406733\": container with ID starting with 1a63c990c98815d08d5309f5eeceb10060e1716d5e2d0bdf46e0f285c8406733 not found: ID does not exist" Dec 11 02:23:46 crc kubenswrapper[4824]: I1211 02:23:46.800142 4824 scope.go:117] "RemoveContainer" containerID="7ffc2a2e7a90bc6a7e90cd703324799658f15e3a108226b58bc2330715e80e0a" Dec 11 02:23:46 crc kubenswrapper[4824]: I1211 02:23:46.800511 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ffc2a2e7a90bc6a7e90cd703324799658f15e3a108226b58bc2330715e80e0a"} err="failed to get container status \"7ffc2a2e7a90bc6a7e90cd703324799658f15e3a108226b58bc2330715e80e0a\": rpc error: code = NotFound desc = could not find container \"7ffc2a2e7a90bc6a7e90cd703324799658f15e3a108226b58bc2330715e80e0a\": container with ID starting with 7ffc2a2e7a90bc6a7e90cd703324799658f15e3a108226b58bc2330715e80e0a not found: ID does not exist" Dec 11 02:23:46 crc kubenswrapper[4824]: I1211 02:23:46.800535 4824 scope.go:117] "RemoveContainer" containerID="1a63c990c98815d08d5309f5eeceb10060e1716d5e2d0bdf46e0f285c8406733" Dec 11 02:23:46 crc kubenswrapper[4824]: I1211 02:23:46.800782 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a63c990c98815d08d5309f5eeceb10060e1716d5e2d0bdf46e0f285c8406733"} err="failed to get container status \"1a63c990c98815d08d5309f5eeceb10060e1716d5e2d0bdf46e0f285c8406733\": rpc error: code = NotFound desc = could not find container \"1a63c990c98815d08d5309f5eeceb10060e1716d5e2d0bdf46e0f285c8406733\": container with ID starting with 1a63c990c98815d08d5309f5eeceb10060e1716d5e2d0bdf46e0f285c8406733 not found: ID does not exist" Dec 11 02:23:46 crc kubenswrapper[4824]: I1211 02:23:46.815477 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 11 02:23:46 crc kubenswrapper[4824]: E1211 02:23:46.815932 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb277b6d-502e-4bcc-801c-0767aac6d4af" containerName="nova-metadata-log" Dec 11 02:23:46 crc kubenswrapper[4824]: I1211 02:23:46.815954 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb277b6d-502e-4bcc-801c-0767aac6d4af" containerName="nova-metadata-log" Dec 11 02:23:46 crc kubenswrapper[4824]: E1211 02:23:46.815978 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb277b6d-502e-4bcc-801c-0767aac6d4af" containerName="nova-metadata-metadata" Dec 11 02:23:46 crc kubenswrapper[4824]: I1211 02:23:46.815988 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb277b6d-502e-4bcc-801c-0767aac6d4af" containerName="nova-metadata-metadata" Dec 11 02:23:46 crc kubenswrapper[4824]: I1211 02:23:46.816229 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb277b6d-502e-4bcc-801c-0767aac6d4af" containerName="nova-metadata-log" Dec 11 02:23:46 crc kubenswrapper[4824]: I1211 02:23:46.816254 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb277b6d-502e-4bcc-801c-0767aac6d4af" containerName="nova-metadata-metadata" Dec 11 02:23:46 crc kubenswrapper[4824]: I1211 02:23:46.817213 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 11 02:23:46 crc kubenswrapper[4824]: I1211 02:23:46.819876 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 11 02:23:46 crc kubenswrapper[4824]: I1211 02:23:46.820074 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 11 02:23:46 crc kubenswrapper[4824]: I1211 02:23:46.828631 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 11 02:23:46 crc kubenswrapper[4824]: I1211 02:23:46.998684 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c9b7563-9ee3-4f4f-959c-67cfc4ad0727-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6c9b7563-9ee3-4f4f-959c-67cfc4ad0727\") " pod="openstack/nova-metadata-0" Dec 11 02:23:46 crc kubenswrapper[4824]: I1211 02:23:46.998778 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6c9b7563-9ee3-4f4f-959c-67cfc4ad0727-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"6c9b7563-9ee3-4f4f-959c-67cfc4ad0727\") " pod="openstack/nova-metadata-0" Dec 11 02:23:46 crc kubenswrapper[4824]: I1211 02:23:46.998799 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6c9b7563-9ee3-4f4f-959c-67cfc4ad0727-logs\") pod \"nova-metadata-0\" (UID: \"6c9b7563-9ee3-4f4f-959c-67cfc4ad0727\") " pod="openstack/nova-metadata-0" Dec 11 02:23:46 crc kubenswrapper[4824]: I1211 02:23:46.999556 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xx9cf\" (UniqueName: \"kubernetes.io/projected/6c9b7563-9ee3-4f4f-959c-67cfc4ad0727-kube-api-access-xx9cf\") pod \"nova-metadata-0\" (UID: \"6c9b7563-9ee3-4f4f-959c-67cfc4ad0727\") " pod="openstack/nova-metadata-0" Dec 11 02:23:46 crc kubenswrapper[4824]: I1211 02:23:46.999665 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c9b7563-9ee3-4f4f-959c-67cfc4ad0727-config-data\") pod \"nova-metadata-0\" (UID: \"6c9b7563-9ee3-4f4f-959c-67cfc4ad0727\") " pod="openstack/nova-metadata-0" Dec 11 02:23:47 crc kubenswrapper[4824]: I1211 02:23:47.101051 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c9b7563-9ee3-4f4f-959c-67cfc4ad0727-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6c9b7563-9ee3-4f4f-959c-67cfc4ad0727\") " pod="openstack/nova-metadata-0" Dec 11 02:23:47 crc kubenswrapper[4824]: I1211 02:23:47.101390 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6c9b7563-9ee3-4f4f-959c-67cfc4ad0727-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"6c9b7563-9ee3-4f4f-959c-67cfc4ad0727\") " pod="openstack/nova-metadata-0" Dec 11 02:23:47 crc kubenswrapper[4824]: I1211 02:23:47.101470 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6c9b7563-9ee3-4f4f-959c-67cfc4ad0727-logs\") pod \"nova-metadata-0\" (UID: \"6c9b7563-9ee3-4f4f-959c-67cfc4ad0727\") " pod="openstack/nova-metadata-0" Dec 11 02:23:47 crc kubenswrapper[4824]: I1211 02:23:47.101564 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xx9cf\" (UniqueName: \"kubernetes.io/projected/6c9b7563-9ee3-4f4f-959c-67cfc4ad0727-kube-api-access-xx9cf\") pod \"nova-metadata-0\" (UID: \"6c9b7563-9ee3-4f4f-959c-67cfc4ad0727\") " pod="openstack/nova-metadata-0" Dec 11 02:23:47 crc kubenswrapper[4824]: I1211 02:23:47.101945 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c9b7563-9ee3-4f4f-959c-67cfc4ad0727-config-data\") pod \"nova-metadata-0\" (UID: \"6c9b7563-9ee3-4f4f-959c-67cfc4ad0727\") " pod="openstack/nova-metadata-0" Dec 11 02:23:47 crc kubenswrapper[4824]: I1211 02:23:47.101878 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6c9b7563-9ee3-4f4f-959c-67cfc4ad0727-logs\") pod \"nova-metadata-0\" (UID: \"6c9b7563-9ee3-4f4f-959c-67cfc4ad0727\") " pod="openstack/nova-metadata-0" Dec 11 02:23:47 crc kubenswrapper[4824]: I1211 02:23:47.108241 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c9b7563-9ee3-4f4f-959c-67cfc4ad0727-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6c9b7563-9ee3-4f4f-959c-67cfc4ad0727\") " pod="openstack/nova-metadata-0" Dec 11 02:23:47 crc kubenswrapper[4824]: I1211 02:23:47.109482 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6c9b7563-9ee3-4f4f-959c-67cfc4ad0727-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"6c9b7563-9ee3-4f4f-959c-67cfc4ad0727\") " pod="openstack/nova-metadata-0" Dec 11 02:23:47 crc kubenswrapper[4824]: I1211 02:23:47.113487 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c9b7563-9ee3-4f4f-959c-67cfc4ad0727-config-data\") pod \"nova-metadata-0\" (UID: \"6c9b7563-9ee3-4f4f-959c-67cfc4ad0727\") " pod="openstack/nova-metadata-0" Dec 11 02:23:47 crc kubenswrapper[4824]: I1211 02:23:47.131481 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xx9cf\" (UniqueName: \"kubernetes.io/projected/6c9b7563-9ee3-4f4f-959c-67cfc4ad0727-kube-api-access-xx9cf\") pod \"nova-metadata-0\" (UID: \"6c9b7563-9ee3-4f4f-959c-67cfc4ad0727\") " pod="openstack/nova-metadata-0" Dec 11 02:23:47 crc kubenswrapper[4824]: I1211 02:23:47.141928 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 11 02:23:47 crc kubenswrapper[4824]: W1211 02:23:47.674555 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6c9b7563_9ee3_4f4f_959c_67cfc4ad0727.slice/crio-b7148eae3dff00fb01489eef660e6e679f6e4996bb3ce74d8625145751d2f74d WatchSource:0}: Error finding container b7148eae3dff00fb01489eef660e6e679f6e4996bb3ce74d8625145751d2f74d: Status 404 returned error can't find the container with id b7148eae3dff00fb01489eef660e6e679f6e4996bb3ce74d8625145751d2f74d Dec 11 02:23:47 crc kubenswrapper[4824]: I1211 02:23:47.701994 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 11 02:23:47 crc kubenswrapper[4824]: I1211 02:23:47.784242 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6c9b7563-9ee3-4f4f-959c-67cfc4ad0727","Type":"ContainerStarted","Data":"b7148eae3dff00fb01489eef660e6e679f6e4996bb3ce74d8625145751d2f74d"} Dec 11 02:23:48 crc kubenswrapper[4824]: I1211 02:23:48.652603 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cb277b6d-502e-4bcc-801c-0767aac6d4af" path="/var/lib/kubelet/pods/cb277b6d-502e-4bcc-801c-0767aac6d4af/volumes" Dec 11 02:23:48 crc kubenswrapper[4824]: I1211 02:23:48.816187 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6c9b7563-9ee3-4f4f-959c-67cfc4ad0727","Type":"ContainerStarted","Data":"6c6b5ca9d57945790a51ad54a14928ae599fc4d950be7ce39b497582610f9a11"} Dec 11 02:23:48 crc kubenswrapper[4824]: I1211 02:23:48.816262 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6c9b7563-9ee3-4f4f-959c-67cfc4ad0727","Type":"ContainerStarted","Data":"f2c54559ef781147992a2bc3c1b26eb2d414ab37f5e26ba7aa8fe3c4389eaecc"} Dec 11 02:23:48 crc kubenswrapper[4824]: I1211 02:23:48.822262 4824 generic.go:334] "Generic (PLEG): container finished" podID="f973a712-53f1-4d52-8767-c65f665dcb41" containerID="adc63ff1b02af6a4a9cc9409bb1319f46d832f8180fd67da9d5a27531bffdcea" exitCode=0 Dec 11 02:23:48 crc kubenswrapper[4824]: I1211 02:23:48.822334 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-lzc98" event={"ID":"f973a712-53f1-4d52-8767-c65f665dcb41","Type":"ContainerDied","Data":"adc63ff1b02af6a4a9cc9409bb1319f46d832f8180fd67da9d5a27531bffdcea"} Dec 11 02:23:48 crc kubenswrapper[4824]: I1211 02:23:48.864409 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.86438296 podStartE2EDuration="2.86438296s" podCreationTimestamp="2025-12-11 02:23:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:23:48.847287811 +0000 UTC m=+1370.536325290" watchObservedRunningTime="2025-12-11 02:23:48.86438296 +0000 UTC m=+1370.553420349" Dec 11 02:23:49 crc kubenswrapper[4824]: I1211 02:23:49.836353 4824 generic.go:334] "Generic (PLEG): container finished" podID="a30aa0d2-55dc-4375-b72d-953a3c57043f" containerID="46278df8294fe1a2c45a468015a14a6cd20406e47e9a25c003822d43a2f84332" exitCode=0 Dec 11 02:23:49 crc kubenswrapper[4824]: I1211 02:23:49.836676 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-2k725" event={"ID":"a30aa0d2-55dc-4375-b72d-953a3c57043f","Type":"ContainerDied","Data":"46278df8294fe1a2c45a468015a14a6cd20406e47e9a25c003822d43a2f84332"} Dec 11 02:23:49 crc kubenswrapper[4824]: I1211 02:23:49.930908 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 11 02:23:49 crc kubenswrapper[4824]: I1211 02:23:49.993756 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 11 02:23:49 crc kubenswrapper[4824]: I1211 02:23:49.993804 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 11 02:23:50 crc kubenswrapper[4824]: I1211 02:23:50.036294 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 11 02:23:50 crc kubenswrapper[4824]: I1211 02:23:50.236099 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-lzc98" Dec 11 02:23:50 crc kubenswrapper[4824]: I1211 02:23:50.275453 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 11 02:23:50 crc kubenswrapper[4824]: I1211 02:23:50.275516 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 11 02:23:50 crc kubenswrapper[4824]: I1211 02:23:50.297325 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-757b4f8459-bjsxw" Dec 11 02:23:50 crc kubenswrapper[4824]: I1211 02:23:50.307687 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-95v4g\" (UniqueName: \"kubernetes.io/projected/f973a712-53f1-4d52-8767-c65f665dcb41-kube-api-access-95v4g\") pod \"f973a712-53f1-4d52-8767-c65f665dcb41\" (UID: \"f973a712-53f1-4d52-8767-c65f665dcb41\") " Dec 11 02:23:50 crc kubenswrapper[4824]: I1211 02:23:50.307793 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f973a712-53f1-4d52-8767-c65f665dcb41-combined-ca-bundle\") pod \"f973a712-53f1-4d52-8767-c65f665dcb41\" (UID: \"f973a712-53f1-4d52-8767-c65f665dcb41\") " Dec 11 02:23:50 crc kubenswrapper[4824]: I1211 02:23:50.308681 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f973a712-53f1-4d52-8767-c65f665dcb41-config-data\") pod \"f973a712-53f1-4d52-8767-c65f665dcb41\" (UID: \"f973a712-53f1-4d52-8767-c65f665dcb41\") " Dec 11 02:23:50 crc kubenswrapper[4824]: I1211 02:23:50.308826 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f973a712-53f1-4d52-8767-c65f665dcb41-scripts\") pod \"f973a712-53f1-4d52-8767-c65f665dcb41\" (UID: \"f973a712-53f1-4d52-8767-c65f665dcb41\") " Dec 11 02:23:50 crc kubenswrapper[4824]: I1211 02:23:50.319611 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f973a712-53f1-4d52-8767-c65f665dcb41-kube-api-access-95v4g" (OuterVolumeSpecName: "kube-api-access-95v4g") pod "f973a712-53f1-4d52-8767-c65f665dcb41" (UID: "f973a712-53f1-4d52-8767-c65f665dcb41"). InnerVolumeSpecName "kube-api-access-95v4g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:23:50 crc kubenswrapper[4824]: I1211 02:23:50.327249 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f973a712-53f1-4d52-8767-c65f665dcb41-scripts" (OuterVolumeSpecName: "scripts") pod "f973a712-53f1-4d52-8767-c65f665dcb41" (UID: "f973a712-53f1-4d52-8767-c65f665dcb41"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:23:50 crc kubenswrapper[4824]: I1211 02:23:50.362250 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f973a712-53f1-4d52-8767-c65f665dcb41-config-data" (OuterVolumeSpecName: "config-data") pod "f973a712-53f1-4d52-8767-c65f665dcb41" (UID: "f973a712-53f1-4d52-8767-c65f665dcb41"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:23:50 crc kubenswrapper[4824]: I1211 02:23:50.404365 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-rng95"] Dec 11 02:23:50 crc kubenswrapper[4824]: I1211 02:23:50.404792 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c9776ccc5-rng95" podUID="ae139cc3-a38e-45fa-a3b4-e507ad5a99d6" containerName="dnsmasq-dns" containerID="cri-o://60188aa532b0f0fa9043b6f15ce868702ccd2495595cff67f37a859aa64a0100" gracePeriod=10 Dec 11 02:23:50 crc kubenswrapper[4824]: I1211 02:23:50.411745 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-95v4g\" (UniqueName: \"kubernetes.io/projected/f973a712-53f1-4d52-8767-c65f665dcb41-kube-api-access-95v4g\") on node \"crc\" DevicePath \"\"" Dec 11 02:23:50 crc kubenswrapper[4824]: I1211 02:23:50.411780 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f973a712-53f1-4d52-8767-c65f665dcb41-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 02:23:50 crc kubenswrapper[4824]: I1211 02:23:50.411848 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f973a712-53f1-4d52-8767-c65f665dcb41-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 02:23:50 crc kubenswrapper[4824]: I1211 02:23:50.436583 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f973a712-53f1-4d52-8767-c65f665dcb41-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f973a712-53f1-4d52-8767-c65f665dcb41" (UID: "f973a712-53f1-4d52-8767-c65f665dcb41"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:23:50 crc kubenswrapper[4824]: I1211 02:23:50.513249 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f973a712-53f1-4d52-8767-c65f665dcb41-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 02:23:50 crc kubenswrapper[4824]: I1211 02:23:50.829834 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-rng95" Dec 11 02:23:50 crc kubenswrapper[4824]: I1211 02:23:50.845367 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-lzc98" event={"ID":"f973a712-53f1-4d52-8767-c65f665dcb41","Type":"ContainerDied","Data":"f81a14ed30c75a3d4a846e5b0083aa49375c740fe4ad2072b6a09eb45923e29a"} Dec 11 02:23:50 crc kubenswrapper[4824]: I1211 02:23:50.845411 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-lzc98" Dec 11 02:23:50 crc kubenswrapper[4824]: I1211 02:23:50.845412 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f81a14ed30c75a3d4a846e5b0083aa49375c740fe4ad2072b6a09eb45923e29a" Dec 11 02:23:50 crc kubenswrapper[4824]: I1211 02:23:50.847587 4824 generic.go:334] "Generic (PLEG): container finished" podID="ae139cc3-a38e-45fa-a3b4-e507ad5a99d6" containerID="60188aa532b0f0fa9043b6f15ce868702ccd2495595cff67f37a859aa64a0100" exitCode=0 Dec 11 02:23:50 crc kubenswrapper[4824]: I1211 02:23:50.847629 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-rng95" Dec 11 02:23:50 crc kubenswrapper[4824]: I1211 02:23:50.847676 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-rng95" event={"ID":"ae139cc3-a38e-45fa-a3b4-e507ad5a99d6","Type":"ContainerDied","Data":"60188aa532b0f0fa9043b6f15ce868702ccd2495595cff67f37a859aa64a0100"} Dec 11 02:23:50 crc kubenswrapper[4824]: I1211 02:23:50.847740 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-rng95" event={"ID":"ae139cc3-a38e-45fa-a3b4-e507ad5a99d6","Type":"ContainerDied","Data":"c7f896f81d1071e759f97e85536990b6e57d9b8c2ae4db5ea3a96cc5b4a8e81d"} Dec 11 02:23:50 crc kubenswrapper[4824]: I1211 02:23:50.847763 4824 scope.go:117] "RemoveContainer" containerID="60188aa532b0f0fa9043b6f15ce868702ccd2495595cff67f37a859aa64a0100" Dec 11 02:23:50 crc kubenswrapper[4824]: I1211 02:23:50.875166 4824 scope.go:117] "RemoveContainer" containerID="4f180d0a5f22f977c7b9cda23a893b756fc73677013a093c38a98c141a1fb5d8" Dec 11 02:23:50 crc kubenswrapper[4824]: I1211 02:23:50.889847 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 11 02:23:50 crc kubenswrapper[4824]: I1211 02:23:50.910182 4824 scope.go:117] "RemoveContainer" containerID="60188aa532b0f0fa9043b6f15ce868702ccd2495595cff67f37a859aa64a0100" Dec 11 02:23:50 crc kubenswrapper[4824]: E1211 02:23:50.918022 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"60188aa532b0f0fa9043b6f15ce868702ccd2495595cff67f37a859aa64a0100\": container with ID starting with 60188aa532b0f0fa9043b6f15ce868702ccd2495595cff67f37a859aa64a0100 not found: ID does not exist" containerID="60188aa532b0f0fa9043b6f15ce868702ccd2495595cff67f37a859aa64a0100" Dec 11 02:23:50 crc kubenswrapper[4824]: I1211 02:23:50.918070 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60188aa532b0f0fa9043b6f15ce868702ccd2495595cff67f37a859aa64a0100"} err="failed to get container status \"60188aa532b0f0fa9043b6f15ce868702ccd2495595cff67f37a859aa64a0100\": rpc error: code = NotFound desc = could not find container \"60188aa532b0f0fa9043b6f15ce868702ccd2495595cff67f37a859aa64a0100\": container with ID starting with 60188aa532b0f0fa9043b6f15ce868702ccd2495595cff67f37a859aa64a0100 not found: ID does not exist" Dec 11 02:23:50 crc kubenswrapper[4824]: I1211 02:23:50.918096 4824 scope.go:117] "RemoveContainer" containerID="4f180d0a5f22f977c7b9cda23a893b756fc73677013a093c38a98c141a1fb5d8" Dec 11 02:23:50 crc kubenswrapper[4824]: I1211 02:23:50.918783 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ae139cc3-a38e-45fa-a3b4-e507ad5a99d6-dns-swift-storage-0\") pod \"ae139cc3-a38e-45fa-a3b4-e507ad5a99d6\" (UID: \"ae139cc3-a38e-45fa-a3b4-e507ad5a99d6\") " Dec 11 02:23:50 crc kubenswrapper[4824]: I1211 02:23:50.918810 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ae139cc3-a38e-45fa-a3b4-e507ad5a99d6-ovsdbserver-sb\") pod \"ae139cc3-a38e-45fa-a3b4-e507ad5a99d6\" (UID: \"ae139cc3-a38e-45fa-a3b4-e507ad5a99d6\") " Dec 11 02:23:50 crc kubenswrapper[4824]: I1211 02:23:50.918857 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jh6td\" (UniqueName: \"kubernetes.io/projected/ae139cc3-a38e-45fa-a3b4-e507ad5a99d6-kube-api-access-jh6td\") pod \"ae139cc3-a38e-45fa-a3b4-e507ad5a99d6\" (UID: \"ae139cc3-a38e-45fa-a3b4-e507ad5a99d6\") " Dec 11 02:23:50 crc kubenswrapper[4824]: I1211 02:23:50.918905 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ae139cc3-a38e-45fa-a3b4-e507ad5a99d6-dns-svc\") pod \"ae139cc3-a38e-45fa-a3b4-e507ad5a99d6\" (UID: \"ae139cc3-a38e-45fa-a3b4-e507ad5a99d6\") " Dec 11 02:23:50 crc kubenswrapper[4824]: I1211 02:23:50.918945 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ae139cc3-a38e-45fa-a3b4-e507ad5a99d6-ovsdbserver-nb\") pod \"ae139cc3-a38e-45fa-a3b4-e507ad5a99d6\" (UID: \"ae139cc3-a38e-45fa-a3b4-e507ad5a99d6\") " Dec 11 02:23:50 crc kubenswrapper[4824]: I1211 02:23:50.919005 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae139cc3-a38e-45fa-a3b4-e507ad5a99d6-config\") pod \"ae139cc3-a38e-45fa-a3b4-e507ad5a99d6\" (UID: \"ae139cc3-a38e-45fa-a3b4-e507ad5a99d6\") " Dec 11 02:23:50 crc kubenswrapper[4824]: E1211 02:23:50.921252 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f180d0a5f22f977c7b9cda23a893b756fc73677013a093c38a98c141a1fb5d8\": container with ID starting with 4f180d0a5f22f977c7b9cda23a893b756fc73677013a093c38a98c141a1fb5d8 not found: ID does not exist" containerID="4f180d0a5f22f977c7b9cda23a893b756fc73677013a093c38a98c141a1fb5d8" Dec 11 02:23:50 crc kubenswrapper[4824]: I1211 02:23:50.921293 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f180d0a5f22f977c7b9cda23a893b756fc73677013a093c38a98c141a1fb5d8"} err="failed to get container status \"4f180d0a5f22f977c7b9cda23a893b756fc73677013a093c38a98c141a1fb5d8\": rpc error: code = NotFound desc = could not find container \"4f180d0a5f22f977c7b9cda23a893b756fc73677013a093c38a98c141a1fb5d8\": container with ID starting with 4f180d0a5f22f977c7b9cda23a893b756fc73677013a093c38a98c141a1fb5d8 not found: ID does not exist" Dec 11 02:23:50 crc kubenswrapper[4824]: I1211 02:23:50.939216 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae139cc3-a38e-45fa-a3b4-e507ad5a99d6-kube-api-access-jh6td" (OuterVolumeSpecName: "kube-api-access-jh6td") pod "ae139cc3-a38e-45fa-a3b4-e507ad5a99d6" (UID: "ae139cc3-a38e-45fa-a3b4-e507ad5a99d6"). InnerVolumeSpecName "kube-api-access-jh6td". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:23:50 crc kubenswrapper[4824]: I1211 02:23:50.973307 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ae139cc3-a38e-45fa-a3b4-e507ad5a99d6-config" (OuterVolumeSpecName: "config") pod "ae139cc3-a38e-45fa-a3b4-e507ad5a99d6" (UID: "ae139cc3-a38e-45fa-a3b4-e507ad5a99d6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:23:50 crc kubenswrapper[4824]: I1211 02:23:50.994459 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ae139cc3-a38e-45fa-a3b4-e507ad5a99d6-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "ae139cc3-a38e-45fa-a3b4-e507ad5a99d6" (UID: "ae139cc3-a38e-45fa-a3b4-e507ad5a99d6"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:23:50 crc kubenswrapper[4824]: I1211 02:23:50.995645 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ae139cc3-a38e-45fa-a3b4-e507ad5a99d6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ae139cc3-a38e-45fa-a3b4-e507ad5a99d6" (UID: "ae139cc3-a38e-45fa-a3b4-e507ad5a99d6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:23:50 crc kubenswrapper[4824]: I1211 02:23:50.998125 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ae139cc3-a38e-45fa-a3b4-e507ad5a99d6-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ae139cc3-a38e-45fa-a3b4-e507ad5a99d6" (UID: "ae139cc3-a38e-45fa-a3b4-e507ad5a99d6"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.019743 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.020139 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="f7af70e6-c4c5-477f-a5af-876490e4ddf7" containerName="nova-api-log" containerID="cri-o://1115cb51cefd19b75e6688302803085b6377a4ad2a750518905c401166fd9d32" gracePeriod=30 Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.020515 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jh6td\" (UniqueName: \"kubernetes.io/projected/ae139cc3-a38e-45fa-a3b4-e507ad5a99d6-kube-api-access-jh6td\") on node \"crc\" DevicePath \"\"" Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.020548 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ae139cc3-a38e-45fa-a3b4-e507ad5a99d6-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.020557 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae139cc3-a38e-45fa-a3b4-e507ad5a99d6-config\") on node \"crc\" DevicePath \"\"" Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.020565 4824 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ae139cc3-a38e-45fa-a3b4-e507ad5a99d6-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.020574 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ae139cc3-a38e-45fa-a3b4-e507ad5a99d6-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.020649 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="f7af70e6-c4c5-477f-a5af-876490e4ddf7" containerName="nova-api-api" containerID="cri-o://4ed05f464df1d6511a61ef661dce449fda7495a35fa8f436707fada58aaee475" gracePeriod=30 Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.029702 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ae139cc3-a38e-45fa-a3b4-e507ad5a99d6-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ae139cc3-a38e-45fa-a3b4-e507ad5a99d6" (UID: "ae139cc3-a38e-45fa-a3b4-e507ad5a99d6"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.067044 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="f7af70e6-c4c5-477f-a5af-876490e4ddf7" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.188:8774/\": EOF" Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.067296 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="f7af70e6-c4c5-477f-a5af-876490e4ddf7" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.188:8774/\": EOF" Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.092008 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.092522 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="6c9b7563-9ee3-4f4f-959c-67cfc4ad0727" containerName="nova-metadata-log" containerID="cri-o://f2c54559ef781147992a2bc3c1b26eb2d414ab37f5e26ba7aa8fe3c4389eaecc" gracePeriod=30 Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.093063 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="6c9b7563-9ee3-4f4f-959c-67cfc4ad0727" containerName="nova-metadata-metadata" containerID="cri-o://6c6b5ca9d57945790a51ad54a14928ae599fc4d950be7ce39b497582610f9a11" gracePeriod=30 Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.122287 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ae139cc3-a38e-45fa-a3b4-e507ad5a99d6-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.370378 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-2k725" Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.394748 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.404371 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-rng95"] Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.411982 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-rng95"] Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.429583 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a30aa0d2-55dc-4375-b72d-953a3c57043f-scripts\") pod \"a30aa0d2-55dc-4375-b72d-953a3c57043f\" (UID: \"a30aa0d2-55dc-4375-b72d-953a3c57043f\") " Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.429737 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ztrqr\" (UniqueName: \"kubernetes.io/projected/a30aa0d2-55dc-4375-b72d-953a3c57043f-kube-api-access-ztrqr\") pod \"a30aa0d2-55dc-4375-b72d-953a3c57043f\" (UID: \"a30aa0d2-55dc-4375-b72d-953a3c57043f\") " Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.429796 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a30aa0d2-55dc-4375-b72d-953a3c57043f-config-data\") pod \"a30aa0d2-55dc-4375-b72d-953a3c57043f\" (UID: \"a30aa0d2-55dc-4375-b72d-953a3c57043f\") " Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.429881 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a30aa0d2-55dc-4375-b72d-953a3c57043f-combined-ca-bundle\") pod \"a30aa0d2-55dc-4375-b72d-953a3c57043f\" (UID: \"a30aa0d2-55dc-4375-b72d-953a3c57043f\") " Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.440643 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a30aa0d2-55dc-4375-b72d-953a3c57043f-kube-api-access-ztrqr" (OuterVolumeSpecName: "kube-api-access-ztrqr") pod "a30aa0d2-55dc-4375-b72d-953a3c57043f" (UID: "a30aa0d2-55dc-4375-b72d-953a3c57043f"). InnerVolumeSpecName "kube-api-access-ztrqr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.479261 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a30aa0d2-55dc-4375-b72d-953a3c57043f-scripts" (OuterVolumeSpecName: "scripts") pod "a30aa0d2-55dc-4375-b72d-953a3c57043f" (UID: "a30aa0d2-55dc-4375-b72d-953a3c57043f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.494275 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a30aa0d2-55dc-4375-b72d-953a3c57043f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a30aa0d2-55dc-4375-b72d-953a3c57043f" (UID: "a30aa0d2-55dc-4375-b72d-953a3c57043f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.517661 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a30aa0d2-55dc-4375-b72d-953a3c57043f-config-data" (OuterVolumeSpecName: "config-data") pod "a30aa0d2-55dc-4375-b72d-953a3c57043f" (UID: "a30aa0d2-55dc-4375-b72d-953a3c57043f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.534756 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a30aa0d2-55dc-4375-b72d-953a3c57043f-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.534816 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a30aa0d2-55dc-4375-b72d-953a3c57043f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.534832 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a30aa0d2-55dc-4375-b72d-953a3c57043f-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.534847 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ztrqr\" (UniqueName: \"kubernetes.io/projected/a30aa0d2-55dc-4375-b72d-953a3c57043f-kube-api-access-ztrqr\") on node \"crc\" DevicePath \"\"" Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.725460 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.741446 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6c9b7563-9ee3-4f4f-959c-67cfc4ad0727-nova-metadata-tls-certs\") pod \"6c9b7563-9ee3-4f4f-959c-67cfc4ad0727\" (UID: \"6c9b7563-9ee3-4f4f-959c-67cfc4ad0727\") " Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.741606 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c9b7563-9ee3-4f4f-959c-67cfc4ad0727-config-data\") pod \"6c9b7563-9ee3-4f4f-959c-67cfc4ad0727\" (UID: \"6c9b7563-9ee3-4f4f-959c-67cfc4ad0727\") " Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.741661 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xx9cf\" (UniqueName: \"kubernetes.io/projected/6c9b7563-9ee3-4f4f-959c-67cfc4ad0727-kube-api-access-xx9cf\") pod \"6c9b7563-9ee3-4f4f-959c-67cfc4ad0727\" (UID: \"6c9b7563-9ee3-4f4f-959c-67cfc4ad0727\") " Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.741710 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6c9b7563-9ee3-4f4f-959c-67cfc4ad0727-logs\") pod \"6c9b7563-9ee3-4f4f-959c-67cfc4ad0727\" (UID: \"6c9b7563-9ee3-4f4f-959c-67cfc4ad0727\") " Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.741747 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c9b7563-9ee3-4f4f-959c-67cfc4ad0727-combined-ca-bundle\") pod \"6c9b7563-9ee3-4f4f-959c-67cfc4ad0727\" (UID: \"6c9b7563-9ee3-4f4f-959c-67cfc4ad0727\") " Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.746070 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6c9b7563-9ee3-4f4f-959c-67cfc4ad0727-logs" (OuterVolumeSpecName: "logs") pod "6c9b7563-9ee3-4f4f-959c-67cfc4ad0727" (UID: "6c9b7563-9ee3-4f4f-959c-67cfc4ad0727"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.767165 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c9b7563-9ee3-4f4f-959c-67cfc4ad0727-kube-api-access-xx9cf" (OuterVolumeSpecName: "kube-api-access-xx9cf") pod "6c9b7563-9ee3-4f4f-959c-67cfc4ad0727" (UID: "6c9b7563-9ee3-4f4f-959c-67cfc4ad0727"). InnerVolumeSpecName "kube-api-access-xx9cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:23:51 crc kubenswrapper[4824]: E1211 02:23:51.773649 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podae139cc3_a38e_45fa_a3b4_e507ad5a99d6.slice/crio-c7f896f81d1071e759f97e85536990b6e57d9b8c2ae4db5ea3a96cc5b4a8e81d\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podae139cc3_a38e_45fa_a3b4_e507ad5a99d6.slice\": RecentStats: unable to find data in memory cache]" Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.777221 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c9b7563-9ee3-4f4f-959c-67cfc4ad0727-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6c9b7563-9ee3-4f4f-959c-67cfc4ad0727" (UID: "6c9b7563-9ee3-4f4f-959c-67cfc4ad0727"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.789051 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c9b7563-9ee3-4f4f-959c-67cfc4ad0727-config-data" (OuterVolumeSpecName: "config-data") pod "6c9b7563-9ee3-4f4f-959c-67cfc4ad0727" (UID: "6c9b7563-9ee3-4f4f-959c-67cfc4ad0727"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.799191 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c9b7563-9ee3-4f4f-959c-67cfc4ad0727-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "6c9b7563-9ee3-4f4f-959c-67cfc4ad0727" (UID: "6c9b7563-9ee3-4f4f-959c-67cfc4ad0727"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.845932 4824 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6c9b7563-9ee3-4f4f-959c-67cfc4ad0727-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.845969 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c9b7563-9ee3-4f4f-959c-67cfc4ad0727-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.845981 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xx9cf\" (UniqueName: \"kubernetes.io/projected/6c9b7563-9ee3-4f4f-959c-67cfc4ad0727-kube-api-access-xx9cf\") on node \"crc\" DevicePath \"\"" Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.845992 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6c9b7563-9ee3-4f4f-959c-67cfc4ad0727-logs\") on node \"crc\" DevicePath \"\"" Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.846002 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c9b7563-9ee3-4f4f-959c-67cfc4ad0727-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.860282 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-2k725" Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.860269 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-2k725" event={"ID":"a30aa0d2-55dc-4375-b72d-953a3c57043f","Type":"ContainerDied","Data":"03f614b0bff13fef3b7ca08a4af0dfbb5c5cf7c6de1d57ae53f583b51b920184"} Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.861160 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="03f614b0bff13fef3b7ca08a4af0dfbb5c5cf7c6de1d57ae53f583b51b920184" Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.864447 4824 generic.go:334] "Generic (PLEG): container finished" podID="6c9b7563-9ee3-4f4f-959c-67cfc4ad0727" containerID="6c6b5ca9d57945790a51ad54a14928ae599fc4d950be7ce39b497582610f9a11" exitCode=0 Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.864480 4824 generic.go:334] "Generic (PLEG): container finished" podID="6c9b7563-9ee3-4f4f-959c-67cfc4ad0727" containerID="f2c54559ef781147992a2bc3c1b26eb2d414ab37f5e26ba7aa8fe3c4389eaecc" exitCode=143 Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.864517 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6c9b7563-9ee3-4f4f-959c-67cfc4ad0727","Type":"ContainerDied","Data":"6c6b5ca9d57945790a51ad54a14928ae599fc4d950be7ce39b497582610f9a11"} Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.864544 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6c9b7563-9ee3-4f4f-959c-67cfc4ad0727","Type":"ContainerDied","Data":"f2c54559ef781147992a2bc3c1b26eb2d414ab37f5e26ba7aa8fe3c4389eaecc"} Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.864553 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6c9b7563-9ee3-4f4f-959c-67cfc4ad0727","Type":"ContainerDied","Data":"b7148eae3dff00fb01489eef660e6e679f6e4996bb3ce74d8625145751d2f74d"} Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.864567 4824 scope.go:117] "RemoveContainer" containerID="6c6b5ca9d57945790a51ad54a14928ae599fc4d950be7ce39b497582610f9a11" Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.864707 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.896716 4824 generic.go:334] "Generic (PLEG): container finished" podID="f7af70e6-c4c5-477f-a5af-876490e4ddf7" containerID="1115cb51cefd19b75e6688302803085b6377a4ad2a750518905c401166fd9d32" exitCode=143 Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.897139 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f7af70e6-c4c5-477f-a5af-876490e4ddf7","Type":"ContainerDied","Data":"1115cb51cefd19b75e6688302803085b6377a4ad2a750518905c401166fd9d32"} Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.943474 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 11 02:23:51 crc kubenswrapper[4824]: E1211 02:23:51.943989 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a30aa0d2-55dc-4375-b72d-953a3c57043f" containerName="nova-cell1-conductor-db-sync" Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.944009 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="a30aa0d2-55dc-4375-b72d-953a3c57043f" containerName="nova-cell1-conductor-db-sync" Dec 11 02:23:51 crc kubenswrapper[4824]: E1211 02:23:51.944021 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c9b7563-9ee3-4f4f-959c-67cfc4ad0727" containerName="nova-metadata-metadata" Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.944028 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c9b7563-9ee3-4f4f-959c-67cfc4ad0727" containerName="nova-metadata-metadata" Dec 11 02:23:51 crc kubenswrapper[4824]: E1211 02:23:51.944039 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae139cc3-a38e-45fa-a3b4-e507ad5a99d6" containerName="dnsmasq-dns" Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.944046 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae139cc3-a38e-45fa-a3b4-e507ad5a99d6" containerName="dnsmasq-dns" Dec 11 02:23:51 crc kubenswrapper[4824]: E1211 02:23:51.944063 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f973a712-53f1-4d52-8767-c65f665dcb41" containerName="nova-manage" Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.944069 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f973a712-53f1-4d52-8767-c65f665dcb41" containerName="nova-manage" Dec 11 02:23:51 crc kubenswrapper[4824]: E1211 02:23:51.944077 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae139cc3-a38e-45fa-a3b4-e507ad5a99d6" containerName="init" Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.944082 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae139cc3-a38e-45fa-a3b4-e507ad5a99d6" containerName="init" Dec 11 02:23:51 crc kubenswrapper[4824]: E1211 02:23:51.944096 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c9b7563-9ee3-4f4f-959c-67cfc4ad0727" containerName="nova-metadata-log" Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.944103 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c9b7563-9ee3-4f4f-959c-67cfc4ad0727" containerName="nova-metadata-log" Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.944306 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c9b7563-9ee3-4f4f-959c-67cfc4ad0727" containerName="nova-metadata-metadata" Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.944321 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f973a712-53f1-4d52-8767-c65f665dcb41" containerName="nova-manage" Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.944333 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c9b7563-9ee3-4f4f-959c-67cfc4ad0727" containerName="nova-metadata-log" Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.944353 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae139cc3-a38e-45fa-a3b4-e507ad5a99d6" containerName="dnsmasq-dns" Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.944364 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="a30aa0d2-55dc-4375-b72d-953a3c57043f" containerName="nova-cell1-conductor-db-sync" Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.945065 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.947649 4824 scope.go:117] "RemoveContainer" containerID="f2c54559ef781147992a2bc3c1b26eb2d414ab37f5e26ba7aa8fe3c4389eaecc" Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.951707 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.957957 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.977768 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.989222 4824 scope.go:117] "RemoveContainer" containerID="6c6b5ca9d57945790a51ad54a14928ae599fc4d950be7ce39b497582610f9a11" Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.990605 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 11 02:23:51 crc kubenswrapper[4824]: E1211 02:23:51.991854 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c6b5ca9d57945790a51ad54a14928ae599fc4d950be7ce39b497582610f9a11\": container with ID starting with 6c6b5ca9d57945790a51ad54a14928ae599fc4d950be7ce39b497582610f9a11 not found: ID does not exist" containerID="6c6b5ca9d57945790a51ad54a14928ae599fc4d950be7ce39b497582610f9a11" Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.991888 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c6b5ca9d57945790a51ad54a14928ae599fc4d950be7ce39b497582610f9a11"} err="failed to get container status \"6c6b5ca9d57945790a51ad54a14928ae599fc4d950be7ce39b497582610f9a11\": rpc error: code = NotFound desc = could not find container \"6c6b5ca9d57945790a51ad54a14928ae599fc4d950be7ce39b497582610f9a11\": container with ID starting with 6c6b5ca9d57945790a51ad54a14928ae599fc4d950be7ce39b497582610f9a11 not found: ID does not exist" Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.991912 4824 scope.go:117] "RemoveContainer" containerID="f2c54559ef781147992a2bc3c1b26eb2d414ab37f5e26ba7aa8fe3c4389eaecc" Dec 11 02:23:51 crc kubenswrapper[4824]: E1211 02:23:51.992205 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f2c54559ef781147992a2bc3c1b26eb2d414ab37f5e26ba7aa8fe3c4389eaecc\": container with ID starting with f2c54559ef781147992a2bc3c1b26eb2d414ab37f5e26ba7aa8fe3c4389eaecc not found: ID does not exist" containerID="f2c54559ef781147992a2bc3c1b26eb2d414ab37f5e26ba7aa8fe3c4389eaecc" Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.992237 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2c54559ef781147992a2bc3c1b26eb2d414ab37f5e26ba7aa8fe3c4389eaecc"} err="failed to get container status \"f2c54559ef781147992a2bc3c1b26eb2d414ab37f5e26ba7aa8fe3c4389eaecc\": rpc error: code = NotFound desc = could not find container \"f2c54559ef781147992a2bc3c1b26eb2d414ab37f5e26ba7aa8fe3c4389eaecc\": container with ID starting with f2c54559ef781147992a2bc3c1b26eb2d414ab37f5e26ba7aa8fe3c4389eaecc not found: ID does not exist" Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.992265 4824 scope.go:117] "RemoveContainer" containerID="6c6b5ca9d57945790a51ad54a14928ae599fc4d950be7ce39b497582610f9a11" Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.992498 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c6b5ca9d57945790a51ad54a14928ae599fc4d950be7ce39b497582610f9a11"} err="failed to get container status \"6c6b5ca9d57945790a51ad54a14928ae599fc4d950be7ce39b497582610f9a11\": rpc error: code = NotFound desc = could not find container \"6c6b5ca9d57945790a51ad54a14928ae599fc4d950be7ce39b497582610f9a11\": container with ID starting with 6c6b5ca9d57945790a51ad54a14928ae599fc4d950be7ce39b497582610f9a11 not found: ID does not exist" Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.992538 4824 scope.go:117] "RemoveContainer" containerID="f2c54559ef781147992a2bc3c1b26eb2d414ab37f5e26ba7aa8fe3c4389eaecc" Dec 11 02:23:51 crc kubenswrapper[4824]: I1211 02:23:51.992714 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2c54559ef781147992a2bc3c1b26eb2d414ab37f5e26ba7aa8fe3c4389eaecc"} err="failed to get container status \"f2c54559ef781147992a2bc3c1b26eb2d414ab37f5e26ba7aa8fe3c4389eaecc\": rpc error: code = NotFound desc = could not find container \"f2c54559ef781147992a2bc3c1b26eb2d414ab37f5e26ba7aa8fe3c4389eaecc\": container with ID starting with f2c54559ef781147992a2bc3c1b26eb2d414ab37f5e26ba7aa8fe3c4389eaecc not found: ID does not exist" Dec 11 02:23:52 crc kubenswrapper[4824]: I1211 02:23:52.000194 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 11 02:23:52 crc kubenswrapper[4824]: I1211 02:23:52.001768 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 11 02:23:52 crc kubenswrapper[4824]: I1211 02:23:52.004389 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 11 02:23:52 crc kubenswrapper[4824]: I1211 02:23:52.004549 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 11 02:23:52 crc kubenswrapper[4824]: I1211 02:23:52.006247 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 11 02:23:52 crc kubenswrapper[4824]: I1211 02:23:52.048644 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2a3020fe-d2be-4c32-9f69-f5365c4a10c4-logs\") pod \"nova-metadata-0\" (UID: \"2a3020fe-d2be-4c32-9f69-f5365c4a10c4\") " pod="openstack/nova-metadata-0" Dec 11 02:23:52 crc kubenswrapper[4824]: I1211 02:23:52.048707 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fddad47b-d680-4be3-8bfa-a7708daef973-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"fddad47b-d680-4be3-8bfa-a7708daef973\") " pod="openstack/nova-cell1-conductor-0" Dec 11 02:23:52 crc kubenswrapper[4824]: I1211 02:23:52.048754 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a3020fe-d2be-4c32-9f69-f5365c4a10c4-config-data\") pod \"nova-metadata-0\" (UID: \"2a3020fe-d2be-4c32-9f69-f5365c4a10c4\") " pod="openstack/nova-metadata-0" Dec 11 02:23:52 crc kubenswrapper[4824]: I1211 02:23:52.049020 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a3020fe-d2be-4c32-9f69-f5365c4a10c4-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2a3020fe-d2be-4c32-9f69-f5365c4a10c4\") " pod="openstack/nova-metadata-0" Dec 11 02:23:52 crc kubenswrapper[4824]: I1211 02:23:52.049267 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fddad47b-d680-4be3-8bfa-a7708daef973-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"fddad47b-d680-4be3-8bfa-a7708daef973\") " pod="openstack/nova-cell1-conductor-0" Dec 11 02:23:52 crc kubenswrapper[4824]: I1211 02:23:52.049369 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wbll5\" (UniqueName: \"kubernetes.io/projected/fddad47b-d680-4be3-8bfa-a7708daef973-kube-api-access-wbll5\") pod \"nova-cell1-conductor-0\" (UID: \"fddad47b-d680-4be3-8bfa-a7708daef973\") " pod="openstack/nova-cell1-conductor-0" Dec 11 02:23:52 crc kubenswrapper[4824]: I1211 02:23:52.049438 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5p9xp\" (UniqueName: \"kubernetes.io/projected/2a3020fe-d2be-4c32-9f69-f5365c4a10c4-kube-api-access-5p9xp\") pod \"nova-metadata-0\" (UID: \"2a3020fe-d2be-4c32-9f69-f5365c4a10c4\") " pod="openstack/nova-metadata-0" Dec 11 02:23:52 crc kubenswrapper[4824]: I1211 02:23:52.049501 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/2a3020fe-d2be-4c32-9f69-f5365c4a10c4-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"2a3020fe-d2be-4c32-9f69-f5365c4a10c4\") " pod="openstack/nova-metadata-0" Dec 11 02:23:52 crc kubenswrapper[4824]: I1211 02:23:52.150261 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wbll5\" (UniqueName: \"kubernetes.io/projected/fddad47b-d680-4be3-8bfa-a7708daef973-kube-api-access-wbll5\") pod \"nova-cell1-conductor-0\" (UID: \"fddad47b-d680-4be3-8bfa-a7708daef973\") " pod="openstack/nova-cell1-conductor-0" Dec 11 02:23:52 crc kubenswrapper[4824]: I1211 02:23:52.150574 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5p9xp\" (UniqueName: \"kubernetes.io/projected/2a3020fe-d2be-4c32-9f69-f5365c4a10c4-kube-api-access-5p9xp\") pod \"nova-metadata-0\" (UID: \"2a3020fe-d2be-4c32-9f69-f5365c4a10c4\") " pod="openstack/nova-metadata-0" Dec 11 02:23:52 crc kubenswrapper[4824]: I1211 02:23:52.150599 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/2a3020fe-d2be-4c32-9f69-f5365c4a10c4-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"2a3020fe-d2be-4c32-9f69-f5365c4a10c4\") " pod="openstack/nova-metadata-0" Dec 11 02:23:52 crc kubenswrapper[4824]: I1211 02:23:52.150654 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2a3020fe-d2be-4c32-9f69-f5365c4a10c4-logs\") pod \"nova-metadata-0\" (UID: \"2a3020fe-d2be-4c32-9f69-f5365c4a10c4\") " pod="openstack/nova-metadata-0" Dec 11 02:23:52 crc kubenswrapper[4824]: I1211 02:23:52.150677 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fddad47b-d680-4be3-8bfa-a7708daef973-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"fddad47b-d680-4be3-8bfa-a7708daef973\") " pod="openstack/nova-cell1-conductor-0" Dec 11 02:23:52 crc kubenswrapper[4824]: I1211 02:23:52.150710 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a3020fe-d2be-4c32-9f69-f5365c4a10c4-config-data\") pod \"nova-metadata-0\" (UID: \"2a3020fe-d2be-4c32-9f69-f5365c4a10c4\") " pod="openstack/nova-metadata-0" Dec 11 02:23:52 crc kubenswrapper[4824]: I1211 02:23:52.150753 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a3020fe-d2be-4c32-9f69-f5365c4a10c4-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2a3020fe-d2be-4c32-9f69-f5365c4a10c4\") " pod="openstack/nova-metadata-0" Dec 11 02:23:52 crc kubenswrapper[4824]: I1211 02:23:52.150780 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fddad47b-d680-4be3-8bfa-a7708daef973-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"fddad47b-d680-4be3-8bfa-a7708daef973\") " pod="openstack/nova-cell1-conductor-0" Dec 11 02:23:52 crc kubenswrapper[4824]: I1211 02:23:52.152323 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2a3020fe-d2be-4c32-9f69-f5365c4a10c4-logs\") pod \"nova-metadata-0\" (UID: \"2a3020fe-d2be-4c32-9f69-f5365c4a10c4\") " pod="openstack/nova-metadata-0" Dec 11 02:23:52 crc kubenswrapper[4824]: I1211 02:23:52.155528 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a3020fe-d2be-4c32-9f69-f5365c4a10c4-config-data\") pod \"nova-metadata-0\" (UID: \"2a3020fe-d2be-4c32-9f69-f5365c4a10c4\") " pod="openstack/nova-metadata-0" Dec 11 02:23:52 crc kubenswrapper[4824]: I1211 02:23:52.156239 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fddad47b-d680-4be3-8bfa-a7708daef973-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"fddad47b-d680-4be3-8bfa-a7708daef973\") " pod="openstack/nova-cell1-conductor-0" Dec 11 02:23:52 crc kubenswrapper[4824]: I1211 02:23:52.157800 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a3020fe-d2be-4c32-9f69-f5365c4a10c4-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2a3020fe-d2be-4c32-9f69-f5365c4a10c4\") " pod="openstack/nova-metadata-0" Dec 11 02:23:52 crc kubenswrapper[4824]: I1211 02:23:52.158186 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fddad47b-d680-4be3-8bfa-a7708daef973-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"fddad47b-d680-4be3-8bfa-a7708daef973\") " pod="openstack/nova-cell1-conductor-0" Dec 11 02:23:52 crc kubenswrapper[4824]: I1211 02:23:52.158728 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/2a3020fe-d2be-4c32-9f69-f5365c4a10c4-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"2a3020fe-d2be-4c32-9f69-f5365c4a10c4\") " pod="openstack/nova-metadata-0" Dec 11 02:23:52 crc kubenswrapper[4824]: I1211 02:23:52.169068 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wbll5\" (UniqueName: \"kubernetes.io/projected/fddad47b-d680-4be3-8bfa-a7708daef973-kube-api-access-wbll5\") pod \"nova-cell1-conductor-0\" (UID: \"fddad47b-d680-4be3-8bfa-a7708daef973\") " pod="openstack/nova-cell1-conductor-0" Dec 11 02:23:52 crc kubenswrapper[4824]: I1211 02:23:52.171241 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5p9xp\" (UniqueName: \"kubernetes.io/projected/2a3020fe-d2be-4c32-9f69-f5365c4a10c4-kube-api-access-5p9xp\") pod \"nova-metadata-0\" (UID: \"2a3020fe-d2be-4c32-9f69-f5365c4a10c4\") " pod="openstack/nova-metadata-0" Dec 11 02:23:52 crc kubenswrapper[4824]: I1211 02:23:52.268274 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 11 02:23:52 crc kubenswrapper[4824]: I1211 02:23:52.321646 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 11 02:23:52 crc kubenswrapper[4824]: I1211 02:23:52.644319 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6c9b7563-9ee3-4f4f-959c-67cfc4ad0727" path="/var/lib/kubelet/pods/6c9b7563-9ee3-4f4f-959c-67cfc4ad0727/volumes" Dec 11 02:23:52 crc kubenswrapper[4824]: I1211 02:23:52.645399 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae139cc3-a38e-45fa-a3b4-e507ad5a99d6" path="/var/lib/kubelet/pods/ae139cc3-a38e-45fa-a3b4-e507ad5a99d6/volumes" Dec 11 02:23:52 crc kubenswrapper[4824]: I1211 02:23:52.752126 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 11 02:23:52 crc kubenswrapper[4824]: I1211 02:23:52.887738 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 11 02:23:52 crc kubenswrapper[4824]: W1211 02:23:52.895621 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2a3020fe_d2be_4c32_9f69_f5365c4a10c4.slice/crio-2f4adeabdb58f8f9ce02c9a1035b476dcd67014565ce40d73f7bb961625a3b4a WatchSource:0}: Error finding container 2f4adeabdb58f8f9ce02c9a1035b476dcd67014565ce40d73f7bb961625a3b4a: Status 404 returned error can't find the container with id 2f4adeabdb58f8f9ce02c9a1035b476dcd67014565ce40d73f7bb961625a3b4a Dec 11 02:23:52 crc kubenswrapper[4824]: I1211 02:23:52.914747 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"fddad47b-d680-4be3-8bfa-a7708daef973","Type":"ContainerStarted","Data":"e5d47980f1c6717a7f20ab8e1e962e51a2144438623aaf21249c21ff8d1a2eb4"} Dec 11 02:23:52 crc kubenswrapper[4824]: I1211 02:23:52.916335 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2a3020fe-d2be-4c32-9f69-f5365c4a10c4","Type":"ContainerStarted","Data":"2f4adeabdb58f8f9ce02c9a1035b476dcd67014565ce40d73f7bb961625a3b4a"} Dec 11 02:23:52 crc kubenswrapper[4824]: I1211 02:23:52.927666 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="d3f62926-29c4-4fd2-8e1b-926a53823b65" containerName="nova-scheduler-scheduler" containerID="cri-o://ae5e31f6a12cd22e1a504146c800e694804f108905cb4e2c880cb196fa831a42" gracePeriod=30 Dec 11 02:23:53 crc kubenswrapper[4824]: I1211 02:23:53.938683 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"fddad47b-d680-4be3-8bfa-a7708daef973","Type":"ContainerStarted","Data":"68064aa1bee88a41fd0998d71e120435c9aadc48bacc309b26811f569ff202fa"} Dec 11 02:23:53 crc kubenswrapper[4824]: I1211 02:23:53.939861 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Dec 11 02:23:53 crc kubenswrapper[4824]: I1211 02:23:53.946624 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2a3020fe-d2be-4c32-9f69-f5365c4a10c4","Type":"ContainerStarted","Data":"1a7e5ed0325ecaba26d9852ff51ba1c50eb8facc2eafffc9eb9fb555445e6a28"} Dec 11 02:23:53 crc kubenswrapper[4824]: I1211 02:23:53.946713 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2a3020fe-d2be-4c32-9f69-f5365c4a10c4","Type":"ContainerStarted","Data":"fc00c4365c55304300d9d6ffee4f4185a4318fb2439681904c5bcb773a9482cc"} Dec 11 02:23:54 crc kubenswrapper[4824]: I1211 02:23:54.028104 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=3.028082828 podStartE2EDuration="3.028082828s" podCreationTimestamp="2025-12-11 02:23:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:23:53.965827385 +0000 UTC m=+1375.654864804" watchObservedRunningTime="2025-12-11 02:23:54.028082828 +0000 UTC m=+1375.717120207" Dec 11 02:23:54 crc kubenswrapper[4824]: I1211 02:23:54.041076 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.041049284 podStartE2EDuration="3.041049284s" podCreationTimestamp="2025-12-11 02:23:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:23:54.017447702 +0000 UTC m=+1375.706485111" watchObservedRunningTime="2025-12-11 02:23:54.041049284 +0000 UTC m=+1375.730086703" Dec 11 02:23:54 crc kubenswrapper[4824]: E1211 02:23:54.995996 4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ae5e31f6a12cd22e1a504146c800e694804f108905cb4e2c880cb196fa831a42" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 11 02:23:54 crc kubenswrapper[4824]: E1211 02:23:54.997191 4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ae5e31f6a12cd22e1a504146c800e694804f108905cb4e2c880cb196fa831a42" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 11 02:23:54 crc kubenswrapper[4824]: E1211 02:23:54.999045 4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ae5e31f6a12cd22e1a504146c800e694804f108905cb4e2c880cb196fa831a42" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 11 02:23:54 crc kubenswrapper[4824]: E1211 02:23:54.999136 4824 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="d3f62926-29c4-4fd2-8e1b-926a53823b65" containerName="nova-scheduler-scheduler" Dec 11 02:23:56 crc kubenswrapper[4824]: I1211 02:23:56.755181 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 11 02:23:56 crc kubenswrapper[4824]: I1211 02:23:56.821258 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 11 02:23:56 crc kubenswrapper[4824]: I1211 02:23:56.937402 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7af70e6-c4c5-477f-a5af-876490e4ddf7-config-data\") pod \"f7af70e6-c4c5-477f-a5af-876490e4ddf7\" (UID: \"f7af70e6-c4c5-477f-a5af-876490e4ddf7\") " Dec 11 02:23:56 crc kubenswrapper[4824]: I1211 02:23:56.937804 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3f62926-29c4-4fd2-8e1b-926a53823b65-config-data\") pod \"d3f62926-29c4-4fd2-8e1b-926a53823b65\" (UID: \"d3f62926-29c4-4fd2-8e1b-926a53823b65\") " Dec 11 02:23:56 crc kubenswrapper[4824]: I1211 02:23:56.937834 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sdxcl\" (UniqueName: \"kubernetes.io/projected/d3f62926-29c4-4fd2-8e1b-926a53823b65-kube-api-access-sdxcl\") pod \"d3f62926-29c4-4fd2-8e1b-926a53823b65\" (UID: \"d3f62926-29c4-4fd2-8e1b-926a53823b65\") " Dec 11 02:23:56 crc kubenswrapper[4824]: I1211 02:23:56.937888 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3f62926-29c4-4fd2-8e1b-926a53823b65-combined-ca-bundle\") pod \"d3f62926-29c4-4fd2-8e1b-926a53823b65\" (UID: \"d3f62926-29c4-4fd2-8e1b-926a53823b65\") " Dec 11 02:23:56 crc kubenswrapper[4824]: I1211 02:23:56.937933 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7af70e6-c4c5-477f-a5af-876490e4ddf7-combined-ca-bundle\") pod \"f7af70e6-c4c5-477f-a5af-876490e4ddf7\" (UID: \"f7af70e6-c4c5-477f-a5af-876490e4ddf7\") " Dec 11 02:23:56 crc kubenswrapper[4824]: I1211 02:23:56.937979 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f7af70e6-c4c5-477f-a5af-876490e4ddf7-logs\") pod \"f7af70e6-c4c5-477f-a5af-876490e4ddf7\" (UID: \"f7af70e6-c4c5-477f-a5af-876490e4ddf7\") " Dec 11 02:23:56 crc kubenswrapper[4824]: I1211 02:23:56.938021 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rfm97\" (UniqueName: \"kubernetes.io/projected/f7af70e6-c4c5-477f-a5af-876490e4ddf7-kube-api-access-rfm97\") pod \"f7af70e6-c4c5-477f-a5af-876490e4ddf7\" (UID: \"f7af70e6-c4c5-477f-a5af-876490e4ddf7\") " Dec 11 02:23:56 crc kubenswrapper[4824]: I1211 02:23:56.938570 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f7af70e6-c4c5-477f-a5af-876490e4ddf7-logs" (OuterVolumeSpecName: "logs") pod "f7af70e6-c4c5-477f-a5af-876490e4ddf7" (UID: "f7af70e6-c4c5-477f-a5af-876490e4ddf7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:23:56 crc kubenswrapper[4824]: I1211 02:23:56.942984 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3f62926-29c4-4fd2-8e1b-926a53823b65-kube-api-access-sdxcl" (OuterVolumeSpecName: "kube-api-access-sdxcl") pod "d3f62926-29c4-4fd2-8e1b-926a53823b65" (UID: "d3f62926-29c4-4fd2-8e1b-926a53823b65"). InnerVolumeSpecName "kube-api-access-sdxcl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:23:56 crc kubenswrapper[4824]: I1211 02:23:56.943344 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7af70e6-c4c5-477f-a5af-876490e4ddf7-kube-api-access-rfm97" (OuterVolumeSpecName: "kube-api-access-rfm97") pod "f7af70e6-c4c5-477f-a5af-876490e4ddf7" (UID: "f7af70e6-c4c5-477f-a5af-876490e4ddf7"). InnerVolumeSpecName "kube-api-access-rfm97". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:23:56 crc kubenswrapper[4824]: I1211 02:23:56.965246 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7af70e6-c4c5-477f-a5af-876490e4ddf7-config-data" (OuterVolumeSpecName: "config-data") pod "f7af70e6-c4c5-477f-a5af-876490e4ddf7" (UID: "f7af70e6-c4c5-477f-a5af-876490e4ddf7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:23:56 crc kubenswrapper[4824]: I1211 02:23:56.966124 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3f62926-29c4-4fd2-8e1b-926a53823b65-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d3f62926-29c4-4fd2-8e1b-926a53823b65" (UID: "d3f62926-29c4-4fd2-8e1b-926a53823b65"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:23:56 crc kubenswrapper[4824]: I1211 02:23:56.972260 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7af70e6-c4c5-477f-a5af-876490e4ddf7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f7af70e6-c4c5-477f-a5af-876490e4ddf7" (UID: "f7af70e6-c4c5-477f-a5af-876490e4ddf7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:23:56 crc kubenswrapper[4824]: I1211 02:23:56.975522 4824 generic.go:334] "Generic (PLEG): container finished" podID="d3f62926-29c4-4fd2-8e1b-926a53823b65" containerID="ae5e31f6a12cd22e1a504146c800e694804f108905cb4e2c880cb196fa831a42" exitCode=0 Dec 11 02:23:56 crc kubenswrapper[4824]: I1211 02:23:56.975661 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 11 02:23:56 crc kubenswrapper[4824]: I1211 02:23:56.975653 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d3f62926-29c4-4fd2-8e1b-926a53823b65","Type":"ContainerDied","Data":"ae5e31f6a12cd22e1a504146c800e694804f108905cb4e2c880cb196fa831a42"} Dec 11 02:23:56 crc kubenswrapper[4824]: I1211 02:23:56.975990 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d3f62926-29c4-4fd2-8e1b-926a53823b65","Type":"ContainerDied","Data":"f4c38321c7e01a8231ee7c762c1b2e29c9b7f44fac81d8bab4a5cedc47bda76b"} Dec 11 02:23:56 crc kubenswrapper[4824]: I1211 02:23:56.976034 4824 scope.go:117] "RemoveContainer" containerID="ae5e31f6a12cd22e1a504146c800e694804f108905cb4e2c880cb196fa831a42" Dec 11 02:23:56 crc kubenswrapper[4824]: I1211 02:23:56.978640 4824 generic.go:334] "Generic (PLEG): container finished" podID="f7af70e6-c4c5-477f-a5af-876490e4ddf7" containerID="4ed05f464df1d6511a61ef661dce449fda7495a35fa8f436707fada58aaee475" exitCode=0 Dec 11 02:23:56 crc kubenswrapper[4824]: I1211 02:23:56.978692 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f7af70e6-c4c5-477f-a5af-876490e4ddf7","Type":"ContainerDied","Data":"4ed05f464df1d6511a61ef661dce449fda7495a35fa8f436707fada58aaee475"} Dec 11 02:23:56 crc kubenswrapper[4824]: I1211 02:23:56.978822 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f7af70e6-c4c5-477f-a5af-876490e4ddf7","Type":"ContainerDied","Data":"260830c63558a99219a76a0c71bcae9a5b8cf7a3b8d34cc6e40391f3f2bff41b"} Dec 11 02:23:56 crc kubenswrapper[4824]: I1211 02:23:56.978714 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 11 02:23:56 crc kubenswrapper[4824]: I1211 02:23:56.983308 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3f62926-29c4-4fd2-8e1b-926a53823b65-config-data" (OuterVolumeSpecName: "config-data") pod "d3f62926-29c4-4fd2-8e1b-926a53823b65" (UID: "d3f62926-29c4-4fd2-8e1b-926a53823b65"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:23:57 crc kubenswrapper[4824]: I1211 02:23:57.012563 4824 scope.go:117] "RemoveContainer" containerID="ae5e31f6a12cd22e1a504146c800e694804f108905cb4e2c880cb196fa831a42" Dec 11 02:23:57 crc kubenswrapper[4824]: E1211 02:23:57.012914 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae5e31f6a12cd22e1a504146c800e694804f108905cb4e2c880cb196fa831a42\": container with ID starting with ae5e31f6a12cd22e1a504146c800e694804f108905cb4e2c880cb196fa831a42 not found: ID does not exist" containerID="ae5e31f6a12cd22e1a504146c800e694804f108905cb4e2c880cb196fa831a42" Dec 11 02:23:57 crc kubenswrapper[4824]: I1211 02:23:57.012945 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae5e31f6a12cd22e1a504146c800e694804f108905cb4e2c880cb196fa831a42"} err="failed to get container status \"ae5e31f6a12cd22e1a504146c800e694804f108905cb4e2c880cb196fa831a42\": rpc error: code = NotFound desc = could not find container \"ae5e31f6a12cd22e1a504146c800e694804f108905cb4e2c880cb196fa831a42\": container with ID starting with ae5e31f6a12cd22e1a504146c800e694804f108905cb4e2c880cb196fa831a42 not found: ID does not exist" Dec 11 02:23:57 crc kubenswrapper[4824]: I1211 02:23:57.012966 4824 scope.go:117] "RemoveContainer" containerID="4ed05f464df1d6511a61ef661dce449fda7495a35fa8f436707fada58aaee475" Dec 11 02:23:57 crc kubenswrapper[4824]: I1211 02:23:57.040735 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7af70e6-c4c5-477f-a5af-876490e4ddf7-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 02:23:57 crc kubenswrapper[4824]: I1211 02:23:57.040767 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3f62926-29c4-4fd2-8e1b-926a53823b65-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 02:23:57 crc kubenswrapper[4824]: I1211 02:23:57.040780 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sdxcl\" (UniqueName: \"kubernetes.io/projected/d3f62926-29c4-4fd2-8e1b-926a53823b65-kube-api-access-sdxcl\") on node \"crc\" DevicePath \"\"" Dec 11 02:23:57 crc kubenswrapper[4824]: I1211 02:23:57.040795 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7af70e6-c4c5-477f-a5af-876490e4ddf7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 02:23:57 crc kubenswrapper[4824]: I1211 02:23:57.040809 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3f62926-29c4-4fd2-8e1b-926a53823b65-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 02:23:57 crc kubenswrapper[4824]: I1211 02:23:57.040820 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f7af70e6-c4c5-477f-a5af-876490e4ddf7-logs\") on node \"crc\" DevicePath \"\"" Dec 11 02:23:57 crc kubenswrapper[4824]: I1211 02:23:57.040831 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rfm97\" (UniqueName: \"kubernetes.io/projected/f7af70e6-c4c5-477f-a5af-876490e4ddf7-kube-api-access-rfm97\") on node \"crc\" DevicePath \"\"" Dec 11 02:23:57 crc kubenswrapper[4824]: I1211 02:23:57.297546 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Dec 11 02:23:57 crc kubenswrapper[4824]: I1211 02:23:57.329874 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 11 02:23:57 crc kubenswrapper[4824]: I1211 02:23:57.330757 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 11 02:23:57 crc kubenswrapper[4824]: I1211 02:23:57.468313 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 11 02:23:57 crc kubenswrapper[4824]: I1211 02:23:57.478642 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 11 02:23:57 crc kubenswrapper[4824]: I1211 02:23:57.486137 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 11 02:23:57 crc kubenswrapper[4824]: I1211 02:23:57.495261 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 11 02:23:57 crc kubenswrapper[4824]: I1211 02:23:57.504286 4824 scope.go:117] "RemoveContainer" containerID="1115cb51cefd19b75e6688302803085b6377a4ad2a750518905c401166fd9d32" Dec 11 02:23:57 crc kubenswrapper[4824]: I1211 02:23:57.512516 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 11 02:23:57 crc kubenswrapper[4824]: E1211 02:23:57.513013 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7af70e6-c4c5-477f-a5af-876490e4ddf7" containerName="nova-api-log" Dec 11 02:23:57 crc kubenswrapper[4824]: I1211 02:23:57.513033 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7af70e6-c4c5-477f-a5af-876490e4ddf7" containerName="nova-api-log" Dec 11 02:23:57 crc kubenswrapper[4824]: E1211 02:23:57.513052 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7af70e6-c4c5-477f-a5af-876490e4ddf7" containerName="nova-api-api" Dec 11 02:23:57 crc kubenswrapper[4824]: I1211 02:23:57.513058 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7af70e6-c4c5-477f-a5af-876490e4ddf7" containerName="nova-api-api" Dec 11 02:23:57 crc kubenswrapper[4824]: E1211 02:23:57.513088 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3f62926-29c4-4fd2-8e1b-926a53823b65" containerName="nova-scheduler-scheduler" Dec 11 02:23:57 crc kubenswrapper[4824]: I1211 02:23:57.513094 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3f62926-29c4-4fd2-8e1b-926a53823b65" containerName="nova-scheduler-scheduler" Dec 11 02:23:57 crc kubenswrapper[4824]: I1211 02:23:57.513293 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7af70e6-c4c5-477f-a5af-876490e4ddf7" containerName="nova-api-api" Dec 11 02:23:57 crc kubenswrapper[4824]: I1211 02:23:57.513318 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3f62926-29c4-4fd2-8e1b-926a53823b65" containerName="nova-scheduler-scheduler" Dec 11 02:23:57 crc kubenswrapper[4824]: I1211 02:23:57.513333 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7af70e6-c4c5-477f-a5af-876490e4ddf7" containerName="nova-api-log" Dec 11 02:23:57 crc kubenswrapper[4824]: I1211 02:23:57.514105 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 11 02:23:57 crc kubenswrapper[4824]: I1211 02:23:57.518533 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 11 02:23:57 crc kubenswrapper[4824]: I1211 02:23:57.552390 4824 scope.go:117] "RemoveContainer" containerID="4ed05f464df1d6511a61ef661dce449fda7495a35fa8f436707fada58aaee475" Dec 11 02:23:57 crc kubenswrapper[4824]: E1211 02:23:57.553121 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4ed05f464df1d6511a61ef661dce449fda7495a35fa8f436707fada58aaee475\": container with ID starting with 4ed05f464df1d6511a61ef661dce449fda7495a35fa8f436707fada58aaee475 not found: ID does not exist" containerID="4ed05f464df1d6511a61ef661dce449fda7495a35fa8f436707fada58aaee475" Dec 11 02:23:57 crc kubenswrapper[4824]: I1211 02:23:57.553191 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ed05f464df1d6511a61ef661dce449fda7495a35fa8f436707fada58aaee475"} err="failed to get container status \"4ed05f464df1d6511a61ef661dce449fda7495a35fa8f436707fada58aaee475\": rpc error: code = NotFound desc = could not find container \"4ed05f464df1d6511a61ef661dce449fda7495a35fa8f436707fada58aaee475\": container with ID starting with 4ed05f464df1d6511a61ef661dce449fda7495a35fa8f436707fada58aaee475 not found: ID does not exist" Dec 11 02:23:57 crc kubenswrapper[4824]: I1211 02:23:57.553222 4824 scope.go:117] "RemoveContainer" containerID="1115cb51cefd19b75e6688302803085b6377a4ad2a750518905c401166fd9d32" Dec 11 02:23:57 crc kubenswrapper[4824]: E1211 02:23:57.553947 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1115cb51cefd19b75e6688302803085b6377a4ad2a750518905c401166fd9d32\": container with ID starting with 1115cb51cefd19b75e6688302803085b6377a4ad2a750518905c401166fd9d32 not found: ID does not exist" containerID="1115cb51cefd19b75e6688302803085b6377a4ad2a750518905c401166fd9d32" Dec 11 02:23:57 crc kubenswrapper[4824]: I1211 02:23:57.553971 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1115cb51cefd19b75e6688302803085b6377a4ad2a750518905c401166fd9d32"} err="failed to get container status \"1115cb51cefd19b75e6688302803085b6377a4ad2a750518905c401166fd9d32\": rpc error: code = NotFound desc = could not find container \"1115cb51cefd19b75e6688302803085b6377a4ad2a750518905c401166fd9d32\": container with ID starting with 1115cb51cefd19b75e6688302803085b6377a4ad2a750518905c401166fd9d32 not found: ID does not exist" Dec 11 02:23:57 crc kubenswrapper[4824]: I1211 02:23:57.570377 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 11 02:23:57 crc kubenswrapper[4824]: I1211 02:23:57.574563 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 11 02:23:57 crc kubenswrapper[4824]: I1211 02:23:57.576343 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 11 02:23:57 crc kubenswrapper[4824]: I1211 02:23:57.586451 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 11 02:23:57 crc kubenswrapper[4824]: I1211 02:23:57.598109 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 11 02:23:57 crc kubenswrapper[4824]: I1211 02:23:57.658199 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-62jnx\" (UniqueName: \"kubernetes.io/projected/cec2d4af-b251-47cf-aa71-b63852b09cce-kube-api-access-62jnx\") pod \"nova-scheduler-0\" (UID: \"cec2d4af-b251-47cf-aa71-b63852b09cce\") " pod="openstack/nova-scheduler-0" Dec 11 02:23:57 crc kubenswrapper[4824]: I1211 02:23:57.658932 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cec2d4af-b251-47cf-aa71-b63852b09cce-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"cec2d4af-b251-47cf-aa71-b63852b09cce\") " pod="openstack/nova-scheduler-0" Dec 11 02:23:57 crc kubenswrapper[4824]: I1211 02:23:57.659054 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cec2d4af-b251-47cf-aa71-b63852b09cce-config-data\") pod \"nova-scheduler-0\" (UID: \"cec2d4af-b251-47cf-aa71-b63852b09cce\") " pod="openstack/nova-scheduler-0" Dec 11 02:23:57 crc kubenswrapper[4824]: I1211 02:23:57.760417 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-62jnx\" (UniqueName: \"kubernetes.io/projected/cec2d4af-b251-47cf-aa71-b63852b09cce-kube-api-access-62jnx\") pod \"nova-scheduler-0\" (UID: \"cec2d4af-b251-47cf-aa71-b63852b09cce\") " pod="openstack/nova-scheduler-0" Dec 11 02:23:57 crc kubenswrapper[4824]: I1211 02:23:57.760480 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ds6ql\" (UniqueName: \"kubernetes.io/projected/420e8783-43b4-43c7-ae74-50df3b388af1-kube-api-access-ds6ql\") pod \"nova-api-0\" (UID: \"420e8783-43b4-43c7-ae74-50df3b388af1\") " pod="openstack/nova-api-0" Dec 11 02:23:57 crc kubenswrapper[4824]: I1211 02:23:57.760575 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/420e8783-43b4-43c7-ae74-50df3b388af1-logs\") pod \"nova-api-0\" (UID: \"420e8783-43b4-43c7-ae74-50df3b388af1\") " pod="openstack/nova-api-0" Dec 11 02:23:57 crc kubenswrapper[4824]: I1211 02:23:57.760606 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/420e8783-43b4-43c7-ae74-50df3b388af1-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"420e8783-43b4-43c7-ae74-50df3b388af1\") " pod="openstack/nova-api-0" Dec 11 02:23:57 crc kubenswrapper[4824]: I1211 02:23:57.760635 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cec2d4af-b251-47cf-aa71-b63852b09cce-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"cec2d4af-b251-47cf-aa71-b63852b09cce\") " pod="openstack/nova-scheduler-0" Dec 11 02:23:57 crc kubenswrapper[4824]: I1211 02:23:57.760742 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cec2d4af-b251-47cf-aa71-b63852b09cce-config-data\") pod \"nova-scheduler-0\" (UID: \"cec2d4af-b251-47cf-aa71-b63852b09cce\") " pod="openstack/nova-scheduler-0" Dec 11 02:23:57 crc kubenswrapper[4824]: I1211 02:23:57.760810 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/420e8783-43b4-43c7-ae74-50df3b388af1-config-data\") pod \"nova-api-0\" (UID: \"420e8783-43b4-43c7-ae74-50df3b388af1\") " pod="openstack/nova-api-0" Dec 11 02:23:57 crc kubenswrapper[4824]: I1211 02:23:57.771647 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cec2d4af-b251-47cf-aa71-b63852b09cce-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"cec2d4af-b251-47cf-aa71-b63852b09cce\") " pod="openstack/nova-scheduler-0" Dec 11 02:23:57 crc kubenswrapper[4824]: I1211 02:23:57.773997 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cec2d4af-b251-47cf-aa71-b63852b09cce-config-data\") pod \"nova-scheduler-0\" (UID: \"cec2d4af-b251-47cf-aa71-b63852b09cce\") " pod="openstack/nova-scheduler-0" Dec 11 02:23:57 crc kubenswrapper[4824]: I1211 02:23:57.778301 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-62jnx\" (UniqueName: \"kubernetes.io/projected/cec2d4af-b251-47cf-aa71-b63852b09cce-kube-api-access-62jnx\") pod \"nova-scheduler-0\" (UID: \"cec2d4af-b251-47cf-aa71-b63852b09cce\") " pod="openstack/nova-scheduler-0" Dec 11 02:23:57 crc kubenswrapper[4824]: I1211 02:23:57.805022 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 11 02:23:57 crc kubenswrapper[4824]: I1211 02:23:57.838924 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 11 02:23:57 crc kubenswrapper[4824]: I1211 02:23:57.862166 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/420e8783-43b4-43c7-ae74-50df3b388af1-config-data\") pod \"nova-api-0\" (UID: \"420e8783-43b4-43c7-ae74-50df3b388af1\") " pod="openstack/nova-api-0" Dec 11 02:23:57 crc kubenswrapper[4824]: I1211 02:23:57.862276 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ds6ql\" (UniqueName: \"kubernetes.io/projected/420e8783-43b4-43c7-ae74-50df3b388af1-kube-api-access-ds6ql\") pod \"nova-api-0\" (UID: \"420e8783-43b4-43c7-ae74-50df3b388af1\") " pod="openstack/nova-api-0" Dec 11 02:23:57 crc kubenswrapper[4824]: I1211 02:23:57.862352 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/420e8783-43b4-43c7-ae74-50df3b388af1-logs\") pod \"nova-api-0\" (UID: \"420e8783-43b4-43c7-ae74-50df3b388af1\") " pod="openstack/nova-api-0" Dec 11 02:23:57 crc kubenswrapper[4824]: I1211 02:23:57.862376 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/420e8783-43b4-43c7-ae74-50df3b388af1-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"420e8783-43b4-43c7-ae74-50df3b388af1\") " pod="openstack/nova-api-0" Dec 11 02:23:57 crc kubenswrapper[4824]: I1211 02:23:57.863583 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/420e8783-43b4-43c7-ae74-50df3b388af1-logs\") pod \"nova-api-0\" (UID: \"420e8783-43b4-43c7-ae74-50df3b388af1\") " pod="openstack/nova-api-0" Dec 11 02:23:57 crc kubenswrapper[4824]: I1211 02:23:57.866950 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/420e8783-43b4-43c7-ae74-50df3b388af1-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"420e8783-43b4-43c7-ae74-50df3b388af1\") " pod="openstack/nova-api-0" Dec 11 02:23:57 crc kubenswrapper[4824]: I1211 02:23:57.867843 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/420e8783-43b4-43c7-ae74-50df3b388af1-config-data\") pod \"nova-api-0\" (UID: \"420e8783-43b4-43c7-ae74-50df3b388af1\") " pod="openstack/nova-api-0" Dec 11 02:23:57 crc kubenswrapper[4824]: I1211 02:23:57.881118 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ds6ql\" (UniqueName: \"kubernetes.io/projected/420e8783-43b4-43c7-ae74-50df3b388af1-kube-api-access-ds6ql\") pod \"nova-api-0\" (UID: \"420e8783-43b4-43c7-ae74-50df3b388af1\") " pod="openstack/nova-api-0" Dec 11 02:23:57 crc kubenswrapper[4824]: I1211 02:23:57.894666 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 11 02:23:58 crc kubenswrapper[4824]: I1211 02:23:58.313833 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 11 02:23:58 crc kubenswrapper[4824]: W1211 02:23:58.318963 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcec2d4af_b251_47cf_aa71_b63852b09cce.slice/crio-7ab6590841d97d01d5b7371461fcc8ab0c9361f245ae70c1ea2f51dbb701e579 WatchSource:0}: Error finding container 7ab6590841d97d01d5b7371461fcc8ab0c9361f245ae70c1ea2f51dbb701e579: Status 404 returned error can't find the container with id 7ab6590841d97d01d5b7371461fcc8ab0c9361f245ae70c1ea2f51dbb701e579 Dec 11 02:23:58 crc kubenswrapper[4824]: I1211 02:23:58.413873 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 11 02:23:58 crc kubenswrapper[4824]: W1211 02:23:58.418358 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod420e8783_43b4_43c7_ae74_50df3b388af1.slice/crio-3b7fd39b5617935f8edcdd5fffb3357f8ac1b85ed77435ddcac6e7bc81b9bb14 WatchSource:0}: Error finding container 3b7fd39b5617935f8edcdd5fffb3357f8ac1b85ed77435ddcac6e7bc81b9bb14: Status 404 returned error can't find the container with id 3b7fd39b5617935f8edcdd5fffb3357f8ac1b85ed77435ddcac6e7bc81b9bb14 Dec 11 02:23:58 crc kubenswrapper[4824]: I1211 02:23:58.656912 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3f62926-29c4-4fd2-8e1b-926a53823b65" path="/var/lib/kubelet/pods/d3f62926-29c4-4fd2-8e1b-926a53823b65/volumes" Dec 11 02:23:58 crc kubenswrapper[4824]: I1211 02:23:58.657980 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7af70e6-c4c5-477f-a5af-876490e4ddf7" path="/var/lib/kubelet/pods/f7af70e6-c4c5-477f-a5af-876490e4ddf7/volumes" Dec 11 02:23:58 crc kubenswrapper[4824]: I1211 02:23:58.999811 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"420e8783-43b4-43c7-ae74-50df3b388af1","Type":"ContainerStarted","Data":"9b98e7345115867a4832021da7738c9d8e895fc4f81273785a40c60ab44ac9d1"} Dec 11 02:23:58 crc kubenswrapper[4824]: I1211 02:23:58.999855 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"420e8783-43b4-43c7-ae74-50df3b388af1","Type":"ContainerStarted","Data":"3b7fd39b5617935f8edcdd5fffb3357f8ac1b85ed77435ddcac6e7bc81b9bb14"} Dec 11 02:23:59 crc kubenswrapper[4824]: I1211 02:23:59.000997 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"cec2d4af-b251-47cf-aa71-b63852b09cce","Type":"ContainerStarted","Data":"f566f44056ef249cb33952ee8373c0ffc800a9261b52edb6c6987191226aed85"} Dec 11 02:23:59 crc kubenswrapper[4824]: I1211 02:23:59.001022 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"cec2d4af-b251-47cf-aa71-b63852b09cce","Type":"ContainerStarted","Data":"7ab6590841d97d01d5b7371461fcc8ab0c9361f245ae70c1ea2f51dbb701e579"} Dec 11 02:23:59 crc kubenswrapper[4824]: I1211 02:23:59.040031 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.040015056 podStartE2EDuration="2.040015056s" podCreationTimestamp="2025-12-11 02:23:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:23:59.032794325 +0000 UTC m=+1380.721831704" watchObservedRunningTime="2025-12-11 02:23:59.040015056 +0000 UTC m=+1380.729052435" Dec 11 02:23:59 crc kubenswrapper[4824]: I1211 02:23:59.583181 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-5hrpm"] Dec 11 02:23:59 crc kubenswrapper[4824]: I1211 02:23:59.596723 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5hrpm" Dec 11 02:23:59 crc kubenswrapper[4824]: I1211 02:23:59.607831 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/171e204b-8e23-4f40-b666-ba76cda27715-catalog-content\") pod \"redhat-marketplace-5hrpm\" (UID: \"171e204b-8e23-4f40-b666-ba76cda27715\") " pod="openshift-marketplace/redhat-marketplace-5hrpm" Dec 11 02:23:59 crc kubenswrapper[4824]: I1211 02:23:59.607958 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/171e204b-8e23-4f40-b666-ba76cda27715-utilities\") pod \"redhat-marketplace-5hrpm\" (UID: \"171e204b-8e23-4f40-b666-ba76cda27715\") " pod="openshift-marketplace/redhat-marketplace-5hrpm" Dec 11 02:23:59 crc kubenswrapper[4824]: I1211 02:23:59.608063 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vr2f8\" (UniqueName: \"kubernetes.io/projected/171e204b-8e23-4f40-b666-ba76cda27715-kube-api-access-vr2f8\") pod \"redhat-marketplace-5hrpm\" (UID: \"171e204b-8e23-4f40-b666-ba76cda27715\") " pod="openshift-marketplace/redhat-marketplace-5hrpm" Dec 11 02:23:59 crc kubenswrapper[4824]: I1211 02:23:59.625837 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5hrpm"] Dec 11 02:23:59 crc kubenswrapper[4824]: I1211 02:23:59.709793 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/171e204b-8e23-4f40-b666-ba76cda27715-utilities\") pod \"redhat-marketplace-5hrpm\" (UID: \"171e204b-8e23-4f40-b666-ba76cda27715\") " pod="openshift-marketplace/redhat-marketplace-5hrpm" Dec 11 02:23:59 crc kubenswrapper[4824]: I1211 02:23:59.709925 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vr2f8\" (UniqueName: \"kubernetes.io/projected/171e204b-8e23-4f40-b666-ba76cda27715-kube-api-access-vr2f8\") pod \"redhat-marketplace-5hrpm\" (UID: \"171e204b-8e23-4f40-b666-ba76cda27715\") " pod="openshift-marketplace/redhat-marketplace-5hrpm" Dec 11 02:23:59 crc kubenswrapper[4824]: I1211 02:23:59.710041 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/171e204b-8e23-4f40-b666-ba76cda27715-catalog-content\") pod \"redhat-marketplace-5hrpm\" (UID: \"171e204b-8e23-4f40-b666-ba76cda27715\") " pod="openshift-marketplace/redhat-marketplace-5hrpm" Dec 11 02:23:59 crc kubenswrapper[4824]: I1211 02:23:59.711379 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/171e204b-8e23-4f40-b666-ba76cda27715-utilities\") pod \"redhat-marketplace-5hrpm\" (UID: \"171e204b-8e23-4f40-b666-ba76cda27715\") " pod="openshift-marketplace/redhat-marketplace-5hrpm" Dec 11 02:23:59 crc kubenswrapper[4824]: I1211 02:23:59.711762 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/171e204b-8e23-4f40-b666-ba76cda27715-catalog-content\") pod \"redhat-marketplace-5hrpm\" (UID: \"171e204b-8e23-4f40-b666-ba76cda27715\") " pod="openshift-marketplace/redhat-marketplace-5hrpm" Dec 11 02:23:59 crc kubenswrapper[4824]: I1211 02:23:59.733156 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vr2f8\" (UniqueName: \"kubernetes.io/projected/171e204b-8e23-4f40-b666-ba76cda27715-kube-api-access-vr2f8\") pod \"redhat-marketplace-5hrpm\" (UID: \"171e204b-8e23-4f40-b666-ba76cda27715\") " pod="openshift-marketplace/redhat-marketplace-5hrpm" Dec 11 02:23:59 crc kubenswrapper[4824]: I1211 02:23:59.926761 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5hrpm" Dec 11 02:24:00 crc kubenswrapper[4824]: I1211 02:24:00.017437 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"420e8783-43b4-43c7-ae74-50df3b388af1","Type":"ContainerStarted","Data":"1809cc49214843841f512f2b18ee339fa99a9a75e7d0725b1b98e8fdb65214ef"} Dec 11 02:24:00 crc kubenswrapper[4824]: I1211 02:24:00.038742 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.038720932 podStartE2EDuration="3.038720932s" podCreationTimestamp="2025-12-11 02:23:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:24:00.035245425 +0000 UTC m=+1381.724282804" watchObservedRunningTime="2025-12-11 02:24:00.038720932 +0000 UTC m=+1381.727758311" Dec 11 02:24:00 crc kubenswrapper[4824]: I1211 02:24:00.435066 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5hrpm"] Dec 11 02:24:01 crc kubenswrapper[4824]: I1211 02:24:01.027550 4824 generic.go:334] "Generic (PLEG): container finished" podID="171e204b-8e23-4f40-b666-ba76cda27715" containerID="04e078ecc7483ffe84663fa4aaf2c3aabb21d8070a637c15e9c8438d289375ab" exitCode=0 Dec 11 02:24:01 crc kubenswrapper[4824]: I1211 02:24:01.027601 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5hrpm" event={"ID":"171e204b-8e23-4f40-b666-ba76cda27715","Type":"ContainerDied","Data":"04e078ecc7483ffe84663fa4aaf2c3aabb21d8070a637c15e9c8438d289375ab"} Dec 11 02:24:01 crc kubenswrapper[4824]: I1211 02:24:01.027884 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5hrpm" event={"ID":"171e204b-8e23-4f40-b666-ba76cda27715","Type":"ContainerStarted","Data":"cc963f0ebf4301760bfb4fcec958dabdc15208ff5fdf946474b5d4ca849c4ac5"} Dec 11 02:24:01 crc kubenswrapper[4824]: I1211 02:24:01.899795 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 11 02:24:01 crc kubenswrapper[4824]: I1211 02:24:01.901166 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="5b375617-b0a6-42cf-be23-ef7f755f9eff" containerName="kube-state-metrics" containerID="cri-o://552c5666299888465ed8c969e420b63ce34abef304f83696d7c2337684b297b3" gracePeriod=30 Dec 11 02:24:02 crc kubenswrapper[4824]: I1211 02:24:02.045747 4824 generic.go:334] "Generic (PLEG): container finished" podID="5b375617-b0a6-42cf-be23-ef7f755f9eff" containerID="552c5666299888465ed8c969e420b63ce34abef304f83696d7c2337684b297b3" exitCode=2 Dec 11 02:24:02 crc kubenswrapper[4824]: I1211 02:24:02.047302 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"5b375617-b0a6-42cf-be23-ef7f755f9eff","Type":"ContainerDied","Data":"552c5666299888465ed8c969e420b63ce34abef304f83696d7c2337684b297b3"} Dec 11 02:24:02 crc kubenswrapper[4824]: E1211 02:24:02.125250 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5b375617_b0a6_42cf_be23_ef7f755f9eff.slice/crio-conmon-552c5666299888465ed8c969e420b63ce34abef304f83696d7c2337684b297b3.scope\": RecentStats: unable to find data in memory cache]" Dec 11 02:24:02 crc kubenswrapper[4824]: I1211 02:24:02.322816 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 11 02:24:02 crc kubenswrapper[4824]: I1211 02:24:02.323045 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 11 02:24:02 crc kubenswrapper[4824]: I1211 02:24:02.486026 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 11 02:24:02 crc kubenswrapper[4824]: I1211 02:24:02.586312 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mk6xs\" (UniqueName: \"kubernetes.io/projected/5b375617-b0a6-42cf-be23-ef7f755f9eff-kube-api-access-mk6xs\") pod \"5b375617-b0a6-42cf-be23-ef7f755f9eff\" (UID: \"5b375617-b0a6-42cf-be23-ef7f755f9eff\") " Dec 11 02:24:02 crc kubenswrapper[4824]: I1211 02:24:02.592100 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b375617-b0a6-42cf-be23-ef7f755f9eff-kube-api-access-mk6xs" (OuterVolumeSpecName: "kube-api-access-mk6xs") pod "5b375617-b0a6-42cf-be23-ef7f755f9eff" (UID: "5b375617-b0a6-42cf-be23-ef7f755f9eff"). InnerVolumeSpecName "kube-api-access-mk6xs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:24:02 crc kubenswrapper[4824]: I1211 02:24:02.689921 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mk6xs\" (UniqueName: \"kubernetes.io/projected/5b375617-b0a6-42cf-be23-ef7f755f9eff-kube-api-access-mk6xs\") on node \"crc\" DevicePath \"\"" Dec 11 02:24:02 crc kubenswrapper[4824]: I1211 02:24:02.839412 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 11 02:24:03 crc kubenswrapper[4824]: I1211 02:24:03.061235 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"5b375617-b0a6-42cf-be23-ef7f755f9eff","Type":"ContainerDied","Data":"07ee1898b339c95b30e9f829c3cb73411381c4743aba8b18ad5e7b8d5c5a3aa6"} Dec 11 02:24:03 crc kubenswrapper[4824]: I1211 02:24:03.061631 4824 scope.go:117] "RemoveContainer" containerID="552c5666299888465ed8c969e420b63ce34abef304f83696d7c2337684b297b3" Dec 11 02:24:03 crc kubenswrapper[4824]: I1211 02:24:03.061472 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 11 02:24:03 crc kubenswrapper[4824]: I1211 02:24:03.067732 4824 generic.go:334] "Generic (PLEG): container finished" podID="171e204b-8e23-4f40-b666-ba76cda27715" containerID="acf0049ba571845043ca3de50fcab3256cf2495fb1fd0003f0cce9ffec3d1666" exitCode=0 Dec 11 02:24:03 crc kubenswrapper[4824]: I1211 02:24:03.067783 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5hrpm" event={"ID":"171e204b-8e23-4f40-b666-ba76cda27715","Type":"ContainerDied","Data":"acf0049ba571845043ca3de50fcab3256cf2495fb1fd0003f0cce9ffec3d1666"} Dec 11 02:24:03 crc kubenswrapper[4824]: I1211 02:24:03.101787 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 11 02:24:03 crc kubenswrapper[4824]: I1211 02:24:03.115531 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 11 02:24:03 crc kubenswrapper[4824]: I1211 02:24:03.125791 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 11 02:24:03 crc kubenswrapper[4824]: E1211 02:24:03.126322 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b375617-b0a6-42cf-be23-ef7f755f9eff" containerName="kube-state-metrics" Dec 11 02:24:03 crc kubenswrapper[4824]: I1211 02:24:03.126343 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b375617-b0a6-42cf-be23-ef7f755f9eff" containerName="kube-state-metrics" Dec 11 02:24:03 crc kubenswrapper[4824]: I1211 02:24:03.126535 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b375617-b0a6-42cf-be23-ef7f755f9eff" containerName="kube-state-metrics" Dec 11 02:24:03 crc kubenswrapper[4824]: I1211 02:24:03.127213 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 11 02:24:03 crc kubenswrapper[4824]: I1211 02:24:03.129935 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Dec 11 02:24:03 crc kubenswrapper[4824]: I1211 02:24:03.136649 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Dec 11 02:24:03 crc kubenswrapper[4824]: I1211 02:24:03.141574 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 11 02:24:03 crc kubenswrapper[4824]: I1211 02:24:03.200385 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/1ed2f6d5-a61f-4c5b-9b73-1089d13cf9b6-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"1ed2f6d5-a61f-4c5b-9b73-1089d13cf9b6\") " pod="openstack/kube-state-metrics-0" Dec 11 02:24:03 crc kubenswrapper[4824]: I1211 02:24:03.200464 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ed2f6d5-a61f-4c5b-9b73-1089d13cf9b6-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"1ed2f6d5-a61f-4c5b-9b73-1089d13cf9b6\") " pod="openstack/kube-state-metrics-0" Dec 11 02:24:03 crc kubenswrapper[4824]: I1211 02:24:03.200829 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5s4vx\" (UniqueName: \"kubernetes.io/projected/1ed2f6d5-a61f-4c5b-9b73-1089d13cf9b6-kube-api-access-5s4vx\") pod \"kube-state-metrics-0\" (UID: \"1ed2f6d5-a61f-4c5b-9b73-1089d13cf9b6\") " pod="openstack/kube-state-metrics-0" Dec 11 02:24:03 crc kubenswrapper[4824]: I1211 02:24:03.200962 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ed2f6d5-a61f-4c5b-9b73-1089d13cf9b6-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"1ed2f6d5-a61f-4c5b-9b73-1089d13cf9b6\") " pod="openstack/kube-state-metrics-0" Dec 11 02:24:03 crc kubenswrapper[4824]: I1211 02:24:03.303251 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/1ed2f6d5-a61f-4c5b-9b73-1089d13cf9b6-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"1ed2f6d5-a61f-4c5b-9b73-1089d13cf9b6\") " pod="openstack/kube-state-metrics-0" Dec 11 02:24:03 crc kubenswrapper[4824]: I1211 02:24:03.303302 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ed2f6d5-a61f-4c5b-9b73-1089d13cf9b6-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"1ed2f6d5-a61f-4c5b-9b73-1089d13cf9b6\") " pod="openstack/kube-state-metrics-0" Dec 11 02:24:03 crc kubenswrapper[4824]: I1211 02:24:03.303378 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5s4vx\" (UniqueName: \"kubernetes.io/projected/1ed2f6d5-a61f-4c5b-9b73-1089d13cf9b6-kube-api-access-5s4vx\") pod \"kube-state-metrics-0\" (UID: \"1ed2f6d5-a61f-4c5b-9b73-1089d13cf9b6\") " pod="openstack/kube-state-metrics-0" Dec 11 02:24:03 crc kubenswrapper[4824]: I1211 02:24:03.303417 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ed2f6d5-a61f-4c5b-9b73-1089d13cf9b6-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"1ed2f6d5-a61f-4c5b-9b73-1089d13cf9b6\") " pod="openstack/kube-state-metrics-0" Dec 11 02:24:03 crc kubenswrapper[4824]: I1211 02:24:03.308341 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ed2f6d5-a61f-4c5b-9b73-1089d13cf9b6-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"1ed2f6d5-a61f-4c5b-9b73-1089d13cf9b6\") " pod="openstack/kube-state-metrics-0" Dec 11 02:24:03 crc kubenswrapper[4824]: I1211 02:24:03.308603 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/1ed2f6d5-a61f-4c5b-9b73-1089d13cf9b6-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"1ed2f6d5-a61f-4c5b-9b73-1089d13cf9b6\") " pod="openstack/kube-state-metrics-0" Dec 11 02:24:03 crc kubenswrapper[4824]: I1211 02:24:03.316198 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ed2f6d5-a61f-4c5b-9b73-1089d13cf9b6-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"1ed2f6d5-a61f-4c5b-9b73-1089d13cf9b6\") " pod="openstack/kube-state-metrics-0" Dec 11 02:24:03 crc kubenswrapper[4824]: I1211 02:24:03.325193 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5s4vx\" (UniqueName: \"kubernetes.io/projected/1ed2f6d5-a61f-4c5b-9b73-1089d13cf9b6-kube-api-access-5s4vx\") pod \"kube-state-metrics-0\" (UID: \"1ed2f6d5-a61f-4c5b-9b73-1089d13cf9b6\") " pod="openstack/kube-state-metrics-0" Dec 11 02:24:03 crc kubenswrapper[4824]: I1211 02:24:03.370265 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="2a3020fe-d2be-4c32-9f69-f5365c4a10c4" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.194:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 11 02:24:03 crc kubenswrapper[4824]: I1211 02:24:03.370319 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="2a3020fe-d2be-4c32-9f69-f5365c4a10c4" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.194:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 11 02:24:03 crc kubenswrapper[4824]: I1211 02:24:03.450152 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 11 02:24:03 crc kubenswrapper[4824]: I1211 02:24:03.934045 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 11 02:24:04 crc kubenswrapper[4824]: I1211 02:24:04.077964 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"1ed2f6d5-a61f-4c5b-9b73-1089d13cf9b6","Type":"ContainerStarted","Data":"4200a82dc88a9b76b0ca2d2bf35c4e67004dae959312665fbf8f14bd50ee3343"} Dec 11 02:24:04 crc kubenswrapper[4824]: I1211 02:24:04.368390 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 11 02:24:04 crc kubenswrapper[4824]: I1211 02:24:04.368910 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c4182e68-340b-43ce-b458-0cac6fa9f7fd" containerName="ceilometer-central-agent" containerID="cri-o://1f88f5ae1bf5cbbcf93220c49f7f08aa3db2f1f5e7c46ed76b5a5919ea60fd63" gracePeriod=30 Dec 11 02:24:04 crc kubenswrapper[4824]: I1211 02:24:04.369389 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c4182e68-340b-43ce-b458-0cac6fa9f7fd" containerName="proxy-httpd" containerID="cri-o://381b142df922b142dcfbebe34bef199920c42f16e2dba9b9cc688f157ba4e937" gracePeriod=30 Dec 11 02:24:04 crc kubenswrapper[4824]: I1211 02:24:04.369442 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c4182e68-340b-43ce-b458-0cac6fa9f7fd" containerName="sg-core" containerID="cri-o://16899e5bc595474bd65445e7d16197f13d56f2ae311171be359b4486c19514f6" gracePeriod=30 Dec 11 02:24:04 crc kubenswrapper[4824]: I1211 02:24:04.369476 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c4182e68-340b-43ce-b458-0cac6fa9f7fd" containerName="ceilometer-notification-agent" containerID="cri-o://ae6988e689b7a435f7dff39aa726ebdb43772d5aad1cd52cf4fb6b62c24ae678" gracePeriod=30 Dec 11 02:24:04 crc kubenswrapper[4824]: I1211 02:24:04.641903 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b375617-b0a6-42cf-be23-ef7f755f9eff" path="/var/lib/kubelet/pods/5b375617-b0a6-42cf-be23-ef7f755f9eff/volumes" Dec 11 02:24:05 crc kubenswrapper[4824]: I1211 02:24:05.092219 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5hrpm" event={"ID":"171e204b-8e23-4f40-b666-ba76cda27715","Type":"ContainerStarted","Data":"992a079394d9659cb31e3d70b0ca545d3f15717620bdc3290217c904935e80f0"} Dec 11 02:24:05 crc kubenswrapper[4824]: I1211 02:24:05.093976 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"1ed2f6d5-a61f-4c5b-9b73-1089d13cf9b6","Type":"ContainerStarted","Data":"f786e40b60542871d3b81a0be245709f7b4a40d88c976fe17b372e6cd16eba1c"} Dec 11 02:24:05 crc kubenswrapper[4824]: I1211 02:24:05.094179 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 11 02:24:05 crc kubenswrapper[4824]: I1211 02:24:05.098033 4824 generic.go:334] "Generic (PLEG): container finished" podID="c4182e68-340b-43ce-b458-0cac6fa9f7fd" containerID="381b142df922b142dcfbebe34bef199920c42f16e2dba9b9cc688f157ba4e937" exitCode=0 Dec 11 02:24:05 crc kubenswrapper[4824]: I1211 02:24:05.098077 4824 generic.go:334] "Generic (PLEG): container finished" podID="c4182e68-340b-43ce-b458-0cac6fa9f7fd" containerID="16899e5bc595474bd65445e7d16197f13d56f2ae311171be359b4486c19514f6" exitCode=2 Dec 11 02:24:05 crc kubenswrapper[4824]: I1211 02:24:05.098084 4824 generic.go:334] "Generic (PLEG): container finished" podID="c4182e68-340b-43ce-b458-0cac6fa9f7fd" containerID="1f88f5ae1bf5cbbcf93220c49f7f08aa3db2f1f5e7c46ed76b5a5919ea60fd63" exitCode=0 Dec 11 02:24:05 crc kubenswrapper[4824]: I1211 02:24:05.098103 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c4182e68-340b-43ce-b458-0cac6fa9f7fd","Type":"ContainerDied","Data":"381b142df922b142dcfbebe34bef199920c42f16e2dba9b9cc688f157ba4e937"} Dec 11 02:24:05 crc kubenswrapper[4824]: I1211 02:24:05.098167 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c4182e68-340b-43ce-b458-0cac6fa9f7fd","Type":"ContainerDied","Data":"16899e5bc595474bd65445e7d16197f13d56f2ae311171be359b4486c19514f6"} Dec 11 02:24:05 crc kubenswrapper[4824]: I1211 02:24:05.098176 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c4182e68-340b-43ce-b458-0cac6fa9f7fd","Type":"ContainerDied","Data":"1f88f5ae1bf5cbbcf93220c49f7f08aa3db2f1f5e7c46ed76b5a5919ea60fd63"} Dec 11 02:24:05 crc kubenswrapper[4824]: I1211 02:24:05.116988 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-5hrpm" podStartSLOduration=3.040520843 podStartE2EDuration="6.116973845s" podCreationTimestamp="2025-12-11 02:23:59 +0000 UTC" firstStartedPulling="2025-12-11 02:24:01.029893988 +0000 UTC m=+1382.718931367" lastFinishedPulling="2025-12-11 02:24:04.10634699 +0000 UTC m=+1385.795384369" observedRunningTime="2025-12-11 02:24:05.114284048 +0000 UTC m=+1386.803321417" watchObservedRunningTime="2025-12-11 02:24:05.116973845 +0000 UTC m=+1386.806011224" Dec 11 02:24:05 crc kubenswrapper[4824]: I1211 02:24:05.134164 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=1.742319588 podStartE2EDuration="2.134149796s" podCreationTimestamp="2025-12-11 02:24:03 +0000 UTC" firstStartedPulling="2025-12-11 02:24:03.937724887 +0000 UTC m=+1385.626762266" lastFinishedPulling="2025-12-11 02:24:04.329555085 +0000 UTC m=+1386.018592474" observedRunningTime="2025-12-11 02:24:05.131057979 +0000 UTC m=+1386.820095358" watchObservedRunningTime="2025-12-11 02:24:05.134149796 +0000 UTC m=+1386.823187175" Dec 11 02:24:07 crc kubenswrapper[4824]: I1211 02:24:07.132060 4824 generic.go:334] "Generic (PLEG): container finished" podID="c4182e68-340b-43ce-b458-0cac6fa9f7fd" containerID="ae6988e689b7a435f7dff39aa726ebdb43772d5aad1cd52cf4fb6b62c24ae678" exitCode=0 Dec 11 02:24:07 crc kubenswrapper[4824]: I1211 02:24:07.132100 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c4182e68-340b-43ce-b458-0cac6fa9f7fd","Type":"ContainerDied","Data":"ae6988e689b7a435f7dff39aa726ebdb43772d5aad1cd52cf4fb6b62c24ae678"} Dec 11 02:24:07 crc kubenswrapper[4824]: I1211 02:24:07.132678 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c4182e68-340b-43ce-b458-0cac6fa9f7fd","Type":"ContainerDied","Data":"5cf0bae8996e79eb913896800a52c69909d05bf2273733a60573cb0c3cac9b46"} Dec 11 02:24:07 crc kubenswrapper[4824]: I1211 02:24:07.132698 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5cf0bae8996e79eb913896800a52c69909d05bf2273733a60573cb0c3cac9b46" Dec 11 02:24:07 crc kubenswrapper[4824]: I1211 02:24:07.166893 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 11 02:24:07 crc kubenswrapper[4824]: I1211 02:24:07.174513 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4182e68-340b-43ce-b458-0cac6fa9f7fd-config-data\") pod \"c4182e68-340b-43ce-b458-0cac6fa9f7fd\" (UID: \"c4182e68-340b-43ce-b458-0cac6fa9f7fd\") " Dec 11 02:24:07 crc kubenswrapper[4824]: I1211 02:24:07.174586 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c4182e68-340b-43ce-b458-0cac6fa9f7fd-run-httpd\") pod \"c4182e68-340b-43ce-b458-0cac6fa9f7fd\" (UID: \"c4182e68-340b-43ce-b458-0cac6fa9f7fd\") " Dec 11 02:24:07 crc kubenswrapper[4824]: I1211 02:24:07.174618 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c4182e68-340b-43ce-b458-0cac6fa9f7fd-sg-core-conf-yaml\") pod \"c4182e68-340b-43ce-b458-0cac6fa9f7fd\" (UID: \"c4182e68-340b-43ce-b458-0cac6fa9f7fd\") " Dec 11 02:24:07 crc kubenswrapper[4824]: I1211 02:24:07.174663 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c4182e68-340b-43ce-b458-0cac6fa9f7fd-scripts\") pod \"c4182e68-340b-43ce-b458-0cac6fa9f7fd\" (UID: \"c4182e68-340b-43ce-b458-0cac6fa9f7fd\") " Dec 11 02:24:07 crc kubenswrapper[4824]: I1211 02:24:07.174861 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2hfrt\" (UniqueName: \"kubernetes.io/projected/c4182e68-340b-43ce-b458-0cac6fa9f7fd-kube-api-access-2hfrt\") pod \"c4182e68-340b-43ce-b458-0cac6fa9f7fd\" (UID: \"c4182e68-340b-43ce-b458-0cac6fa9f7fd\") " Dec 11 02:24:07 crc kubenswrapper[4824]: I1211 02:24:07.174912 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4182e68-340b-43ce-b458-0cac6fa9f7fd-combined-ca-bundle\") pod \"c4182e68-340b-43ce-b458-0cac6fa9f7fd\" (UID: \"c4182e68-340b-43ce-b458-0cac6fa9f7fd\") " Dec 11 02:24:07 crc kubenswrapper[4824]: I1211 02:24:07.174943 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c4182e68-340b-43ce-b458-0cac6fa9f7fd-log-httpd\") pod \"c4182e68-340b-43ce-b458-0cac6fa9f7fd\" (UID: \"c4182e68-340b-43ce-b458-0cac6fa9f7fd\") " Dec 11 02:24:07 crc kubenswrapper[4824]: I1211 02:24:07.175048 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c4182e68-340b-43ce-b458-0cac6fa9f7fd-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "c4182e68-340b-43ce-b458-0cac6fa9f7fd" (UID: "c4182e68-340b-43ce-b458-0cac6fa9f7fd"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:24:07 crc kubenswrapper[4824]: I1211 02:24:07.175327 4824 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c4182e68-340b-43ce-b458-0cac6fa9f7fd-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 11 02:24:07 crc kubenswrapper[4824]: I1211 02:24:07.175364 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c4182e68-340b-43ce-b458-0cac6fa9f7fd-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "c4182e68-340b-43ce-b458-0cac6fa9f7fd" (UID: "c4182e68-340b-43ce-b458-0cac6fa9f7fd"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:24:07 crc kubenswrapper[4824]: I1211 02:24:07.179723 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4182e68-340b-43ce-b458-0cac6fa9f7fd-kube-api-access-2hfrt" (OuterVolumeSpecName: "kube-api-access-2hfrt") pod "c4182e68-340b-43ce-b458-0cac6fa9f7fd" (UID: "c4182e68-340b-43ce-b458-0cac6fa9f7fd"). InnerVolumeSpecName "kube-api-access-2hfrt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:24:07 crc kubenswrapper[4824]: I1211 02:24:07.181338 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4182e68-340b-43ce-b458-0cac6fa9f7fd-scripts" (OuterVolumeSpecName: "scripts") pod "c4182e68-340b-43ce-b458-0cac6fa9f7fd" (UID: "c4182e68-340b-43ce-b458-0cac6fa9f7fd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:24:07 crc kubenswrapper[4824]: I1211 02:24:07.214047 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4182e68-340b-43ce-b458-0cac6fa9f7fd-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "c4182e68-340b-43ce-b458-0cac6fa9f7fd" (UID: "c4182e68-340b-43ce-b458-0cac6fa9f7fd"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:24:07 crc kubenswrapper[4824]: I1211 02:24:07.277805 4824 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c4182e68-340b-43ce-b458-0cac6fa9f7fd-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 11 02:24:07 crc kubenswrapper[4824]: I1211 02:24:07.277832 4824 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c4182e68-340b-43ce-b458-0cac6fa9f7fd-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 11 02:24:07 crc kubenswrapper[4824]: I1211 02:24:07.277840 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c4182e68-340b-43ce-b458-0cac6fa9f7fd-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 02:24:07 crc kubenswrapper[4824]: I1211 02:24:07.277849 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2hfrt\" (UniqueName: \"kubernetes.io/projected/c4182e68-340b-43ce-b458-0cac6fa9f7fd-kube-api-access-2hfrt\") on node \"crc\" DevicePath \"\"" Dec 11 02:24:07 crc kubenswrapper[4824]: I1211 02:24:07.286199 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4182e68-340b-43ce-b458-0cac6fa9f7fd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c4182e68-340b-43ce-b458-0cac6fa9f7fd" (UID: "c4182e68-340b-43ce-b458-0cac6fa9f7fd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:24:07 crc kubenswrapper[4824]: I1211 02:24:07.286579 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4182e68-340b-43ce-b458-0cac6fa9f7fd-config-data" (OuterVolumeSpecName: "config-data") pod "c4182e68-340b-43ce-b458-0cac6fa9f7fd" (UID: "c4182e68-340b-43ce-b458-0cac6fa9f7fd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:24:07 crc kubenswrapper[4824]: I1211 02:24:07.379270 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4182e68-340b-43ce-b458-0cac6fa9f7fd-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 02:24:07 crc kubenswrapper[4824]: I1211 02:24:07.379303 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4182e68-340b-43ce-b458-0cac6fa9f7fd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 02:24:07 crc kubenswrapper[4824]: I1211 02:24:07.840074 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 11 02:24:07 crc kubenswrapper[4824]: I1211 02:24:07.878453 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 11 02:24:07 crc kubenswrapper[4824]: I1211 02:24:07.895678 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 11 02:24:07 crc kubenswrapper[4824]: I1211 02:24:07.895728 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 11 02:24:08 crc kubenswrapper[4824]: I1211 02:24:08.144325 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 11 02:24:08 crc kubenswrapper[4824]: I1211 02:24:08.206187 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 11 02:24:08 crc kubenswrapper[4824]: I1211 02:24:08.211501 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 11 02:24:08 crc kubenswrapper[4824]: I1211 02:24:08.213347 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 11 02:24:08 crc kubenswrapper[4824]: I1211 02:24:08.255185 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 11 02:24:08 crc kubenswrapper[4824]: E1211 02:24:08.255627 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4182e68-340b-43ce-b458-0cac6fa9f7fd" containerName="proxy-httpd" Dec 11 02:24:08 crc kubenswrapper[4824]: I1211 02:24:08.255645 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4182e68-340b-43ce-b458-0cac6fa9f7fd" containerName="proxy-httpd" Dec 11 02:24:08 crc kubenswrapper[4824]: E1211 02:24:08.255661 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4182e68-340b-43ce-b458-0cac6fa9f7fd" containerName="ceilometer-central-agent" Dec 11 02:24:08 crc kubenswrapper[4824]: I1211 02:24:08.255668 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4182e68-340b-43ce-b458-0cac6fa9f7fd" containerName="ceilometer-central-agent" Dec 11 02:24:08 crc kubenswrapper[4824]: E1211 02:24:08.255674 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4182e68-340b-43ce-b458-0cac6fa9f7fd" containerName="ceilometer-notification-agent" Dec 11 02:24:08 crc kubenswrapper[4824]: I1211 02:24:08.255680 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4182e68-340b-43ce-b458-0cac6fa9f7fd" containerName="ceilometer-notification-agent" Dec 11 02:24:08 crc kubenswrapper[4824]: E1211 02:24:08.255689 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4182e68-340b-43ce-b458-0cac6fa9f7fd" containerName="sg-core" Dec 11 02:24:08 crc kubenswrapper[4824]: I1211 02:24:08.255695 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4182e68-340b-43ce-b458-0cac6fa9f7fd" containerName="sg-core" Dec 11 02:24:08 crc kubenswrapper[4824]: I1211 02:24:08.255856 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4182e68-340b-43ce-b458-0cac6fa9f7fd" containerName="sg-core" Dec 11 02:24:08 crc kubenswrapper[4824]: I1211 02:24:08.255870 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4182e68-340b-43ce-b458-0cac6fa9f7fd" containerName="proxy-httpd" Dec 11 02:24:08 crc kubenswrapper[4824]: I1211 02:24:08.255889 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4182e68-340b-43ce-b458-0cac6fa9f7fd" containerName="ceilometer-notification-agent" Dec 11 02:24:08 crc kubenswrapper[4824]: I1211 02:24:08.255903 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4182e68-340b-43ce-b458-0cac6fa9f7fd" containerName="ceilometer-central-agent" Dec 11 02:24:08 crc kubenswrapper[4824]: I1211 02:24:08.257630 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 11 02:24:08 crc kubenswrapper[4824]: I1211 02:24:08.264898 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 11 02:24:08 crc kubenswrapper[4824]: I1211 02:24:08.265076 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 11 02:24:08 crc kubenswrapper[4824]: I1211 02:24:08.265258 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 11 02:24:08 crc kubenswrapper[4824]: I1211 02:24:08.270765 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 11 02:24:08 crc kubenswrapper[4824]: I1211 02:24:08.298307 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2d9a1d28-7094-46c4-ac74-d837b864b3c1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2d9a1d28-7094-46c4-ac74-d837b864b3c1\") " pod="openstack/ceilometer-0" Dec 11 02:24:08 crc kubenswrapper[4824]: I1211 02:24:08.298349 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/2d9a1d28-7094-46c4-ac74-d837b864b3c1-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"2d9a1d28-7094-46c4-ac74-d837b864b3c1\") " pod="openstack/ceilometer-0" Dec 11 02:24:08 crc kubenswrapper[4824]: I1211 02:24:08.298382 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d9a1d28-7094-46c4-ac74-d837b864b3c1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2d9a1d28-7094-46c4-ac74-d837b864b3c1\") " pod="openstack/ceilometer-0" Dec 11 02:24:08 crc kubenswrapper[4824]: I1211 02:24:08.298408 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2d9a1d28-7094-46c4-ac74-d837b864b3c1-scripts\") pod \"ceilometer-0\" (UID: \"2d9a1d28-7094-46c4-ac74-d837b864b3c1\") " pod="openstack/ceilometer-0" Dec 11 02:24:08 crc kubenswrapper[4824]: I1211 02:24:08.298433 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8q5xc\" (UniqueName: \"kubernetes.io/projected/2d9a1d28-7094-46c4-ac74-d837b864b3c1-kube-api-access-8q5xc\") pod \"ceilometer-0\" (UID: \"2d9a1d28-7094-46c4-ac74-d837b864b3c1\") " pod="openstack/ceilometer-0" Dec 11 02:24:08 crc kubenswrapper[4824]: I1211 02:24:08.298506 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d9a1d28-7094-46c4-ac74-d837b864b3c1-config-data\") pod \"ceilometer-0\" (UID: \"2d9a1d28-7094-46c4-ac74-d837b864b3c1\") " pod="openstack/ceilometer-0" Dec 11 02:24:08 crc kubenswrapper[4824]: I1211 02:24:08.298525 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2d9a1d28-7094-46c4-ac74-d837b864b3c1-run-httpd\") pod \"ceilometer-0\" (UID: \"2d9a1d28-7094-46c4-ac74-d837b864b3c1\") " pod="openstack/ceilometer-0" Dec 11 02:24:08 crc kubenswrapper[4824]: I1211 02:24:08.298547 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2d9a1d28-7094-46c4-ac74-d837b864b3c1-log-httpd\") pod \"ceilometer-0\" (UID: \"2d9a1d28-7094-46c4-ac74-d837b864b3c1\") " pod="openstack/ceilometer-0" Dec 11 02:24:08 crc kubenswrapper[4824]: I1211 02:24:08.400487 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2d9a1d28-7094-46c4-ac74-d837b864b3c1-log-httpd\") pod \"ceilometer-0\" (UID: \"2d9a1d28-7094-46c4-ac74-d837b864b3c1\") " pod="openstack/ceilometer-0" Dec 11 02:24:08 crc kubenswrapper[4824]: I1211 02:24:08.400575 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2d9a1d28-7094-46c4-ac74-d837b864b3c1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2d9a1d28-7094-46c4-ac74-d837b864b3c1\") " pod="openstack/ceilometer-0" Dec 11 02:24:08 crc kubenswrapper[4824]: I1211 02:24:08.400605 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/2d9a1d28-7094-46c4-ac74-d837b864b3c1-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"2d9a1d28-7094-46c4-ac74-d837b864b3c1\") " pod="openstack/ceilometer-0" Dec 11 02:24:08 crc kubenswrapper[4824]: I1211 02:24:08.400640 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d9a1d28-7094-46c4-ac74-d837b864b3c1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2d9a1d28-7094-46c4-ac74-d837b864b3c1\") " pod="openstack/ceilometer-0" Dec 11 02:24:08 crc kubenswrapper[4824]: I1211 02:24:08.400666 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2d9a1d28-7094-46c4-ac74-d837b864b3c1-scripts\") pod \"ceilometer-0\" (UID: \"2d9a1d28-7094-46c4-ac74-d837b864b3c1\") " pod="openstack/ceilometer-0" Dec 11 02:24:08 crc kubenswrapper[4824]: I1211 02:24:08.400688 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8q5xc\" (UniqueName: \"kubernetes.io/projected/2d9a1d28-7094-46c4-ac74-d837b864b3c1-kube-api-access-8q5xc\") pod \"ceilometer-0\" (UID: \"2d9a1d28-7094-46c4-ac74-d837b864b3c1\") " pod="openstack/ceilometer-0" Dec 11 02:24:08 crc kubenswrapper[4824]: I1211 02:24:08.400764 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d9a1d28-7094-46c4-ac74-d837b864b3c1-config-data\") pod \"ceilometer-0\" (UID: \"2d9a1d28-7094-46c4-ac74-d837b864b3c1\") " pod="openstack/ceilometer-0" Dec 11 02:24:08 crc kubenswrapper[4824]: I1211 02:24:08.400792 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2d9a1d28-7094-46c4-ac74-d837b864b3c1-run-httpd\") pod \"ceilometer-0\" (UID: \"2d9a1d28-7094-46c4-ac74-d837b864b3c1\") " pod="openstack/ceilometer-0" Dec 11 02:24:08 crc kubenswrapper[4824]: I1211 02:24:08.401073 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2d9a1d28-7094-46c4-ac74-d837b864b3c1-log-httpd\") pod \"ceilometer-0\" (UID: \"2d9a1d28-7094-46c4-ac74-d837b864b3c1\") " pod="openstack/ceilometer-0" Dec 11 02:24:08 crc kubenswrapper[4824]: I1211 02:24:08.401202 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2d9a1d28-7094-46c4-ac74-d837b864b3c1-run-httpd\") pod \"ceilometer-0\" (UID: \"2d9a1d28-7094-46c4-ac74-d837b864b3c1\") " pod="openstack/ceilometer-0" Dec 11 02:24:08 crc kubenswrapper[4824]: I1211 02:24:08.407359 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d9a1d28-7094-46c4-ac74-d837b864b3c1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2d9a1d28-7094-46c4-ac74-d837b864b3c1\") " pod="openstack/ceilometer-0" Dec 11 02:24:08 crc kubenswrapper[4824]: I1211 02:24:08.410454 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/2d9a1d28-7094-46c4-ac74-d837b864b3c1-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"2d9a1d28-7094-46c4-ac74-d837b864b3c1\") " pod="openstack/ceilometer-0" Dec 11 02:24:08 crc kubenswrapper[4824]: I1211 02:24:08.410855 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d9a1d28-7094-46c4-ac74-d837b864b3c1-config-data\") pod \"ceilometer-0\" (UID: \"2d9a1d28-7094-46c4-ac74-d837b864b3c1\") " pod="openstack/ceilometer-0" Dec 11 02:24:08 crc kubenswrapper[4824]: I1211 02:24:08.422921 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2d9a1d28-7094-46c4-ac74-d837b864b3c1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2d9a1d28-7094-46c4-ac74-d837b864b3c1\") " pod="openstack/ceilometer-0" Dec 11 02:24:08 crc kubenswrapper[4824]: I1211 02:24:08.423719 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2d9a1d28-7094-46c4-ac74-d837b864b3c1-scripts\") pod \"ceilometer-0\" (UID: \"2d9a1d28-7094-46c4-ac74-d837b864b3c1\") " pod="openstack/ceilometer-0" Dec 11 02:24:08 crc kubenswrapper[4824]: I1211 02:24:08.433344 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8q5xc\" (UniqueName: \"kubernetes.io/projected/2d9a1d28-7094-46c4-ac74-d837b864b3c1-kube-api-access-8q5xc\") pod \"ceilometer-0\" (UID: \"2d9a1d28-7094-46c4-ac74-d837b864b3c1\") " pod="openstack/ceilometer-0" Dec 11 02:24:08 crc kubenswrapper[4824]: I1211 02:24:08.594511 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 11 02:24:08 crc kubenswrapper[4824]: I1211 02:24:08.643510 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c4182e68-340b-43ce-b458-0cac6fa9f7fd" path="/var/lib/kubelet/pods/c4182e68-340b-43ce-b458-0cac6fa9f7fd/volumes" Dec 11 02:24:08 crc kubenswrapper[4824]: I1211 02:24:08.978420 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="420e8783-43b4-43c7-ae74-50df3b388af1" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.196:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 11 02:24:08 crc kubenswrapper[4824]: I1211 02:24:08.978456 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="420e8783-43b4-43c7-ae74-50df3b388af1" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.196:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 11 02:24:09 crc kubenswrapper[4824]: W1211 02:24:09.091676 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2d9a1d28_7094_46c4_ac74_d837b864b3c1.slice/crio-4706c470c78f687b936933b5a4232d3c41bec123752d0ee32ef0ad2e5476d39d WatchSource:0}: Error finding container 4706c470c78f687b936933b5a4232d3c41bec123752d0ee32ef0ad2e5476d39d: Status 404 returned error can't find the container with id 4706c470c78f687b936933b5a4232d3c41bec123752d0ee32ef0ad2e5476d39d Dec 11 02:24:09 crc kubenswrapper[4824]: I1211 02:24:09.093512 4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 11 02:24:09 crc kubenswrapper[4824]: I1211 02:24:09.102319 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 11 02:24:09 crc kubenswrapper[4824]: I1211 02:24:09.153234 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2d9a1d28-7094-46c4-ac74-d837b864b3c1","Type":"ContainerStarted","Data":"4706c470c78f687b936933b5a4232d3c41bec123752d0ee32ef0ad2e5476d39d"} Dec 11 02:24:09 crc kubenswrapper[4824]: I1211 02:24:09.927088 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-5hrpm" Dec 11 02:24:09 crc kubenswrapper[4824]: I1211 02:24:09.927410 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-5hrpm" Dec 11 02:24:09 crc kubenswrapper[4824]: I1211 02:24:09.977618 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-5hrpm" Dec 11 02:24:10 crc kubenswrapper[4824]: I1211 02:24:10.163212 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2d9a1d28-7094-46c4-ac74-d837b864b3c1","Type":"ContainerStarted","Data":"aae77d6d224333792761bd1e3c6f11759691c276741783edf3cf0e36ad3e9596"} Dec 11 02:24:10 crc kubenswrapper[4824]: I1211 02:24:10.213432 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-5hrpm" Dec 11 02:24:10 crc kubenswrapper[4824]: I1211 02:24:10.264665 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5hrpm"] Dec 11 02:24:11 crc kubenswrapper[4824]: I1211 02:24:11.176991 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2d9a1d28-7094-46c4-ac74-d837b864b3c1","Type":"ContainerStarted","Data":"78d943742b416bfd57c49237836050355f0627a2657e50c8486cb5d00fb512f9"} Dec 11 02:24:12 crc kubenswrapper[4824]: I1211 02:24:12.190372 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2d9a1d28-7094-46c4-ac74-d837b864b3c1","Type":"ContainerStarted","Data":"3055a188a673ecc45cabc09804530c6a11fc33359f7b5b614d7fc109d114bf88"} Dec 11 02:24:12 crc kubenswrapper[4824]: I1211 02:24:12.190572 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-5hrpm" podUID="171e204b-8e23-4f40-b666-ba76cda27715" containerName="registry-server" containerID="cri-o://992a079394d9659cb31e3d70b0ca545d3f15717620bdc3290217c904935e80f0" gracePeriod=2 Dec 11 02:24:12 crc kubenswrapper[4824]: I1211 02:24:12.337845 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 11 02:24:12 crc kubenswrapper[4824]: I1211 02:24:12.347213 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 11 02:24:12 crc kubenswrapper[4824]: I1211 02:24:12.350602 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 11 02:24:12 crc kubenswrapper[4824]: E1211 02:24:12.389575 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod171e204b_8e23_4f40_b666_ba76cda27715.slice/crio-992a079394d9659cb31e3d70b0ca545d3f15717620bdc3290217c904935e80f0.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod171e204b_8e23_4f40_b666_ba76cda27715.slice/crio-conmon-992a079394d9659cb31e3d70b0ca545d3f15717620bdc3290217c904935e80f0.scope\": RecentStats: unable to find data in memory cache]" Dec 11 02:24:12 crc kubenswrapper[4824]: I1211 02:24:12.684102 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5hrpm" Dec 11 02:24:12 crc kubenswrapper[4824]: I1211 02:24:12.833176 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/171e204b-8e23-4f40-b666-ba76cda27715-catalog-content\") pod \"171e204b-8e23-4f40-b666-ba76cda27715\" (UID: \"171e204b-8e23-4f40-b666-ba76cda27715\") " Dec 11 02:24:12 crc kubenswrapper[4824]: I1211 02:24:12.833235 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/171e204b-8e23-4f40-b666-ba76cda27715-utilities\") pod \"171e204b-8e23-4f40-b666-ba76cda27715\" (UID: \"171e204b-8e23-4f40-b666-ba76cda27715\") " Dec 11 02:24:12 crc kubenswrapper[4824]: I1211 02:24:12.833428 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vr2f8\" (UniqueName: \"kubernetes.io/projected/171e204b-8e23-4f40-b666-ba76cda27715-kube-api-access-vr2f8\") pod \"171e204b-8e23-4f40-b666-ba76cda27715\" (UID: \"171e204b-8e23-4f40-b666-ba76cda27715\") " Dec 11 02:24:12 crc kubenswrapper[4824]: I1211 02:24:12.833934 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/171e204b-8e23-4f40-b666-ba76cda27715-utilities" (OuterVolumeSpecName: "utilities") pod "171e204b-8e23-4f40-b666-ba76cda27715" (UID: "171e204b-8e23-4f40-b666-ba76cda27715"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:24:12 crc kubenswrapper[4824]: I1211 02:24:12.838594 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/171e204b-8e23-4f40-b666-ba76cda27715-kube-api-access-vr2f8" (OuterVolumeSpecName: "kube-api-access-vr2f8") pod "171e204b-8e23-4f40-b666-ba76cda27715" (UID: "171e204b-8e23-4f40-b666-ba76cda27715"). InnerVolumeSpecName "kube-api-access-vr2f8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:24:12 crc kubenswrapper[4824]: I1211 02:24:12.860270 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/171e204b-8e23-4f40-b666-ba76cda27715-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "171e204b-8e23-4f40-b666-ba76cda27715" (UID: "171e204b-8e23-4f40-b666-ba76cda27715"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:24:12 crc kubenswrapper[4824]: I1211 02:24:12.937653 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vr2f8\" (UniqueName: \"kubernetes.io/projected/171e204b-8e23-4f40-b666-ba76cda27715-kube-api-access-vr2f8\") on node \"crc\" DevicePath \"\"" Dec 11 02:24:12 crc kubenswrapper[4824]: I1211 02:24:12.937685 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/171e204b-8e23-4f40-b666-ba76cda27715-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 02:24:12 crc kubenswrapper[4824]: I1211 02:24:12.937694 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/171e204b-8e23-4f40-b666-ba76cda27715-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 02:24:13 crc kubenswrapper[4824]: I1211 02:24:13.203051 4824 generic.go:334] "Generic (PLEG): container finished" podID="171e204b-8e23-4f40-b666-ba76cda27715" containerID="992a079394d9659cb31e3d70b0ca545d3f15717620bdc3290217c904935e80f0" exitCode=0 Dec 11 02:24:13 crc kubenswrapper[4824]: I1211 02:24:13.203096 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5hrpm" event={"ID":"171e204b-8e23-4f40-b666-ba76cda27715","Type":"ContainerDied","Data":"992a079394d9659cb31e3d70b0ca545d3f15717620bdc3290217c904935e80f0"} Dec 11 02:24:13 crc kubenswrapper[4824]: I1211 02:24:13.203476 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5hrpm" event={"ID":"171e204b-8e23-4f40-b666-ba76cda27715","Type":"ContainerDied","Data":"cc963f0ebf4301760bfb4fcec958dabdc15208ff5fdf946474b5d4ca849c4ac5"} Dec 11 02:24:13 crc kubenswrapper[4824]: I1211 02:24:13.203500 4824 scope.go:117] "RemoveContainer" containerID="992a079394d9659cb31e3d70b0ca545d3f15717620bdc3290217c904935e80f0" Dec 11 02:24:13 crc kubenswrapper[4824]: I1211 02:24:13.203178 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5hrpm" Dec 11 02:24:13 crc kubenswrapper[4824]: I1211 02:24:13.216184 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2d9a1d28-7094-46c4-ac74-d837b864b3c1","Type":"ContainerStarted","Data":"e45f0d6a802642667bc97e4a89a33af55e11779129fd173c303abe4c4736a6af"} Dec 11 02:24:13 crc kubenswrapper[4824]: I1211 02:24:13.216343 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 11 02:24:13 crc kubenswrapper[4824]: I1211 02:24:13.225925 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 11 02:24:13 crc kubenswrapper[4824]: I1211 02:24:13.250539 4824 scope.go:117] "RemoveContainer" containerID="acf0049ba571845043ca3de50fcab3256cf2495fb1fd0003f0cce9ffec3d1666" Dec 11 02:24:13 crc kubenswrapper[4824]: I1211 02:24:13.250901 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.492889673 podStartE2EDuration="5.250872828s" podCreationTimestamp="2025-12-11 02:24:08 +0000 UTC" firstStartedPulling="2025-12-11 02:24:09.093299911 +0000 UTC m=+1390.782337290" lastFinishedPulling="2025-12-11 02:24:12.851283066 +0000 UTC m=+1394.540320445" observedRunningTime="2025-12-11 02:24:13.237383199 +0000 UTC m=+1394.926420588" watchObservedRunningTime="2025-12-11 02:24:13.250872828 +0000 UTC m=+1394.939910227" Dec 11 02:24:13 crc kubenswrapper[4824]: I1211 02:24:13.284143 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5hrpm"] Dec 11 02:24:13 crc kubenswrapper[4824]: I1211 02:24:13.309743 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-5hrpm"] Dec 11 02:24:13 crc kubenswrapper[4824]: I1211 02:24:13.324240 4824 scope.go:117] "RemoveContainer" containerID="04e078ecc7483ffe84663fa4aaf2c3aabb21d8070a637c15e9c8438d289375ab" Dec 11 02:24:13 crc kubenswrapper[4824]: I1211 02:24:13.404039 4824 scope.go:117] "RemoveContainer" containerID="992a079394d9659cb31e3d70b0ca545d3f15717620bdc3290217c904935e80f0" Dec 11 02:24:13 crc kubenswrapper[4824]: E1211 02:24:13.404642 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"992a079394d9659cb31e3d70b0ca545d3f15717620bdc3290217c904935e80f0\": container with ID starting with 992a079394d9659cb31e3d70b0ca545d3f15717620bdc3290217c904935e80f0 not found: ID does not exist" containerID="992a079394d9659cb31e3d70b0ca545d3f15717620bdc3290217c904935e80f0" Dec 11 02:24:13 crc kubenswrapper[4824]: I1211 02:24:13.404671 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"992a079394d9659cb31e3d70b0ca545d3f15717620bdc3290217c904935e80f0"} err="failed to get container status \"992a079394d9659cb31e3d70b0ca545d3f15717620bdc3290217c904935e80f0\": rpc error: code = NotFound desc = could not find container \"992a079394d9659cb31e3d70b0ca545d3f15717620bdc3290217c904935e80f0\": container with ID starting with 992a079394d9659cb31e3d70b0ca545d3f15717620bdc3290217c904935e80f0 not found: ID does not exist" Dec 11 02:24:13 crc kubenswrapper[4824]: I1211 02:24:13.404691 4824 scope.go:117] "RemoveContainer" containerID="acf0049ba571845043ca3de50fcab3256cf2495fb1fd0003f0cce9ffec3d1666" Dec 11 02:24:13 crc kubenswrapper[4824]: E1211 02:24:13.404959 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"acf0049ba571845043ca3de50fcab3256cf2495fb1fd0003f0cce9ffec3d1666\": container with ID starting with acf0049ba571845043ca3de50fcab3256cf2495fb1fd0003f0cce9ffec3d1666 not found: ID does not exist" containerID="acf0049ba571845043ca3de50fcab3256cf2495fb1fd0003f0cce9ffec3d1666" Dec 11 02:24:13 crc kubenswrapper[4824]: I1211 02:24:13.405004 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"acf0049ba571845043ca3de50fcab3256cf2495fb1fd0003f0cce9ffec3d1666"} err="failed to get container status \"acf0049ba571845043ca3de50fcab3256cf2495fb1fd0003f0cce9ffec3d1666\": rpc error: code = NotFound desc = could not find container \"acf0049ba571845043ca3de50fcab3256cf2495fb1fd0003f0cce9ffec3d1666\": container with ID starting with acf0049ba571845043ca3de50fcab3256cf2495fb1fd0003f0cce9ffec3d1666 not found: ID does not exist" Dec 11 02:24:13 crc kubenswrapper[4824]: I1211 02:24:13.405039 4824 scope.go:117] "RemoveContainer" containerID="04e078ecc7483ffe84663fa4aaf2c3aabb21d8070a637c15e9c8438d289375ab" Dec 11 02:24:13 crc kubenswrapper[4824]: E1211 02:24:13.405514 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"04e078ecc7483ffe84663fa4aaf2c3aabb21d8070a637c15e9c8438d289375ab\": container with ID starting with 04e078ecc7483ffe84663fa4aaf2c3aabb21d8070a637c15e9c8438d289375ab not found: ID does not exist" containerID="04e078ecc7483ffe84663fa4aaf2c3aabb21d8070a637c15e9c8438d289375ab" Dec 11 02:24:13 crc kubenswrapper[4824]: I1211 02:24:13.405550 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"04e078ecc7483ffe84663fa4aaf2c3aabb21d8070a637c15e9c8438d289375ab"} err="failed to get container status \"04e078ecc7483ffe84663fa4aaf2c3aabb21d8070a637c15e9c8438d289375ab\": rpc error: code = NotFound desc = could not find container \"04e078ecc7483ffe84663fa4aaf2c3aabb21d8070a637c15e9c8438d289375ab\": container with ID starting with 04e078ecc7483ffe84663fa4aaf2c3aabb21d8070a637c15e9c8438d289375ab not found: ID does not exist" Dec 11 02:24:13 crc kubenswrapper[4824]: I1211 02:24:13.464569 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 11 02:24:14 crc kubenswrapper[4824]: I1211 02:24:14.675966 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="171e204b-8e23-4f40-b666-ba76cda27715" path="/var/lib/kubelet/pods/171e204b-8e23-4f40-b666-ba76cda27715/volumes" Dec 11 02:24:16 crc kubenswrapper[4824]: I1211 02:24:16.185320 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 11 02:24:16 crc kubenswrapper[4824]: I1211 02:24:16.245223 4824 generic.go:334] "Generic (PLEG): container finished" podID="8f850bc8-5b69-478a-b8d4-1bd17f34b4b2" containerID="f9a6eef8a9b612afb5ddf1eb1a6fbf0a6f1270d7de14586e824ce8b0e4c638df" exitCode=137 Dec 11 02:24:16 crc kubenswrapper[4824]: I1211 02:24:16.245263 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"8f850bc8-5b69-478a-b8d4-1bd17f34b4b2","Type":"ContainerDied","Data":"f9a6eef8a9b612afb5ddf1eb1a6fbf0a6f1270d7de14586e824ce8b0e4c638df"} Dec 11 02:24:16 crc kubenswrapper[4824]: I1211 02:24:16.245288 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"8f850bc8-5b69-478a-b8d4-1bd17f34b4b2","Type":"ContainerDied","Data":"3ae43ba0ce737bb48adf1d1b9861b7e276443787156939a8d9ef7f8dd9a0ca5c"} Dec 11 02:24:16 crc kubenswrapper[4824]: I1211 02:24:16.245305 4824 scope.go:117] "RemoveContainer" containerID="f9a6eef8a9b612afb5ddf1eb1a6fbf0a6f1270d7de14586e824ce8b0e4c638df" Dec 11 02:24:16 crc kubenswrapper[4824]: I1211 02:24:16.245375 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 11 02:24:16 crc kubenswrapper[4824]: I1211 02:24:16.267628 4824 scope.go:117] "RemoveContainer" containerID="f9a6eef8a9b612afb5ddf1eb1a6fbf0a6f1270d7de14586e824ce8b0e4c638df" Dec 11 02:24:16 crc kubenswrapper[4824]: E1211 02:24:16.268339 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f9a6eef8a9b612afb5ddf1eb1a6fbf0a6f1270d7de14586e824ce8b0e4c638df\": container with ID starting with f9a6eef8a9b612afb5ddf1eb1a6fbf0a6f1270d7de14586e824ce8b0e4c638df not found: ID does not exist" containerID="f9a6eef8a9b612afb5ddf1eb1a6fbf0a6f1270d7de14586e824ce8b0e4c638df" Dec 11 02:24:16 crc kubenswrapper[4824]: I1211 02:24:16.268392 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f9a6eef8a9b612afb5ddf1eb1a6fbf0a6f1270d7de14586e824ce8b0e4c638df"} err="failed to get container status \"f9a6eef8a9b612afb5ddf1eb1a6fbf0a6f1270d7de14586e824ce8b0e4c638df\": rpc error: code = NotFound desc = could not find container \"f9a6eef8a9b612afb5ddf1eb1a6fbf0a6f1270d7de14586e824ce8b0e4c638df\": container with ID starting with f9a6eef8a9b612afb5ddf1eb1a6fbf0a6f1270d7de14586e824ce8b0e4c638df not found: ID does not exist" Dec 11 02:24:16 crc kubenswrapper[4824]: I1211 02:24:16.301890 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f850bc8-5b69-478a-b8d4-1bd17f34b4b2-combined-ca-bundle\") pod \"8f850bc8-5b69-478a-b8d4-1bd17f34b4b2\" (UID: \"8f850bc8-5b69-478a-b8d4-1bd17f34b4b2\") " Dec 11 02:24:16 crc kubenswrapper[4824]: I1211 02:24:16.301980 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f850bc8-5b69-478a-b8d4-1bd17f34b4b2-config-data\") pod \"8f850bc8-5b69-478a-b8d4-1bd17f34b4b2\" (UID: \"8f850bc8-5b69-478a-b8d4-1bd17f34b4b2\") " Dec 11 02:24:16 crc kubenswrapper[4824]: I1211 02:24:16.302000 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dgrpt\" (UniqueName: \"kubernetes.io/projected/8f850bc8-5b69-478a-b8d4-1bd17f34b4b2-kube-api-access-dgrpt\") pod \"8f850bc8-5b69-478a-b8d4-1bd17f34b4b2\" (UID: \"8f850bc8-5b69-478a-b8d4-1bd17f34b4b2\") " Dec 11 02:24:16 crc kubenswrapper[4824]: I1211 02:24:16.310256 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f850bc8-5b69-478a-b8d4-1bd17f34b4b2-kube-api-access-dgrpt" (OuterVolumeSpecName: "kube-api-access-dgrpt") pod "8f850bc8-5b69-478a-b8d4-1bd17f34b4b2" (UID: "8f850bc8-5b69-478a-b8d4-1bd17f34b4b2"). InnerVolumeSpecName "kube-api-access-dgrpt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:24:16 crc kubenswrapper[4824]: I1211 02:24:16.345583 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f850bc8-5b69-478a-b8d4-1bd17f34b4b2-config-data" (OuterVolumeSpecName: "config-data") pod "8f850bc8-5b69-478a-b8d4-1bd17f34b4b2" (UID: "8f850bc8-5b69-478a-b8d4-1bd17f34b4b2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:24:16 crc kubenswrapper[4824]: I1211 02:24:16.347976 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f850bc8-5b69-478a-b8d4-1bd17f34b4b2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8f850bc8-5b69-478a-b8d4-1bd17f34b4b2" (UID: "8f850bc8-5b69-478a-b8d4-1bd17f34b4b2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:24:16 crc kubenswrapper[4824]: I1211 02:24:16.404421 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f850bc8-5b69-478a-b8d4-1bd17f34b4b2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 02:24:16 crc kubenswrapper[4824]: I1211 02:24:16.404733 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f850bc8-5b69-478a-b8d4-1bd17f34b4b2-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 02:24:16 crc kubenswrapper[4824]: I1211 02:24:16.404747 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dgrpt\" (UniqueName: \"kubernetes.io/projected/8f850bc8-5b69-478a-b8d4-1bd17f34b4b2-kube-api-access-dgrpt\") on node \"crc\" DevicePath \"\"" Dec 11 02:24:16 crc kubenswrapper[4824]: I1211 02:24:16.589794 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 11 02:24:16 crc kubenswrapper[4824]: I1211 02:24:16.609656 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 11 02:24:16 crc kubenswrapper[4824]: I1211 02:24:16.625026 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 11 02:24:16 crc kubenswrapper[4824]: E1211 02:24:16.625421 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f850bc8-5b69-478a-b8d4-1bd17f34b4b2" containerName="nova-cell1-novncproxy-novncproxy" Dec 11 02:24:16 crc kubenswrapper[4824]: I1211 02:24:16.625439 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f850bc8-5b69-478a-b8d4-1bd17f34b4b2" containerName="nova-cell1-novncproxy-novncproxy" Dec 11 02:24:16 crc kubenswrapper[4824]: E1211 02:24:16.625472 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="171e204b-8e23-4f40-b666-ba76cda27715" containerName="extract-content" Dec 11 02:24:16 crc kubenswrapper[4824]: I1211 02:24:16.625479 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="171e204b-8e23-4f40-b666-ba76cda27715" containerName="extract-content" Dec 11 02:24:16 crc kubenswrapper[4824]: E1211 02:24:16.625486 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="171e204b-8e23-4f40-b666-ba76cda27715" containerName="registry-server" Dec 11 02:24:16 crc kubenswrapper[4824]: I1211 02:24:16.625492 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="171e204b-8e23-4f40-b666-ba76cda27715" containerName="registry-server" Dec 11 02:24:16 crc kubenswrapper[4824]: E1211 02:24:16.625514 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="171e204b-8e23-4f40-b666-ba76cda27715" containerName="extract-utilities" Dec 11 02:24:16 crc kubenswrapper[4824]: I1211 02:24:16.625520 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="171e204b-8e23-4f40-b666-ba76cda27715" containerName="extract-utilities" Dec 11 02:24:16 crc kubenswrapper[4824]: I1211 02:24:16.625686 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="171e204b-8e23-4f40-b666-ba76cda27715" containerName="registry-server" Dec 11 02:24:16 crc kubenswrapper[4824]: I1211 02:24:16.625710 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f850bc8-5b69-478a-b8d4-1bd17f34b4b2" containerName="nova-cell1-novncproxy-novncproxy" Dec 11 02:24:16 crc kubenswrapper[4824]: I1211 02:24:16.626475 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 11 02:24:16 crc kubenswrapper[4824]: I1211 02:24:16.631097 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Dec 11 02:24:16 crc kubenswrapper[4824]: I1211 02:24:16.633507 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 11 02:24:16 crc kubenswrapper[4824]: I1211 02:24:16.633831 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Dec 11 02:24:16 crc kubenswrapper[4824]: I1211 02:24:16.643602 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f850bc8-5b69-478a-b8d4-1bd17f34b4b2" path="/var/lib/kubelet/pods/8f850bc8-5b69-478a-b8d4-1bd17f34b4b2/volumes" Dec 11 02:24:16 crc kubenswrapper[4824]: I1211 02:24:16.649774 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 11 02:24:16 crc kubenswrapper[4824]: I1211 02:24:16.710489 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c246a287-b22f-405b-8a24-c50765f5b153-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"c246a287-b22f-405b-8a24-c50765f5b153\") " pod="openstack/nova-cell1-novncproxy-0" Dec 11 02:24:16 crc kubenswrapper[4824]: I1211 02:24:16.710574 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7pntw\" (UniqueName: \"kubernetes.io/projected/c246a287-b22f-405b-8a24-c50765f5b153-kube-api-access-7pntw\") pod \"nova-cell1-novncproxy-0\" (UID: \"c246a287-b22f-405b-8a24-c50765f5b153\") " pod="openstack/nova-cell1-novncproxy-0" Dec 11 02:24:16 crc kubenswrapper[4824]: I1211 02:24:16.710616 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c246a287-b22f-405b-8a24-c50765f5b153-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"c246a287-b22f-405b-8a24-c50765f5b153\") " pod="openstack/nova-cell1-novncproxy-0" Dec 11 02:24:16 crc kubenswrapper[4824]: I1211 02:24:16.710645 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/c246a287-b22f-405b-8a24-c50765f5b153-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"c246a287-b22f-405b-8a24-c50765f5b153\") " pod="openstack/nova-cell1-novncproxy-0" Dec 11 02:24:16 crc kubenswrapper[4824]: I1211 02:24:16.710675 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/c246a287-b22f-405b-8a24-c50765f5b153-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"c246a287-b22f-405b-8a24-c50765f5b153\") " pod="openstack/nova-cell1-novncproxy-0" Dec 11 02:24:16 crc kubenswrapper[4824]: I1211 02:24:16.813676 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c246a287-b22f-405b-8a24-c50765f5b153-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"c246a287-b22f-405b-8a24-c50765f5b153\") " pod="openstack/nova-cell1-novncproxy-0" Dec 11 02:24:16 crc kubenswrapper[4824]: I1211 02:24:16.815007 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7pntw\" (UniqueName: \"kubernetes.io/projected/c246a287-b22f-405b-8a24-c50765f5b153-kube-api-access-7pntw\") pod \"nova-cell1-novncproxy-0\" (UID: \"c246a287-b22f-405b-8a24-c50765f5b153\") " pod="openstack/nova-cell1-novncproxy-0" Dec 11 02:24:16 crc kubenswrapper[4824]: I1211 02:24:16.815334 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c246a287-b22f-405b-8a24-c50765f5b153-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"c246a287-b22f-405b-8a24-c50765f5b153\") " pod="openstack/nova-cell1-novncproxy-0" Dec 11 02:24:16 crc kubenswrapper[4824]: I1211 02:24:16.816136 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/c246a287-b22f-405b-8a24-c50765f5b153-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"c246a287-b22f-405b-8a24-c50765f5b153\") " pod="openstack/nova-cell1-novncproxy-0" Dec 11 02:24:16 crc kubenswrapper[4824]: I1211 02:24:16.816450 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/c246a287-b22f-405b-8a24-c50765f5b153-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"c246a287-b22f-405b-8a24-c50765f5b153\") " pod="openstack/nova-cell1-novncproxy-0" Dec 11 02:24:16 crc kubenswrapper[4824]: I1211 02:24:16.819554 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c246a287-b22f-405b-8a24-c50765f5b153-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"c246a287-b22f-405b-8a24-c50765f5b153\") " pod="openstack/nova-cell1-novncproxy-0" Dec 11 02:24:16 crc kubenswrapper[4824]: I1211 02:24:16.819616 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/c246a287-b22f-405b-8a24-c50765f5b153-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"c246a287-b22f-405b-8a24-c50765f5b153\") " pod="openstack/nova-cell1-novncproxy-0" Dec 11 02:24:16 crc kubenswrapper[4824]: I1211 02:24:16.820171 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c246a287-b22f-405b-8a24-c50765f5b153-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"c246a287-b22f-405b-8a24-c50765f5b153\") " pod="openstack/nova-cell1-novncproxy-0" Dec 11 02:24:16 crc kubenswrapper[4824]: I1211 02:24:16.831879 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/c246a287-b22f-405b-8a24-c50765f5b153-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"c246a287-b22f-405b-8a24-c50765f5b153\") " pod="openstack/nova-cell1-novncproxy-0" Dec 11 02:24:16 crc kubenswrapper[4824]: I1211 02:24:16.839355 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7pntw\" (UniqueName: \"kubernetes.io/projected/c246a287-b22f-405b-8a24-c50765f5b153-kube-api-access-7pntw\") pod \"nova-cell1-novncproxy-0\" (UID: \"c246a287-b22f-405b-8a24-c50765f5b153\") " pod="openstack/nova-cell1-novncproxy-0" Dec 11 02:24:16 crc kubenswrapper[4824]: I1211 02:24:16.978375 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 11 02:24:17 crc kubenswrapper[4824]: I1211 02:24:17.486844 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 11 02:24:17 crc kubenswrapper[4824]: I1211 02:24:17.900053 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 11 02:24:17 crc kubenswrapper[4824]: I1211 02:24:17.900708 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 11 02:24:17 crc kubenswrapper[4824]: I1211 02:24:17.903547 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 11 02:24:17 crc kubenswrapper[4824]: I1211 02:24:17.903781 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 11 02:24:18 crc kubenswrapper[4824]: I1211 02:24:18.287879 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"c246a287-b22f-405b-8a24-c50765f5b153","Type":"ContainerStarted","Data":"1a071d2ff7ce7fe1cddee032a425dbe88845678077d6e993ef00843d73eb8200"} Dec 11 02:24:18 crc kubenswrapper[4824]: I1211 02:24:18.288362 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"c246a287-b22f-405b-8a24-c50765f5b153","Type":"ContainerStarted","Data":"d415e3cda38c6d8e6414ab4d805028562a91a394e5d5b0b35032f74861712573"} Dec 11 02:24:18 crc kubenswrapper[4824]: I1211 02:24:18.288432 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 11 02:24:18 crc kubenswrapper[4824]: I1211 02:24:18.294034 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 11 02:24:18 crc kubenswrapper[4824]: I1211 02:24:18.320738 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.32071167 podStartE2EDuration="2.32071167s" podCreationTimestamp="2025-12-11 02:24:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:24:18.309411297 +0000 UTC m=+1399.998448706" watchObservedRunningTime="2025-12-11 02:24:18.32071167 +0000 UTC m=+1400.009749079" Dec 11 02:24:18 crc kubenswrapper[4824]: I1211 02:24:18.564583 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-n9sd5"] Dec 11 02:24:18 crc kubenswrapper[4824]: I1211 02:24:18.570203 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-n9sd5" Dec 11 02:24:18 crc kubenswrapper[4824]: I1211 02:24:18.583206 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-n9sd5"] Dec 11 02:24:18 crc kubenswrapper[4824]: I1211 02:24:18.756043 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c358674b-3741-4d28-a629-029d94cdc3d4-dns-swift-storage-0\") pod \"dnsmasq-dns-89c5cd4d5-n9sd5\" (UID: \"c358674b-3741-4d28-a629-029d94cdc3d4\") " pod="openstack/dnsmasq-dns-89c5cd4d5-n9sd5" Dec 11 02:24:18 crc kubenswrapper[4824]: I1211 02:24:18.756084 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c358674b-3741-4d28-a629-029d94cdc3d4-ovsdbserver-sb\") pod \"dnsmasq-dns-89c5cd4d5-n9sd5\" (UID: \"c358674b-3741-4d28-a629-029d94cdc3d4\") " pod="openstack/dnsmasq-dns-89c5cd4d5-n9sd5" Dec 11 02:24:18 crc kubenswrapper[4824]: I1211 02:24:18.756167 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nmnsp\" (UniqueName: \"kubernetes.io/projected/c358674b-3741-4d28-a629-029d94cdc3d4-kube-api-access-nmnsp\") pod \"dnsmasq-dns-89c5cd4d5-n9sd5\" (UID: \"c358674b-3741-4d28-a629-029d94cdc3d4\") " pod="openstack/dnsmasq-dns-89c5cd4d5-n9sd5" Dec 11 02:24:18 crc kubenswrapper[4824]: I1211 02:24:18.756223 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c358674b-3741-4d28-a629-029d94cdc3d4-dns-svc\") pod \"dnsmasq-dns-89c5cd4d5-n9sd5\" (UID: \"c358674b-3741-4d28-a629-029d94cdc3d4\") " pod="openstack/dnsmasq-dns-89c5cd4d5-n9sd5" Dec 11 02:24:18 crc kubenswrapper[4824]: I1211 02:24:18.756245 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c358674b-3741-4d28-a629-029d94cdc3d4-ovsdbserver-nb\") pod \"dnsmasq-dns-89c5cd4d5-n9sd5\" (UID: \"c358674b-3741-4d28-a629-029d94cdc3d4\") " pod="openstack/dnsmasq-dns-89c5cd4d5-n9sd5" Dec 11 02:24:18 crc kubenswrapper[4824]: I1211 02:24:18.757758 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c358674b-3741-4d28-a629-029d94cdc3d4-config\") pod \"dnsmasq-dns-89c5cd4d5-n9sd5\" (UID: \"c358674b-3741-4d28-a629-029d94cdc3d4\") " pod="openstack/dnsmasq-dns-89c5cd4d5-n9sd5" Dec 11 02:24:18 crc kubenswrapper[4824]: I1211 02:24:18.859292 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c358674b-3741-4d28-a629-029d94cdc3d4-ovsdbserver-nb\") pod \"dnsmasq-dns-89c5cd4d5-n9sd5\" (UID: \"c358674b-3741-4d28-a629-029d94cdc3d4\") " pod="openstack/dnsmasq-dns-89c5cd4d5-n9sd5" Dec 11 02:24:18 crc kubenswrapper[4824]: I1211 02:24:18.859370 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c358674b-3741-4d28-a629-029d94cdc3d4-config\") pod \"dnsmasq-dns-89c5cd4d5-n9sd5\" (UID: \"c358674b-3741-4d28-a629-029d94cdc3d4\") " pod="openstack/dnsmasq-dns-89c5cd4d5-n9sd5" Dec 11 02:24:18 crc kubenswrapper[4824]: I1211 02:24:18.859412 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c358674b-3741-4d28-a629-029d94cdc3d4-dns-swift-storage-0\") pod \"dnsmasq-dns-89c5cd4d5-n9sd5\" (UID: \"c358674b-3741-4d28-a629-029d94cdc3d4\") " pod="openstack/dnsmasq-dns-89c5cd4d5-n9sd5" Dec 11 02:24:18 crc kubenswrapper[4824]: I1211 02:24:18.859429 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c358674b-3741-4d28-a629-029d94cdc3d4-ovsdbserver-sb\") pod \"dnsmasq-dns-89c5cd4d5-n9sd5\" (UID: \"c358674b-3741-4d28-a629-029d94cdc3d4\") " pod="openstack/dnsmasq-dns-89c5cd4d5-n9sd5" Dec 11 02:24:18 crc kubenswrapper[4824]: I1211 02:24:18.859498 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nmnsp\" (UniqueName: \"kubernetes.io/projected/c358674b-3741-4d28-a629-029d94cdc3d4-kube-api-access-nmnsp\") pod \"dnsmasq-dns-89c5cd4d5-n9sd5\" (UID: \"c358674b-3741-4d28-a629-029d94cdc3d4\") " pod="openstack/dnsmasq-dns-89c5cd4d5-n9sd5" Dec 11 02:24:18 crc kubenswrapper[4824]: I1211 02:24:18.859553 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c358674b-3741-4d28-a629-029d94cdc3d4-dns-svc\") pod \"dnsmasq-dns-89c5cd4d5-n9sd5\" (UID: \"c358674b-3741-4d28-a629-029d94cdc3d4\") " pod="openstack/dnsmasq-dns-89c5cd4d5-n9sd5" Dec 11 02:24:18 crc kubenswrapper[4824]: I1211 02:24:18.860145 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c358674b-3741-4d28-a629-029d94cdc3d4-ovsdbserver-nb\") pod \"dnsmasq-dns-89c5cd4d5-n9sd5\" (UID: \"c358674b-3741-4d28-a629-029d94cdc3d4\") " pod="openstack/dnsmasq-dns-89c5cd4d5-n9sd5" Dec 11 02:24:18 crc kubenswrapper[4824]: I1211 02:24:18.860263 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c358674b-3741-4d28-a629-029d94cdc3d4-dns-swift-storage-0\") pod \"dnsmasq-dns-89c5cd4d5-n9sd5\" (UID: \"c358674b-3741-4d28-a629-029d94cdc3d4\") " pod="openstack/dnsmasq-dns-89c5cd4d5-n9sd5" Dec 11 02:24:18 crc kubenswrapper[4824]: I1211 02:24:18.860352 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c358674b-3741-4d28-a629-029d94cdc3d4-config\") pod \"dnsmasq-dns-89c5cd4d5-n9sd5\" (UID: \"c358674b-3741-4d28-a629-029d94cdc3d4\") " pod="openstack/dnsmasq-dns-89c5cd4d5-n9sd5" Dec 11 02:24:18 crc kubenswrapper[4824]: I1211 02:24:18.860541 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c358674b-3741-4d28-a629-029d94cdc3d4-dns-svc\") pod \"dnsmasq-dns-89c5cd4d5-n9sd5\" (UID: \"c358674b-3741-4d28-a629-029d94cdc3d4\") " pod="openstack/dnsmasq-dns-89c5cd4d5-n9sd5" Dec 11 02:24:18 crc kubenswrapper[4824]: I1211 02:24:18.861185 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c358674b-3741-4d28-a629-029d94cdc3d4-ovsdbserver-sb\") pod \"dnsmasq-dns-89c5cd4d5-n9sd5\" (UID: \"c358674b-3741-4d28-a629-029d94cdc3d4\") " pod="openstack/dnsmasq-dns-89c5cd4d5-n9sd5" Dec 11 02:24:18 crc kubenswrapper[4824]: I1211 02:24:18.879172 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nmnsp\" (UniqueName: \"kubernetes.io/projected/c358674b-3741-4d28-a629-029d94cdc3d4-kube-api-access-nmnsp\") pod \"dnsmasq-dns-89c5cd4d5-n9sd5\" (UID: \"c358674b-3741-4d28-a629-029d94cdc3d4\") " pod="openstack/dnsmasq-dns-89c5cd4d5-n9sd5" Dec 11 02:24:18 crc kubenswrapper[4824]: I1211 02:24:18.890337 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-n9sd5" Dec 11 02:24:19 crc kubenswrapper[4824]: I1211 02:24:19.356676 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-n9sd5"] Dec 11 02:24:19 crc kubenswrapper[4824]: W1211 02:24:19.365859 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc358674b_3741_4d28_a629_029d94cdc3d4.slice/crio-353d3c2d26c496b382830a8aeaa639e60dcca462352b24b7bf941ab2a4d0d71e WatchSource:0}: Error finding container 353d3c2d26c496b382830a8aeaa639e60dcca462352b24b7bf941ab2a4d0d71e: Status 404 returned error can't find the container with id 353d3c2d26c496b382830a8aeaa639e60dcca462352b24b7bf941ab2a4d0d71e Dec 11 02:24:20 crc kubenswrapper[4824]: I1211 02:24:20.306085 4824 generic.go:334] "Generic (PLEG): container finished" podID="c358674b-3741-4d28-a629-029d94cdc3d4" containerID="67044c3ed0dfd1e496a077f8d2c586480c8b3e34954ce85885b81960285fd53d" exitCode=0 Dec 11 02:24:20 crc kubenswrapper[4824]: I1211 02:24:20.306161 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-n9sd5" event={"ID":"c358674b-3741-4d28-a629-029d94cdc3d4","Type":"ContainerDied","Data":"67044c3ed0dfd1e496a077f8d2c586480c8b3e34954ce85885b81960285fd53d"} Dec 11 02:24:20 crc kubenswrapper[4824]: I1211 02:24:20.306449 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-n9sd5" event={"ID":"c358674b-3741-4d28-a629-029d94cdc3d4","Type":"ContainerStarted","Data":"353d3c2d26c496b382830a8aeaa639e60dcca462352b24b7bf941ab2a4d0d71e"} Dec 11 02:24:20 crc kubenswrapper[4824]: I1211 02:24:20.433478 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 11 02:24:20 crc kubenswrapper[4824]: I1211 02:24:20.434038 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2d9a1d28-7094-46c4-ac74-d837b864b3c1" containerName="proxy-httpd" containerID="cri-o://e45f0d6a802642667bc97e4a89a33af55e11779129fd173c303abe4c4736a6af" gracePeriod=30 Dec 11 02:24:20 crc kubenswrapper[4824]: I1211 02:24:20.434407 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2d9a1d28-7094-46c4-ac74-d837b864b3c1" containerName="sg-core" containerID="cri-o://3055a188a673ecc45cabc09804530c6a11fc33359f7b5b614d7fc109d114bf88" gracePeriod=30 Dec 11 02:24:20 crc kubenswrapper[4824]: I1211 02:24:20.434496 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2d9a1d28-7094-46c4-ac74-d837b864b3c1" containerName="ceilometer-notification-agent" containerID="cri-o://78d943742b416bfd57c49237836050355f0627a2657e50c8486cb5d00fb512f9" gracePeriod=30 Dec 11 02:24:20 crc kubenswrapper[4824]: I1211 02:24:20.436193 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2d9a1d28-7094-46c4-ac74-d837b864b3c1" containerName="ceilometer-central-agent" containerID="cri-o://aae77d6d224333792761bd1e3c6f11759691c276741783edf3cf0e36ad3e9596" gracePeriod=30 Dec 11 02:24:21 crc kubenswrapper[4824]: I1211 02:24:21.168338 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 11 02:24:21 crc kubenswrapper[4824]: I1211 02:24:21.314631 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-n9sd5" event={"ID":"c358674b-3741-4d28-a629-029d94cdc3d4","Type":"ContainerStarted","Data":"bd89a5a921ae19774166b4f494c2bb2e0ce856de55580b2d902ebdb5c24f7d9e"} Dec 11 02:24:21 crc kubenswrapper[4824]: I1211 02:24:21.314797 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-89c5cd4d5-n9sd5" Dec 11 02:24:21 crc kubenswrapper[4824]: I1211 02:24:21.316738 4824 generic.go:334] "Generic (PLEG): container finished" podID="2d9a1d28-7094-46c4-ac74-d837b864b3c1" containerID="e45f0d6a802642667bc97e4a89a33af55e11779129fd173c303abe4c4736a6af" exitCode=0 Dec 11 02:24:21 crc kubenswrapper[4824]: I1211 02:24:21.316781 4824 generic.go:334] "Generic (PLEG): container finished" podID="2d9a1d28-7094-46c4-ac74-d837b864b3c1" containerID="3055a188a673ecc45cabc09804530c6a11fc33359f7b5b614d7fc109d114bf88" exitCode=2 Dec 11 02:24:21 crc kubenswrapper[4824]: I1211 02:24:21.316799 4824 generic.go:334] "Generic (PLEG): container finished" podID="2d9a1d28-7094-46c4-ac74-d837b864b3c1" containerID="aae77d6d224333792761bd1e3c6f11759691c276741783edf3cf0e36ad3e9596" exitCode=0 Dec 11 02:24:21 crc kubenswrapper[4824]: I1211 02:24:21.316801 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2d9a1d28-7094-46c4-ac74-d837b864b3c1","Type":"ContainerDied","Data":"e45f0d6a802642667bc97e4a89a33af55e11779129fd173c303abe4c4736a6af"} Dec 11 02:24:21 crc kubenswrapper[4824]: I1211 02:24:21.316847 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2d9a1d28-7094-46c4-ac74-d837b864b3c1","Type":"ContainerDied","Data":"3055a188a673ecc45cabc09804530c6a11fc33359f7b5b614d7fc109d114bf88"} Dec 11 02:24:21 crc kubenswrapper[4824]: I1211 02:24:21.316857 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2d9a1d28-7094-46c4-ac74-d837b864b3c1","Type":"ContainerDied","Data":"aae77d6d224333792761bd1e3c6f11759691c276741783edf3cf0e36ad3e9596"} Dec 11 02:24:21 crc kubenswrapper[4824]: I1211 02:24:21.317033 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="420e8783-43b4-43c7-ae74-50df3b388af1" containerName="nova-api-log" containerID="cri-o://9b98e7345115867a4832021da7738c9d8e895fc4f81273785a40c60ab44ac9d1" gracePeriod=30 Dec 11 02:24:21 crc kubenswrapper[4824]: I1211 02:24:21.317058 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="420e8783-43b4-43c7-ae74-50df3b388af1" containerName="nova-api-api" containerID="cri-o://1809cc49214843841f512f2b18ee339fa99a9a75e7d0725b1b98e8fdb65214ef" gracePeriod=30 Dec 11 02:24:21 crc kubenswrapper[4824]: I1211 02:24:21.352830 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-89c5cd4d5-n9sd5" podStartSLOduration=3.352802709 podStartE2EDuration="3.352802709s" podCreationTimestamp="2025-12-11 02:24:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:24:21.338395697 +0000 UTC m=+1403.027433076" watchObservedRunningTime="2025-12-11 02:24:21.352802709 +0000 UTC m=+1403.041840118" Dec 11 02:24:21 crc kubenswrapper[4824]: I1211 02:24:21.979534 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 11 02:24:22 crc kubenswrapper[4824]: I1211 02:24:22.339477 4824 generic.go:334] "Generic (PLEG): container finished" podID="420e8783-43b4-43c7-ae74-50df3b388af1" containerID="9b98e7345115867a4832021da7738c9d8e895fc4f81273785a40c60ab44ac9d1" exitCode=143 Dec 11 02:24:22 crc kubenswrapper[4824]: I1211 02:24:22.339856 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"420e8783-43b4-43c7-ae74-50df3b388af1","Type":"ContainerDied","Data":"9b98e7345115867a4832021da7738c9d8e895fc4f81273785a40c60ab44ac9d1"} Dec 11 02:24:22 crc kubenswrapper[4824]: I1211 02:24:22.349447 4824 generic.go:334] "Generic (PLEG): container finished" podID="2d9a1d28-7094-46c4-ac74-d837b864b3c1" containerID="78d943742b416bfd57c49237836050355f0627a2657e50c8486cb5d00fb512f9" exitCode=0 Dec 11 02:24:22 crc kubenswrapper[4824]: I1211 02:24:22.350228 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2d9a1d28-7094-46c4-ac74-d837b864b3c1","Type":"ContainerDied","Data":"78d943742b416bfd57c49237836050355f0627a2657e50c8486cb5d00fb512f9"} Dec 11 02:24:22 crc kubenswrapper[4824]: I1211 02:24:22.549192 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 11 02:24:22 crc kubenswrapper[4824]: I1211 02:24:22.723457 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2d9a1d28-7094-46c4-ac74-d837b864b3c1-run-httpd\") pod \"2d9a1d28-7094-46c4-ac74-d837b864b3c1\" (UID: \"2d9a1d28-7094-46c4-ac74-d837b864b3c1\") " Dec 11 02:24:22 crc kubenswrapper[4824]: I1211 02:24:22.723639 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2d9a1d28-7094-46c4-ac74-d837b864b3c1-log-httpd\") pod \"2d9a1d28-7094-46c4-ac74-d837b864b3c1\" (UID: \"2d9a1d28-7094-46c4-ac74-d837b864b3c1\") " Dec 11 02:24:22 crc kubenswrapper[4824]: I1211 02:24:22.723694 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/2d9a1d28-7094-46c4-ac74-d837b864b3c1-ceilometer-tls-certs\") pod \"2d9a1d28-7094-46c4-ac74-d837b864b3c1\" (UID: \"2d9a1d28-7094-46c4-ac74-d837b864b3c1\") " Dec 11 02:24:22 crc kubenswrapper[4824]: I1211 02:24:22.723734 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8q5xc\" (UniqueName: \"kubernetes.io/projected/2d9a1d28-7094-46c4-ac74-d837b864b3c1-kube-api-access-8q5xc\") pod \"2d9a1d28-7094-46c4-ac74-d837b864b3c1\" (UID: \"2d9a1d28-7094-46c4-ac74-d837b864b3c1\") " Dec 11 02:24:22 crc kubenswrapper[4824]: I1211 02:24:22.723773 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2d9a1d28-7094-46c4-ac74-d837b864b3c1-sg-core-conf-yaml\") pod \"2d9a1d28-7094-46c4-ac74-d837b864b3c1\" (UID: \"2d9a1d28-7094-46c4-ac74-d837b864b3c1\") " Dec 11 02:24:22 crc kubenswrapper[4824]: I1211 02:24:22.723825 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d9a1d28-7094-46c4-ac74-d837b864b3c1-config-data\") pod \"2d9a1d28-7094-46c4-ac74-d837b864b3c1\" (UID: \"2d9a1d28-7094-46c4-ac74-d837b864b3c1\") " Dec 11 02:24:22 crc kubenswrapper[4824]: I1211 02:24:22.723871 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2d9a1d28-7094-46c4-ac74-d837b864b3c1-scripts\") pod \"2d9a1d28-7094-46c4-ac74-d837b864b3c1\" (UID: \"2d9a1d28-7094-46c4-ac74-d837b864b3c1\") " Dec 11 02:24:22 crc kubenswrapper[4824]: I1211 02:24:22.723977 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d9a1d28-7094-46c4-ac74-d837b864b3c1-combined-ca-bundle\") pod \"2d9a1d28-7094-46c4-ac74-d837b864b3c1\" (UID: \"2d9a1d28-7094-46c4-ac74-d837b864b3c1\") " Dec 11 02:24:22 crc kubenswrapper[4824]: I1211 02:24:22.725436 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2d9a1d28-7094-46c4-ac74-d837b864b3c1-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "2d9a1d28-7094-46c4-ac74-d837b864b3c1" (UID: "2d9a1d28-7094-46c4-ac74-d837b864b3c1"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:24:22 crc kubenswrapper[4824]: I1211 02:24:22.725731 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2d9a1d28-7094-46c4-ac74-d837b864b3c1-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "2d9a1d28-7094-46c4-ac74-d837b864b3c1" (UID: "2d9a1d28-7094-46c4-ac74-d837b864b3c1"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:24:22 crc kubenswrapper[4824]: I1211 02:24:22.737462 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d9a1d28-7094-46c4-ac74-d837b864b3c1-scripts" (OuterVolumeSpecName: "scripts") pod "2d9a1d28-7094-46c4-ac74-d837b864b3c1" (UID: "2d9a1d28-7094-46c4-ac74-d837b864b3c1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:24:22 crc kubenswrapper[4824]: I1211 02:24:22.738413 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d9a1d28-7094-46c4-ac74-d837b864b3c1-kube-api-access-8q5xc" (OuterVolumeSpecName: "kube-api-access-8q5xc") pod "2d9a1d28-7094-46c4-ac74-d837b864b3c1" (UID: "2d9a1d28-7094-46c4-ac74-d837b864b3c1"). InnerVolumeSpecName "kube-api-access-8q5xc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:24:22 crc kubenswrapper[4824]: I1211 02:24:22.768843 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d9a1d28-7094-46c4-ac74-d837b864b3c1-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "2d9a1d28-7094-46c4-ac74-d837b864b3c1" (UID: "2d9a1d28-7094-46c4-ac74-d837b864b3c1"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:24:22 crc kubenswrapper[4824]: I1211 02:24:22.826808 4824 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2d9a1d28-7094-46c4-ac74-d837b864b3c1-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 11 02:24:22 crc kubenswrapper[4824]: I1211 02:24:22.826847 4824 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2d9a1d28-7094-46c4-ac74-d837b864b3c1-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 11 02:24:22 crc kubenswrapper[4824]: I1211 02:24:22.826860 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8q5xc\" (UniqueName: \"kubernetes.io/projected/2d9a1d28-7094-46c4-ac74-d837b864b3c1-kube-api-access-8q5xc\") on node \"crc\" DevicePath \"\"" Dec 11 02:24:22 crc kubenswrapper[4824]: I1211 02:24:22.826872 4824 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2d9a1d28-7094-46c4-ac74-d837b864b3c1-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 11 02:24:22 crc kubenswrapper[4824]: I1211 02:24:22.826882 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2d9a1d28-7094-46c4-ac74-d837b864b3c1-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 02:24:22 crc kubenswrapper[4824]: I1211 02:24:22.840757 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d9a1d28-7094-46c4-ac74-d837b864b3c1-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "2d9a1d28-7094-46c4-ac74-d837b864b3c1" (UID: "2d9a1d28-7094-46c4-ac74-d837b864b3c1"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:24:22 crc kubenswrapper[4824]: I1211 02:24:22.849019 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d9a1d28-7094-46c4-ac74-d837b864b3c1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2d9a1d28-7094-46c4-ac74-d837b864b3c1" (UID: "2d9a1d28-7094-46c4-ac74-d837b864b3c1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:24:22 crc kubenswrapper[4824]: I1211 02:24:22.883260 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d9a1d28-7094-46c4-ac74-d837b864b3c1-config-data" (OuterVolumeSpecName: "config-data") pod "2d9a1d28-7094-46c4-ac74-d837b864b3c1" (UID: "2d9a1d28-7094-46c4-ac74-d837b864b3c1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:24:22 crc kubenswrapper[4824]: I1211 02:24:22.928048 4824 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/2d9a1d28-7094-46c4-ac74-d837b864b3c1-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 11 02:24:22 crc kubenswrapper[4824]: I1211 02:24:22.928085 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d9a1d28-7094-46c4-ac74-d837b864b3c1-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 02:24:22 crc kubenswrapper[4824]: I1211 02:24:22.928095 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d9a1d28-7094-46c4-ac74-d837b864b3c1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 02:24:23 crc kubenswrapper[4824]: I1211 02:24:23.362024 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2d9a1d28-7094-46c4-ac74-d837b864b3c1","Type":"ContainerDied","Data":"4706c470c78f687b936933b5a4232d3c41bec123752d0ee32ef0ad2e5476d39d"} Dec 11 02:24:23 crc kubenswrapper[4824]: I1211 02:24:23.363044 4824 scope.go:117] "RemoveContainer" containerID="e45f0d6a802642667bc97e4a89a33af55e11779129fd173c303abe4c4736a6af" Dec 11 02:24:23 crc kubenswrapper[4824]: I1211 02:24:23.362345 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 11 02:24:23 crc kubenswrapper[4824]: I1211 02:24:23.414879 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 11 02:24:23 crc kubenswrapper[4824]: I1211 02:24:23.429821 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 11 02:24:23 crc kubenswrapper[4824]: I1211 02:24:23.433070 4824 scope.go:117] "RemoveContainer" containerID="3055a188a673ecc45cabc09804530c6a11fc33359f7b5b614d7fc109d114bf88" Dec 11 02:24:23 crc kubenswrapper[4824]: I1211 02:24:23.449921 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 11 02:24:23 crc kubenswrapper[4824]: E1211 02:24:23.450365 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d9a1d28-7094-46c4-ac74-d837b864b3c1" containerName="proxy-httpd" Dec 11 02:24:23 crc kubenswrapper[4824]: I1211 02:24:23.450382 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d9a1d28-7094-46c4-ac74-d837b864b3c1" containerName="proxy-httpd" Dec 11 02:24:23 crc kubenswrapper[4824]: E1211 02:24:23.450406 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d9a1d28-7094-46c4-ac74-d837b864b3c1" containerName="sg-core" Dec 11 02:24:23 crc kubenswrapper[4824]: I1211 02:24:23.450412 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d9a1d28-7094-46c4-ac74-d837b864b3c1" containerName="sg-core" Dec 11 02:24:23 crc kubenswrapper[4824]: E1211 02:24:23.450426 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d9a1d28-7094-46c4-ac74-d837b864b3c1" containerName="ceilometer-notification-agent" Dec 11 02:24:23 crc kubenswrapper[4824]: I1211 02:24:23.450433 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d9a1d28-7094-46c4-ac74-d837b864b3c1" containerName="ceilometer-notification-agent" Dec 11 02:24:23 crc kubenswrapper[4824]: E1211 02:24:23.450447 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d9a1d28-7094-46c4-ac74-d837b864b3c1" containerName="ceilometer-central-agent" Dec 11 02:24:23 crc kubenswrapper[4824]: I1211 02:24:23.450452 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d9a1d28-7094-46c4-ac74-d837b864b3c1" containerName="ceilometer-central-agent" Dec 11 02:24:23 crc kubenswrapper[4824]: I1211 02:24:23.450651 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d9a1d28-7094-46c4-ac74-d837b864b3c1" containerName="ceilometer-notification-agent" Dec 11 02:24:23 crc kubenswrapper[4824]: I1211 02:24:23.450660 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d9a1d28-7094-46c4-ac74-d837b864b3c1" containerName="sg-core" Dec 11 02:24:23 crc kubenswrapper[4824]: I1211 02:24:23.450672 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d9a1d28-7094-46c4-ac74-d837b864b3c1" containerName="proxy-httpd" Dec 11 02:24:23 crc kubenswrapper[4824]: I1211 02:24:23.450683 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d9a1d28-7094-46c4-ac74-d837b864b3c1" containerName="ceilometer-central-agent" Dec 11 02:24:23 crc kubenswrapper[4824]: I1211 02:24:23.452307 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 11 02:24:23 crc kubenswrapper[4824]: I1211 02:24:23.456564 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 11 02:24:23 crc kubenswrapper[4824]: I1211 02:24:23.458094 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 11 02:24:23 crc kubenswrapper[4824]: I1211 02:24:23.459593 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 11 02:24:23 crc kubenswrapper[4824]: I1211 02:24:23.481002 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 11 02:24:23 crc kubenswrapper[4824]: I1211 02:24:23.486277 4824 scope.go:117] "RemoveContainer" containerID="78d943742b416bfd57c49237836050355f0627a2657e50c8486cb5d00fb512f9" Dec 11 02:24:23 crc kubenswrapper[4824]: I1211 02:24:23.510646 4824 scope.go:117] "RemoveContainer" containerID="aae77d6d224333792761bd1e3c6f11759691c276741783edf3cf0e36ad3e9596" Dec 11 02:24:23 crc kubenswrapper[4824]: I1211 02:24:23.537425 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c65e7e08-e393-4cd6-b17e-3d9e81783e39-log-httpd\") pod \"ceilometer-0\" (UID: \"c65e7e08-e393-4cd6-b17e-3d9e81783e39\") " pod="openstack/ceilometer-0" Dec 11 02:24:23 crc kubenswrapper[4824]: I1211 02:24:23.537487 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c65e7e08-e393-4cd6-b17e-3d9e81783e39-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c65e7e08-e393-4cd6-b17e-3d9e81783e39\") " pod="openstack/ceilometer-0" Dec 11 02:24:23 crc kubenswrapper[4824]: I1211 02:24:23.537520 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c65e7e08-e393-4cd6-b17e-3d9e81783e39-config-data\") pod \"ceilometer-0\" (UID: \"c65e7e08-e393-4cd6-b17e-3d9e81783e39\") " pod="openstack/ceilometer-0" Dec 11 02:24:23 crc kubenswrapper[4824]: I1211 02:24:23.537552 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c65e7e08-e393-4cd6-b17e-3d9e81783e39-scripts\") pod \"ceilometer-0\" (UID: \"c65e7e08-e393-4cd6-b17e-3d9e81783e39\") " pod="openstack/ceilometer-0" Dec 11 02:24:23 crc kubenswrapper[4824]: I1211 02:24:23.537568 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c65e7e08-e393-4cd6-b17e-3d9e81783e39-run-httpd\") pod \"ceilometer-0\" (UID: \"c65e7e08-e393-4cd6-b17e-3d9e81783e39\") " pod="openstack/ceilometer-0" Dec 11 02:24:23 crc kubenswrapper[4824]: I1211 02:24:23.537583 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c65e7e08-e393-4cd6-b17e-3d9e81783e39-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c65e7e08-e393-4cd6-b17e-3d9e81783e39\") " pod="openstack/ceilometer-0" Dec 11 02:24:23 crc kubenswrapper[4824]: I1211 02:24:23.537763 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m87xg\" (UniqueName: \"kubernetes.io/projected/c65e7e08-e393-4cd6-b17e-3d9e81783e39-kube-api-access-m87xg\") pod \"ceilometer-0\" (UID: \"c65e7e08-e393-4cd6-b17e-3d9e81783e39\") " pod="openstack/ceilometer-0" Dec 11 02:24:23 crc kubenswrapper[4824]: I1211 02:24:23.537840 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c65e7e08-e393-4cd6-b17e-3d9e81783e39-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"c65e7e08-e393-4cd6-b17e-3d9e81783e39\") " pod="openstack/ceilometer-0" Dec 11 02:24:23 crc kubenswrapper[4824]: I1211 02:24:23.638884 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c65e7e08-e393-4cd6-b17e-3d9e81783e39-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c65e7e08-e393-4cd6-b17e-3d9e81783e39\") " pod="openstack/ceilometer-0" Dec 11 02:24:23 crc kubenswrapper[4824]: I1211 02:24:23.638926 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c65e7e08-e393-4cd6-b17e-3d9e81783e39-config-data\") pod \"ceilometer-0\" (UID: \"c65e7e08-e393-4cd6-b17e-3d9e81783e39\") " pod="openstack/ceilometer-0" Dec 11 02:24:23 crc kubenswrapper[4824]: I1211 02:24:23.638960 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c65e7e08-e393-4cd6-b17e-3d9e81783e39-scripts\") pod \"ceilometer-0\" (UID: \"c65e7e08-e393-4cd6-b17e-3d9e81783e39\") " pod="openstack/ceilometer-0" Dec 11 02:24:23 crc kubenswrapper[4824]: I1211 02:24:23.638974 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c65e7e08-e393-4cd6-b17e-3d9e81783e39-run-httpd\") pod \"ceilometer-0\" (UID: \"c65e7e08-e393-4cd6-b17e-3d9e81783e39\") " pod="openstack/ceilometer-0" Dec 11 02:24:23 crc kubenswrapper[4824]: I1211 02:24:23.638991 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c65e7e08-e393-4cd6-b17e-3d9e81783e39-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c65e7e08-e393-4cd6-b17e-3d9e81783e39\") " pod="openstack/ceilometer-0" Dec 11 02:24:23 crc kubenswrapper[4824]: I1211 02:24:23.639006 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m87xg\" (UniqueName: \"kubernetes.io/projected/c65e7e08-e393-4cd6-b17e-3d9e81783e39-kube-api-access-m87xg\") pod \"ceilometer-0\" (UID: \"c65e7e08-e393-4cd6-b17e-3d9e81783e39\") " pod="openstack/ceilometer-0" Dec 11 02:24:23 crc kubenswrapper[4824]: I1211 02:24:23.639092 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c65e7e08-e393-4cd6-b17e-3d9e81783e39-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"c65e7e08-e393-4cd6-b17e-3d9e81783e39\") " pod="openstack/ceilometer-0" Dec 11 02:24:23 crc kubenswrapper[4824]: I1211 02:24:23.639151 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c65e7e08-e393-4cd6-b17e-3d9e81783e39-log-httpd\") pod \"ceilometer-0\" (UID: \"c65e7e08-e393-4cd6-b17e-3d9e81783e39\") " pod="openstack/ceilometer-0" Dec 11 02:24:23 crc kubenswrapper[4824]: I1211 02:24:23.639513 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c65e7e08-e393-4cd6-b17e-3d9e81783e39-log-httpd\") pod \"ceilometer-0\" (UID: \"c65e7e08-e393-4cd6-b17e-3d9e81783e39\") " pod="openstack/ceilometer-0" Dec 11 02:24:23 crc kubenswrapper[4824]: I1211 02:24:23.639826 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c65e7e08-e393-4cd6-b17e-3d9e81783e39-run-httpd\") pod \"ceilometer-0\" (UID: \"c65e7e08-e393-4cd6-b17e-3d9e81783e39\") " pod="openstack/ceilometer-0" Dec 11 02:24:23 crc kubenswrapper[4824]: I1211 02:24:23.643256 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c65e7e08-e393-4cd6-b17e-3d9e81783e39-scripts\") pod \"ceilometer-0\" (UID: \"c65e7e08-e393-4cd6-b17e-3d9e81783e39\") " pod="openstack/ceilometer-0" Dec 11 02:24:23 crc kubenswrapper[4824]: I1211 02:24:23.644399 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c65e7e08-e393-4cd6-b17e-3d9e81783e39-config-data\") pod \"ceilometer-0\" (UID: \"c65e7e08-e393-4cd6-b17e-3d9e81783e39\") " pod="openstack/ceilometer-0" Dec 11 02:24:23 crc kubenswrapper[4824]: I1211 02:24:23.644592 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c65e7e08-e393-4cd6-b17e-3d9e81783e39-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c65e7e08-e393-4cd6-b17e-3d9e81783e39\") " pod="openstack/ceilometer-0" Dec 11 02:24:23 crc kubenswrapper[4824]: I1211 02:24:23.644655 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c65e7e08-e393-4cd6-b17e-3d9e81783e39-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"c65e7e08-e393-4cd6-b17e-3d9e81783e39\") " pod="openstack/ceilometer-0" Dec 11 02:24:23 crc kubenswrapper[4824]: I1211 02:24:23.649477 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c65e7e08-e393-4cd6-b17e-3d9e81783e39-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c65e7e08-e393-4cd6-b17e-3d9e81783e39\") " pod="openstack/ceilometer-0" Dec 11 02:24:23 crc kubenswrapper[4824]: I1211 02:24:23.657295 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m87xg\" (UniqueName: \"kubernetes.io/projected/c65e7e08-e393-4cd6-b17e-3d9e81783e39-kube-api-access-m87xg\") pod \"ceilometer-0\" (UID: \"c65e7e08-e393-4cd6-b17e-3d9e81783e39\") " pod="openstack/ceilometer-0" Dec 11 02:24:23 crc kubenswrapper[4824]: I1211 02:24:23.776675 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 11 02:24:24 crc kubenswrapper[4824]: I1211 02:24:24.066291 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 11 02:24:24 crc kubenswrapper[4824]: W1211 02:24:24.072241 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc65e7e08_e393_4cd6_b17e_3d9e81783e39.slice/crio-d161f45fc5ab3104e89418eca0910b7607a3a5a6185d3d4a6cc6d4c350ddb497 WatchSource:0}: Error finding container d161f45fc5ab3104e89418eca0910b7607a3a5a6185d3d4a6cc6d4c350ddb497: Status 404 returned error can't find the container with id d161f45fc5ab3104e89418eca0910b7607a3a5a6185d3d4a6cc6d4c350ddb497 Dec 11 02:24:24 crc kubenswrapper[4824]: I1211 02:24:24.373611 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c65e7e08-e393-4cd6-b17e-3d9e81783e39","Type":"ContainerStarted","Data":"d161f45fc5ab3104e89418eca0910b7607a3a5a6185d3d4a6cc6d4c350ddb497"} Dec 11 02:24:24 crc kubenswrapper[4824]: I1211 02:24:24.645033 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d9a1d28-7094-46c4-ac74-d837b864b3c1" path="/var/lib/kubelet/pods/2d9a1d28-7094-46c4-ac74-d837b864b3c1/volumes" Dec 11 02:24:24 crc kubenswrapper[4824]: I1211 02:24:24.984636 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 11 02:24:25 crc kubenswrapper[4824]: I1211 02:24:25.074717 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ds6ql\" (UniqueName: \"kubernetes.io/projected/420e8783-43b4-43c7-ae74-50df3b388af1-kube-api-access-ds6ql\") pod \"420e8783-43b4-43c7-ae74-50df3b388af1\" (UID: \"420e8783-43b4-43c7-ae74-50df3b388af1\") " Dec 11 02:24:25 crc kubenswrapper[4824]: I1211 02:24:25.074786 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/420e8783-43b4-43c7-ae74-50df3b388af1-combined-ca-bundle\") pod \"420e8783-43b4-43c7-ae74-50df3b388af1\" (UID: \"420e8783-43b4-43c7-ae74-50df3b388af1\") " Dec 11 02:24:25 crc kubenswrapper[4824]: I1211 02:24:25.074916 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/420e8783-43b4-43c7-ae74-50df3b388af1-config-data\") pod \"420e8783-43b4-43c7-ae74-50df3b388af1\" (UID: \"420e8783-43b4-43c7-ae74-50df3b388af1\") " Dec 11 02:24:25 crc kubenswrapper[4824]: I1211 02:24:25.074947 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/420e8783-43b4-43c7-ae74-50df3b388af1-logs\") pod \"420e8783-43b4-43c7-ae74-50df3b388af1\" (UID: \"420e8783-43b4-43c7-ae74-50df3b388af1\") " Dec 11 02:24:25 crc kubenswrapper[4824]: I1211 02:24:25.075663 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/420e8783-43b4-43c7-ae74-50df3b388af1-logs" (OuterVolumeSpecName: "logs") pod "420e8783-43b4-43c7-ae74-50df3b388af1" (UID: "420e8783-43b4-43c7-ae74-50df3b388af1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:24:25 crc kubenswrapper[4824]: I1211 02:24:25.081591 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/420e8783-43b4-43c7-ae74-50df3b388af1-kube-api-access-ds6ql" (OuterVolumeSpecName: "kube-api-access-ds6ql") pod "420e8783-43b4-43c7-ae74-50df3b388af1" (UID: "420e8783-43b4-43c7-ae74-50df3b388af1"). InnerVolumeSpecName "kube-api-access-ds6ql". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:24:25 crc kubenswrapper[4824]: I1211 02:24:25.115740 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/420e8783-43b4-43c7-ae74-50df3b388af1-config-data" (OuterVolumeSpecName: "config-data") pod "420e8783-43b4-43c7-ae74-50df3b388af1" (UID: "420e8783-43b4-43c7-ae74-50df3b388af1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:24:25 crc kubenswrapper[4824]: I1211 02:24:25.116163 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/420e8783-43b4-43c7-ae74-50df3b388af1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "420e8783-43b4-43c7-ae74-50df3b388af1" (UID: "420e8783-43b4-43c7-ae74-50df3b388af1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:24:25 crc kubenswrapper[4824]: I1211 02:24:25.176480 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ds6ql\" (UniqueName: \"kubernetes.io/projected/420e8783-43b4-43c7-ae74-50df3b388af1-kube-api-access-ds6ql\") on node \"crc\" DevicePath \"\"" Dec 11 02:24:25 crc kubenswrapper[4824]: I1211 02:24:25.176507 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/420e8783-43b4-43c7-ae74-50df3b388af1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 02:24:25 crc kubenswrapper[4824]: I1211 02:24:25.176517 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/420e8783-43b4-43c7-ae74-50df3b388af1-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 02:24:25 crc kubenswrapper[4824]: I1211 02:24:25.176527 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/420e8783-43b4-43c7-ae74-50df3b388af1-logs\") on node \"crc\" DevicePath \"\"" Dec 11 02:24:25 crc kubenswrapper[4824]: I1211 02:24:25.383079 4824 generic.go:334] "Generic (PLEG): container finished" podID="420e8783-43b4-43c7-ae74-50df3b388af1" containerID="1809cc49214843841f512f2b18ee339fa99a9a75e7d0725b1b98e8fdb65214ef" exitCode=0 Dec 11 02:24:25 crc kubenswrapper[4824]: I1211 02:24:25.383236 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"420e8783-43b4-43c7-ae74-50df3b388af1","Type":"ContainerDied","Data":"1809cc49214843841f512f2b18ee339fa99a9a75e7d0725b1b98e8fdb65214ef"} Dec 11 02:24:25 crc kubenswrapper[4824]: I1211 02:24:25.383276 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"420e8783-43b4-43c7-ae74-50df3b388af1","Type":"ContainerDied","Data":"3b7fd39b5617935f8edcdd5fffb3357f8ac1b85ed77435ddcac6e7bc81b9bb14"} Dec 11 02:24:25 crc kubenswrapper[4824]: I1211 02:24:25.383285 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 11 02:24:25 crc kubenswrapper[4824]: I1211 02:24:25.383304 4824 scope.go:117] "RemoveContainer" containerID="1809cc49214843841f512f2b18ee339fa99a9a75e7d0725b1b98e8fdb65214ef" Dec 11 02:24:25 crc kubenswrapper[4824]: I1211 02:24:25.386261 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c65e7e08-e393-4cd6-b17e-3d9e81783e39","Type":"ContainerStarted","Data":"1e051935efac1e1fc37c2a2e140f55e9bac38d593133325033c87e37289d086f"} Dec 11 02:24:25 crc kubenswrapper[4824]: I1211 02:24:25.424154 4824 scope.go:117] "RemoveContainer" containerID="9b98e7345115867a4832021da7738c9d8e895fc4f81273785a40c60ab44ac9d1" Dec 11 02:24:25 crc kubenswrapper[4824]: I1211 02:24:25.429474 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 11 02:24:25 crc kubenswrapper[4824]: I1211 02:24:25.459189 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 11 02:24:25 crc kubenswrapper[4824]: I1211 02:24:25.466754 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 11 02:24:25 crc kubenswrapper[4824]: E1211 02:24:25.467464 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="420e8783-43b4-43c7-ae74-50df3b388af1" containerName="nova-api-api" Dec 11 02:24:25 crc kubenswrapper[4824]: I1211 02:24:25.467497 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="420e8783-43b4-43c7-ae74-50df3b388af1" containerName="nova-api-api" Dec 11 02:24:25 crc kubenswrapper[4824]: E1211 02:24:25.467530 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="420e8783-43b4-43c7-ae74-50df3b388af1" containerName="nova-api-log" Dec 11 02:24:25 crc kubenswrapper[4824]: I1211 02:24:25.467545 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="420e8783-43b4-43c7-ae74-50df3b388af1" containerName="nova-api-log" Dec 11 02:24:25 crc kubenswrapper[4824]: I1211 02:24:25.466792 4824 scope.go:117] "RemoveContainer" containerID="1809cc49214843841f512f2b18ee339fa99a9a75e7d0725b1b98e8fdb65214ef" Dec 11 02:24:25 crc kubenswrapper[4824]: I1211 02:24:25.467925 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="420e8783-43b4-43c7-ae74-50df3b388af1" containerName="nova-api-log" Dec 11 02:24:25 crc kubenswrapper[4824]: I1211 02:24:25.467976 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="420e8783-43b4-43c7-ae74-50df3b388af1" containerName="nova-api-api" Dec 11 02:24:25 crc kubenswrapper[4824]: I1211 02:24:25.469804 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 11 02:24:25 crc kubenswrapper[4824]: E1211 02:24:25.471233 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1809cc49214843841f512f2b18ee339fa99a9a75e7d0725b1b98e8fdb65214ef\": container with ID starting with 1809cc49214843841f512f2b18ee339fa99a9a75e7d0725b1b98e8fdb65214ef not found: ID does not exist" containerID="1809cc49214843841f512f2b18ee339fa99a9a75e7d0725b1b98e8fdb65214ef" Dec 11 02:24:25 crc kubenswrapper[4824]: I1211 02:24:25.471290 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1809cc49214843841f512f2b18ee339fa99a9a75e7d0725b1b98e8fdb65214ef"} err="failed to get container status \"1809cc49214843841f512f2b18ee339fa99a9a75e7d0725b1b98e8fdb65214ef\": rpc error: code = NotFound desc = could not find container \"1809cc49214843841f512f2b18ee339fa99a9a75e7d0725b1b98e8fdb65214ef\": container with ID starting with 1809cc49214843841f512f2b18ee339fa99a9a75e7d0725b1b98e8fdb65214ef not found: ID does not exist" Dec 11 02:24:25 crc kubenswrapper[4824]: I1211 02:24:25.471328 4824 scope.go:117] "RemoveContainer" containerID="9b98e7345115867a4832021da7738c9d8e895fc4f81273785a40c60ab44ac9d1" Dec 11 02:24:25 crc kubenswrapper[4824]: E1211 02:24:25.471797 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9b98e7345115867a4832021da7738c9d8e895fc4f81273785a40c60ab44ac9d1\": container with ID starting with 9b98e7345115867a4832021da7738c9d8e895fc4f81273785a40c60ab44ac9d1 not found: ID does not exist" containerID="9b98e7345115867a4832021da7738c9d8e895fc4f81273785a40c60ab44ac9d1" Dec 11 02:24:25 crc kubenswrapper[4824]: I1211 02:24:25.471842 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b98e7345115867a4832021da7738c9d8e895fc4f81273785a40c60ab44ac9d1"} err="failed to get container status \"9b98e7345115867a4832021da7738c9d8e895fc4f81273785a40c60ab44ac9d1\": rpc error: code = NotFound desc = could not find container \"9b98e7345115867a4832021da7738c9d8e895fc4f81273785a40c60ab44ac9d1\": container with ID starting with 9b98e7345115867a4832021da7738c9d8e895fc4f81273785a40c60ab44ac9d1 not found: ID does not exist" Dec 11 02:24:25 crc kubenswrapper[4824]: I1211 02:24:25.474515 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 11 02:24:25 crc kubenswrapper[4824]: I1211 02:24:25.478242 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 11 02:24:25 crc kubenswrapper[4824]: I1211 02:24:25.479193 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 11 02:24:25 crc kubenswrapper[4824]: I1211 02:24:25.480324 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 11 02:24:25 crc kubenswrapper[4824]: I1211 02:24:25.583973 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f36004e0-4967-40f2-b291-e55fc0bf49b0-internal-tls-certs\") pod \"nova-api-0\" (UID: \"f36004e0-4967-40f2-b291-e55fc0bf49b0\") " pod="openstack/nova-api-0" Dec 11 02:24:25 crc kubenswrapper[4824]: I1211 02:24:25.584411 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2llf8\" (UniqueName: \"kubernetes.io/projected/f36004e0-4967-40f2-b291-e55fc0bf49b0-kube-api-access-2llf8\") pod \"nova-api-0\" (UID: \"f36004e0-4967-40f2-b291-e55fc0bf49b0\") " pod="openstack/nova-api-0" Dec 11 02:24:25 crc kubenswrapper[4824]: I1211 02:24:25.584560 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f36004e0-4967-40f2-b291-e55fc0bf49b0-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f36004e0-4967-40f2-b291-e55fc0bf49b0\") " pod="openstack/nova-api-0" Dec 11 02:24:25 crc kubenswrapper[4824]: I1211 02:24:25.584587 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f36004e0-4967-40f2-b291-e55fc0bf49b0-public-tls-certs\") pod \"nova-api-0\" (UID: \"f36004e0-4967-40f2-b291-e55fc0bf49b0\") " pod="openstack/nova-api-0" Dec 11 02:24:25 crc kubenswrapper[4824]: I1211 02:24:25.584611 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f36004e0-4967-40f2-b291-e55fc0bf49b0-config-data\") pod \"nova-api-0\" (UID: \"f36004e0-4967-40f2-b291-e55fc0bf49b0\") " pod="openstack/nova-api-0" Dec 11 02:24:25 crc kubenswrapper[4824]: I1211 02:24:25.584638 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f36004e0-4967-40f2-b291-e55fc0bf49b0-logs\") pod \"nova-api-0\" (UID: \"f36004e0-4967-40f2-b291-e55fc0bf49b0\") " pod="openstack/nova-api-0" Dec 11 02:24:25 crc kubenswrapper[4824]: I1211 02:24:25.686923 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2llf8\" (UniqueName: \"kubernetes.io/projected/f36004e0-4967-40f2-b291-e55fc0bf49b0-kube-api-access-2llf8\") pod \"nova-api-0\" (UID: \"f36004e0-4967-40f2-b291-e55fc0bf49b0\") " pod="openstack/nova-api-0" Dec 11 02:24:25 crc kubenswrapper[4824]: I1211 02:24:25.687125 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f36004e0-4967-40f2-b291-e55fc0bf49b0-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f36004e0-4967-40f2-b291-e55fc0bf49b0\") " pod="openstack/nova-api-0" Dec 11 02:24:25 crc kubenswrapper[4824]: I1211 02:24:25.687157 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f36004e0-4967-40f2-b291-e55fc0bf49b0-public-tls-certs\") pod \"nova-api-0\" (UID: \"f36004e0-4967-40f2-b291-e55fc0bf49b0\") " pod="openstack/nova-api-0" Dec 11 02:24:25 crc kubenswrapper[4824]: I1211 02:24:25.687182 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f36004e0-4967-40f2-b291-e55fc0bf49b0-config-data\") pod \"nova-api-0\" (UID: \"f36004e0-4967-40f2-b291-e55fc0bf49b0\") " pod="openstack/nova-api-0" Dec 11 02:24:25 crc kubenswrapper[4824]: I1211 02:24:25.687207 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f36004e0-4967-40f2-b291-e55fc0bf49b0-logs\") pod \"nova-api-0\" (UID: \"f36004e0-4967-40f2-b291-e55fc0bf49b0\") " pod="openstack/nova-api-0" Dec 11 02:24:25 crc kubenswrapper[4824]: I1211 02:24:25.687245 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f36004e0-4967-40f2-b291-e55fc0bf49b0-internal-tls-certs\") pod \"nova-api-0\" (UID: \"f36004e0-4967-40f2-b291-e55fc0bf49b0\") " pod="openstack/nova-api-0" Dec 11 02:24:25 crc kubenswrapper[4824]: I1211 02:24:25.693082 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f36004e0-4967-40f2-b291-e55fc0bf49b0-logs\") pod \"nova-api-0\" (UID: \"f36004e0-4967-40f2-b291-e55fc0bf49b0\") " pod="openstack/nova-api-0" Dec 11 02:24:25 crc kubenswrapper[4824]: I1211 02:24:25.714228 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f36004e0-4967-40f2-b291-e55fc0bf49b0-config-data\") pod \"nova-api-0\" (UID: \"f36004e0-4967-40f2-b291-e55fc0bf49b0\") " pod="openstack/nova-api-0" Dec 11 02:24:25 crc kubenswrapper[4824]: I1211 02:24:25.714263 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f36004e0-4967-40f2-b291-e55fc0bf49b0-internal-tls-certs\") pod \"nova-api-0\" (UID: \"f36004e0-4967-40f2-b291-e55fc0bf49b0\") " pod="openstack/nova-api-0" Dec 11 02:24:25 crc kubenswrapper[4824]: I1211 02:24:25.714898 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f36004e0-4967-40f2-b291-e55fc0bf49b0-public-tls-certs\") pod \"nova-api-0\" (UID: \"f36004e0-4967-40f2-b291-e55fc0bf49b0\") " pod="openstack/nova-api-0" Dec 11 02:24:25 crc kubenswrapper[4824]: I1211 02:24:25.715106 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f36004e0-4967-40f2-b291-e55fc0bf49b0-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f36004e0-4967-40f2-b291-e55fc0bf49b0\") " pod="openstack/nova-api-0" Dec 11 02:24:25 crc kubenswrapper[4824]: I1211 02:24:25.731985 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2llf8\" (UniqueName: \"kubernetes.io/projected/f36004e0-4967-40f2-b291-e55fc0bf49b0-kube-api-access-2llf8\") pod \"nova-api-0\" (UID: \"f36004e0-4967-40f2-b291-e55fc0bf49b0\") " pod="openstack/nova-api-0" Dec 11 02:24:25 crc kubenswrapper[4824]: I1211 02:24:25.798520 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 11 02:24:26 crc kubenswrapper[4824]: I1211 02:24:26.282828 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 11 02:24:26 crc kubenswrapper[4824]: I1211 02:24:26.408230 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c65e7e08-e393-4cd6-b17e-3d9e81783e39","Type":"ContainerStarted","Data":"9248b340ff0309c379679c03911f4193b985dd86cad07ee8994f8f88822b29c4"} Dec 11 02:24:26 crc kubenswrapper[4824]: I1211 02:24:26.411011 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f36004e0-4967-40f2-b291-e55fc0bf49b0","Type":"ContainerStarted","Data":"c3b31a326396e3b4ce025aee6a11f15b7107d549e6d85db9a3debaa6b7bbce3a"} Dec 11 02:24:26 crc kubenswrapper[4824]: I1211 02:24:26.652366 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="420e8783-43b4-43c7-ae74-50df3b388af1" path="/var/lib/kubelet/pods/420e8783-43b4-43c7-ae74-50df3b388af1/volumes" Dec 11 02:24:26 crc kubenswrapper[4824]: I1211 02:24:26.979102 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Dec 11 02:24:26 crc kubenswrapper[4824]: I1211 02:24:26.999078 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Dec 11 02:24:27 crc kubenswrapper[4824]: I1211 02:24:27.425702 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c65e7e08-e393-4cd6-b17e-3d9e81783e39","Type":"ContainerStarted","Data":"bacb95ae86b7a27604ad086a81a63cecc732578fee49413d4b8130a01311c811"} Dec 11 02:24:27 crc kubenswrapper[4824]: I1211 02:24:27.432131 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f36004e0-4967-40f2-b291-e55fc0bf49b0","Type":"ContainerStarted","Data":"e1f5282b07b947a76641d38f716e922be52d7b20881f7735b0a445c8789b887a"} Dec 11 02:24:27 crc kubenswrapper[4824]: I1211 02:24:27.432490 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f36004e0-4967-40f2-b291-e55fc0bf49b0","Type":"ContainerStarted","Data":"eae48b788f653486c7102a8163072c1986cb344f908e39f1e8f20f8afac50a79"} Dec 11 02:24:27 crc kubenswrapper[4824]: I1211 02:24:27.459681 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Dec 11 02:24:27 crc kubenswrapper[4824]: I1211 02:24:27.472986 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.472952122 podStartE2EDuration="2.472952122s" podCreationTimestamp="2025-12-11 02:24:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:24:27.469740252 +0000 UTC m=+1409.158777641" watchObservedRunningTime="2025-12-11 02:24:27.472952122 +0000 UTC m=+1409.161989511" Dec 11 02:24:27 crc kubenswrapper[4824]: I1211 02:24:27.672410 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-trgbn"] Dec 11 02:24:27 crc kubenswrapper[4824]: I1211 02:24:27.673746 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-trgbn" Dec 11 02:24:27 crc kubenswrapper[4824]: I1211 02:24:27.681512 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Dec 11 02:24:27 crc kubenswrapper[4824]: I1211 02:24:27.681702 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Dec 11 02:24:27 crc kubenswrapper[4824]: I1211 02:24:27.683643 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-trgbn"] Dec 11 02:24:27 crc kubenswrapper[4824]: I1211 02:24:27.833426 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/767d0d22-091d-4768-8fda-f51e6bac2504-config-data\") pod \"nova-cell1-cell-mapping-trgbn\" (UID: \"767d0d22-091d-4768-8fda-f51e6bac2504\") " pod="openstack/nova-cell1-cell-mapping-trgbn" Dec 11 02:24:27 crc kubenswrapper[4824]: I1211 02:24:27.833694 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gjrhg\" (UniqueName: \"kubernetes.io/projected/767d0d22-091d-4768-8fda-f51e6bac2504-kube-api-access-gjrhg\") pod \"nova-cell1-cell-mapping-trgbn\" (UID: \"767d0d22-091d-4768-8fda-f51e6bac2504\") " pod="openstack/nova-cell1-cell-mapping-trgbn" Dec 11 02:24:27 crc kubenswrapper[4824]: I1211 02:24:27.833738 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/767d0d22-091d-4768-8fda-f51e6bac2504-scripts\") pod \"nova-cell1-cell-mapping-trgbn\" (UID: \"767d0d22-091d-4768-8fda-f51e6bac2504\") " pod="openstack/nova-cell1-cell-mapping-trgbn" Dec 11 02:24:27 crc kubenswrapper[4824]: I1211 02:24:27.833817 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/767d0d22-091d-4768-8fda-f51e6bac2504-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-trgbn\" (UID: \"767d0d22-091d-4768-8fda-f51e6bac2504\") " pod="openstack/nova-cell1-cell-mapping-trgbn" Dec 11 02:24:27 crc kubenswrapper[4824]: I1211 02:24:27.935217 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/767d0d22-091d-4768-8fda-f51e6bac2504-config-data\") pod \"nova-cell1-cell-mapping-trgbn\" (UID: \"767d0d22-091d-4768-8fda-f51e6bac2504\") " pod="openstack/nova-cell1-cell-mapping-trgbn" Dec 11 02:24:27 crc kubenswrapper[4824]: I1211 02:24:27.935290 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gjrhg\" (UniqueName: \"kubernetes.io/projected/767d0d22-091d-4768-8fda-f51e6bac2504-kube-api-access-gjrhg\") pod \"nova-cell1-cell-mapping-trgbn\" (UID: \"767d0d22-091d-4768-8fda-f51e6bac2504\") " pod="openstack/nova-cell1-cell-mapping-trgbn" Dec 11 02:24:27 crc kubenswrapper[4824]: I1211 02:24:27.935329 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/767d0d22-091d-4768-8fda-f51e6bac2504-scripts\") pod \"nova-cell1-cell-mapping-trgbn\" (UID: \"767d0d22-091d-4768-8fda-f51e6bac2504\") " pod="openstack/nova-cell1-cell-mapping-trgbn" Dec 11 02:24:27 crc kubenswrapper[4824]: I1211 02:24:27.935445 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/767d0d22-091d-4768-8fda-f51e6bac2504-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-trgbn\" (UID: \"767d0d22-091d-4768-8fda-f51e6bac2504\") " pod="openstack/nova-cell1-cell-mapping-trgbn" Dec 11 02:24:27 crc kubenswrapper[4824]: I1211 02:24:27.941703 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/767d0d22-091d-4768-8fda-f51e6bac2504-scripts\") pod \"nova-cell1-cell-mapping-trgbn\" (UID: \"767d0d22-091d-4768-8fda-f51e6bac2504\") " pod="openstack/nova-cell1-cell-mapping-trgbn" Dec 11 02:24:27 crc kubenswrapper[4824]: I1211 02:24:27.943963 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/767d0d22-091d-4768-8fda-f51e6bac2504-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-trgbn\" (UID: \"767d0d22-091d-4768-8fda-f51e6bac2504\") " pod="openstack/nova-cell1-cell-mapping-trgbn" Dec 11 02:24:27 crc kubenswrapper[4824]: I1211 02:24:27.953081 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/767d0d22-091d-4768-8fda-f51e6bac2504-config-data\") pod \"nova-cell1-cell-mapping-trgbn\" (UID: \"767d0d22-091d-4768-8fda-f51e6bac2504\") " pod="openstack/nova-cell1-cell-mapping-trgbn" Dec 11 02:24:27 crc kubenswrapper[4824]: I1211 02:24:27.965554 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gjrhg\" (UniqueName: \"kubernetes.io/projected/767d0d22-091d-4768-8fda-f51e6bac2504-kube-api-access-gjrhg\") pod \"nova-cell1-cell-mapping-trgbn\" (UID: \"767d0d22-091d-4768-8fda-f51e6bac2504\") " pod="openstack/nova-cell1-cell-mapping-trgbn" Dec 11 02:24:28 crc kubenswrapper[4824]: I1211 02:24:28.006020 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-trgbn" Dec 11 02:24:28 crc kubenswrapper[4824]: I1211 02:24:28.443184 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c65e7e08-e393-4cd6-b17e-3d9e81783e39","Type":"ContainerStarted","Data":"bf07ad4097588519c67a81a0b3640dc5b46df844e156311004078dd0bfe4d72e"} Dec 11 02:24:28 crc kubenswrapper[4824]: I1211 02:24:28.444399 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 11 02:24:28 crc kubenswrapper[4824]: I1211 02:24:28.501803 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.617087878 podStartE2EDuration="5.501785084s" podCreationTimestamp="2025-12-11 02:24:23 +0000 UTC" firstStartedPulling="2025-12-11 02:24:24.074607358 +0000 UTC m=+1405.763644737" lastFinishedPulling="2025-12-11 02:24:27.959304524 +0000 UTC m=+1409.648341943" observedRunningTime="2025-12-11 02:24:28.477577066 +0000 UTC m=+1410.166614455" watchObservedRunningTime="2025-12-11 02:24:28.501785084 +0000 UTC m=+1410.190822473" Dec 11 02:24:28 crc kubenswrapper[4824]: I1211 02:24:28.507214 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-trgbn"] Dec 11 02:24:28 crc kubenswrapper[4824]: W1211 02:24:28.516363 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod767d0d22_091d_4768_8fda_f51e6bac2504.slice/crio-1d0e48afda179166c270f598cdd15fdf1e51490fd0bcea5c15bd90066fedd5cb WatchSource:0}: Error finding container 1d0e48afda179166c270f598cdd15fdf1e51490fd0bcea5c15bd90066fedd5cb: Status 404 returned error can't find the container with id 1d0e48afda179166c270f598cdd15fdf1e51490fd0bcea5c15bd90066fedd5cb Dec 11 02:24:28 crc kubenswrapper[4824]: I1211 02:24:28.891223 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-89c5cd4d5-n9sd5" Dec 11 02:24:28 crc kubenswrapper[4824]: I1211 02:24:28.961316 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-bjsxw"] Dec 11 02:24:28 crc kubenswrapper[4824]: I1211 02:24:28.961604 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-757b4f8459-bjsxw" podUID="aa930a8a-6891-4790-9c51-2165f75270c5" containerName="dnsmasq-dns" containerID="cri-o://11952cf39c9a35bfbe8a40322a6d5e89939669557153ae881d9eddb635214b4a" gracePeriod=10 Dec 11 02:24:29 crc kubenswrapper[4824]: I1211 02:24:29.404400 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-bjsxw" Dec 11 02:24:29 crc kubenswrapper[4824]: I1211 02:24:29.463752 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-trgbn" event={"ID":"767d0d22-091d-4768-8fda-f51e6bac2504","Type":"ContainerStarted","Data":"ecb08a99e948f9fdfcffdae254cf531db2878ad91679263a52a6ed6e24f48c97"} Dec 11 02:24:29 crc kubenswrapper[4824]: I1211 02:24:29.464028 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-trgbn" event={"ID":"767d0d22-091d-4768-8fda-f51e6bac2504","Type":"ContainerStarted","Data":"1d0e48afda179166c270f598cdd15fdf1e51490fd0bcea5c15bd90066fedd5cb"} Dec 11 02:24:29 crc kubenswrapper[4824]: I1211 02:24:29.467648 4824 generic.go:334] "Generic (PLEG): container finished" podID="aa930a8a-6891-4790-9c51-2165f75270c5" containerID="11952cf39c9a35bfbe8a40322a6d5e89939669557153ae881d9eddb635214b4a" exitCode=0 Dec 11 02:24:29 crc kubenswrapper[4824]: I1211 02:24:29.468277 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-bjsxw" Dec 11 02:24:29 crc kubenswrapper[4824]: I1211 02:24:29.468431 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-bjsxw" event={"ID":"aa930a8a-6891-4790-9c51-2165f75270c5","Type":"ContainerDied","Data":"11952cf39c9a35bfbe8a40322a6d5e89939669557153ae881d9eddb635214b4a"} Dec 11 02:24:29 crc kubenswrapper[4824]: I1211 02:24:29.468456 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-bjsxw" event={"ID":"aa930a8a-6891-4790-9c51-2165f75270c5","Type":"ContainerDied","Data":"108296ee90d9a63c87e2b07b29f3723b29c9ff3a59aa496123c325c0d21bd524"} Dec 11 02:24:29 crc kubenswrapper[4824]: I1211 02:24:29.468483 4824 scope.go:117] "RemoveContainer" containerID="11952cf39c9a35bfbe8a40322a6d5e89939669557153ae881d9eddb635214b4a" Dec 11 02:24:29 crc kubenswrapper[4824]: I1211 02:24:29.486548 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-trgbn" podStartSLOduration=2.486524079 podStartE2EDuration="2.486524079s" podCreationTimestamp="2025-12-11 02:24:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:24:29.476937127 +0000 UTC m=+1411.165974506" watchObservedRunningTime="2025-12-11 02:24:29.486524079 +0000 UTC m=+1411.175561468" Dec 11 02:24:29 crc kubenswrapper[4824]: I1211 02:24:29.505280 4824 scope.go:117] "RemoveContainer" containerID="044c445997a4ae9b4eb965768ef2443c7b5408a6b12599b224f8d29a366069b2" Dec 11 02:24:29 crc kubenswrapper[4824]: I1211 02:24:29.528425 4824 scope.go:117] "RemoveContainer" containerID="11952cf39c9a35bfbe8a40322a6d5e89939669557153ae881d9eddb635214b4a" Dec 11 02:24:29 crc kubenswrapper[4824]: E1211 02:24:29.531917 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11952cf39c9a35bfbe8a40322a6d5e89939669557153ae881d9eddb635214b4a\": container with ID starting with 11952cf39c9a35bfbe8a40322a6d5e89939669557153ae881d9eddb635214b4a not found: ID does not exist" containerID="11952cf39c9a35bfbe8a40322a6d5e89939669557153ae881d9eddb635214b4a" Dec 11 02:24:29 crc kubenswrapper[4824]: I1211 02:24:29.531946 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11952cf39c9a35bfbe8a40322a6d5e89939669557153ae881d9eddb635214b4a"} err="failed to get container status \"11952cf39c9a35bfbe8a40322a6d5e89939669557153ae881d9eddb635214b4a\": rpc error: code = NotFound desc = could not find container \"11952cf39c9a35bfbe8a40322a6d5e89939669557153ae881d9eddb635214b4a\": container with ID starting with 11952cf39c9a35bfbe8a40322a6d5e89939669557153ae881d9eddb635214b4a not found: ID does not exist" Dec 11 02:24:29 crc kubenswrapper[4824]: I1211 02:24:29.531964 4824 scope.go:117] "RemoveContainer" containerID="044c445997a4ae9b4eb965768ef2443c7b5408a6b12599b224f8d29a366069b2" Dec 11 02:24:29 crc kubenswrapper[4824]: E1211 02:24:29.537260 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"044c445997a4ae9b4eb965768ef2443c7b5408a6b12599b224f8d29a366069b2\": container with ID starting with 044c445997a4ae9b4eb965768ef2443c7b5408a6b12599b224f8d29a366069b2 not found: ID does not exist" containerID="044c445997a4ae9b4eb965768ef2443c7b5408a6b12599b224f8d29a366069b2" Dec 11 02:24:29 crc kubenswrapper[4824]: I1211 02:24:29.537292 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"044c445997a4ae9b4eb965768ef2443c7b5408a6b12599b224f8d29a366069b2"} err="failed to get container status \"044c445997a4ae9b4eb965768ef2443c7b5408a6b12599b224f8d29a366069b2\": rpc error: code = NotFound desc = could not find container \"044c445997a4ae9b4eb965768ef2443c7b5408a6b12599b224f8d29a366069b2\": container with ID starting with 044c445997a4ae9b4eb965768ef2443c7b5408a6b12599b224f8d29a366069b2 not found: ID does not exist" Dec 11 02:24:29 crc kubenswrapper[4824]: I1211 02:24:29.565722 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aa930a8a-6891-4790-9c51-2165f75270c5-dns-svc\") pod \"aa930a8a-6891-4790-9c51-2165f75270c5\" (UID: \"aa930a8a-6891-4790-9c51-2165f75270c5\") " Dec 11 02:24:29 crc kubenswrapper[4824]: I1211 02:24:29.565843 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa930a8a-6891-4790-9c51-2165f75270c5-config\") pod \"aa930a8a-6891-4790-9c51-2165f75270c5\" (UID: \"aa930a8a-6891-4790-9c51-2165f75270c5\") " Dec 11 02:24:29 crc kubenswrapper[4824]: I1211 02:24:29.565865 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aa930a8a-6891-4790-9c51-2165f75270c5-ovsdbserver-sb\") pod \"aa930a8a-6891-4790-9c51-2165f75270c5\" (UID: \"aa930a8a-6891-4790-9c51-2165f75270c5\") " Dec 11 02:24:29 crc kubenswrapper[4824]: I1211 02:24:29.565906 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d8tw9\" (UniqueName: \"kubernetes.io/projected/aa930a8a-6891-4790-9c51-2165f75270c5-kube-api-access-d8tw9\") pod \"aa930a8a-6891-4790-9c51-2165f75270c5\" (UID: \"aa930a8a-6891-4790-9c51-2165f75270c5\") " Dec 11 02:24:29 crc kubenswrapper[4824]: I1211 02:24:29.565949 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aa930a8a-6891-4790-9c51-2165f75270c5-dns-swift-storage-0\") pod \"aa930a8a-6891-4790-9c51-2165f75270c5\" (UID: \"aa930a8a-6891-4790-9c51-2165f75270c5\") " Dec 11 02:24:29 crc kubenswrapper[4824]: I1211 02:24:29.565992 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aa930a8a-6891-4790-9c51-2165f75270c5-ovsdbserver-nb\") pod \"aa930a8a-6891-4790-9c51-2165f75270c5\" (UID: \"aa930a8a-6891-4790-9c51-2165f75270c5\") " Dec 11 02:24:29 crc kubenswrapper[4824]: I1211 02:24:29.587386 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa930a8a-6891-4790-9c51-2165f75270c5-kube-api-access-d8tw9" (OuterVolumeSpecName: "kube-api-access-d8tw9") pod "aa930a8a-6891-4790-9c51-2165f75270c5" (UID: "aa930a8a-6891-4790-9c51-2165f75270c5"). InnerVolumeSpecName "kube-api-access-d8tw9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:24:29 crc kubenswrapper[4824]: I1211 02:24:29.623307 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa930a8a-6891-4790-9c51-2165f75270c5-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "aa930a8a-6891-4790-9c51-2165f75270c5" (UID: "aa930a8a-6891-4790-9c51-2165f75270c5"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:24:29 crc kubenswrapper[4824]: I1211 02:24:29.624681 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa930a8a-6891-4790-9c51-2165f75270c5-config" (OuterVolumeSpecName: "config") pod "aa930a8a-6891-4790-9c51-2165f75270c5" (UID: "aa930a8a-6891-4790-9c51-2165f75270c5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:24:29 crc kubenswrapper[4824]: I1211 02:24:29.653417 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa930a8a-6891-4790-9c51-2165f75270c5-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "aa930a8a-6891-4790-9c51-2165f75270c5" (UID: "aa930a8a-6891-4790-9c51-2165f75270c5"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:24:29 crc kubenswrapper[4824]: I1211 02:24:29.665332 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa930a8a-6891-4790-9c51-2165f75270c5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "aa930a8a-6891-4790-9c51-2165f75270c5" (UID: "aa930a8a-6891-4790-9c51-2165f75270c5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:24:29 crc kubenswrapper[4824]: I1211 02:24:29.668655 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aa930a8a-6891-4790-9c51-2165f75270c5-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 11 02:24:29 crc kubenswrapper[4824]: I1211 02:24:29.668689 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aa930a8a-6891-4790-9c51-2165f75270c5-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 11 02:24:29 crc kubenswrapper[4824]: I1211 02:24:29.668699 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa930a8a-6891-4790-9c51-2165f75270c5-config\") on node \"crc\" DevicePath \"\"" Dec 11 02:24:29 crc kubenswrapper[4824]: I1211 02:24:29.668708 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aa930a8a-6891-4790-9c51-2165f75270c5-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 11 02:24:29 crc kubenswrapper[4824]: I1211 02:24:29.668717 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d8tw9\" (UniqueName: \"kubernetes.io/projected/aa930a8a-6891-4790-9c51-2165f75270c5-kube-api-access-d8tw9\") on node \"crc\" DevicePath \"\"" Dec 11 02:24:29 crc kubenswrapper[4824]: I1211 02:24:29.675134 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa930a8a-6891-4790-9c51-2165f75270c5-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "aa930a8a-6891-4790-9c51-2165f75270c5" (UID: "aa930a8a-6891-4790-9c51-2165f75270c5"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:24:29 crc kubenswrapper[4824]: I1211 02:24:29.770377 4824 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aa930a8a-6891-4790-9c51-2165f75270c5-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 11 02:24:29 crc kubenswrapper[4824]: I1211 02:24:29.802988 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-bjsxw"] Dec 11 02:24:29 crc kubenswrapper[4824]: I1211 02:24:29.811429 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-bjsxw"] Dec 11 02:24:30 crc kubenswrapper[4824]: I1211 02:24:30.652918 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa930a8a-6891-4790-9c51-2165f75270c5" path="/var/lib/kubelet/pods/aa930a8a-6891-4790-9c51-2165f75270c5/volumes" Dec 11 02:24:33 crc kubenswrapper[4824]: I1211 02:24:33.526053 4824 generic.go:334] "Generic (PLEG): container finished" podID="767d0d22-091d-4768-8fda-f51e6bac2504" containerID="ecb08a99e948f9fdfcffdae254cf531db2878ad91679263a52a6ed6e24f48c97" exitCode=0 Dec 11 02:24:33 crc kubenswrapper[4824]: I1211 02:24:33.526153 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-trgbn" event={"ID":"767d0d22-091d-4768-8fda-f51e6bac2504","Type":"ContainerDied","Data":"ecb08a99e948f9fdfcffdae254cf531db2878ad91679263a52a6ed6e24f48c97"} Dec 11 02:24:35 crc kubenswrapper[4824]: I1211 02:24:35.031016 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-trgbn" Dec 11 02:24:35 crc kubenswrapper[4824]: I1211 02:24:35.095949 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gjrhg\" (UniqueName: \"kubernetes.io/projected/767d0d22-091d-4768-8fda-f51e6bac2504-kube-api-access-gjrhg\") pod \"767d0d22-091d-4768-8fda-f51e6bac2504\" (UID: \"767d0d22-091d-4768-8fda-f51e6bac2504\") " Dec 11 02:24:35 crc kubenswrapper[4824]: I1211 02:24:35.096118 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/767d0d22-091d-4768-8fda-f51e6bac2504-scripts\") pod \"767d0d22-091d-4768-8fda-f51e6bac2504\" (UID: \"767d0d22-091d-4768-8fda-f51e6bac2504\") " Dec 11 02:24:35 crc kubenswrapper[4824]: I1211 02:24:35.096183 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/767d0d22-091d-4768-8fda-f51e6bac2504-combined-ca-bundle\") pod \"767d0d22-091d-4768-8fda-f51e6bac2504\" (UID: \"767d0d22-091d-4768-8fda-f51e6bac2504\") " Dec 11 02:24:35 crc kubenswrapper[4824]: I1211 02:24:35.096239 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/767d0d22-091d-4768-8fda-f51e6bac2504-config-data\") pod \"767d0d22-091d-4768-8fda-f51e6bac2504\" (UID: \"767d0d22-091d-4768-8fda-f51e6bac2504\") " Dec 11 02:24:35 crc kubenswrapper[4824]: I1211 02:24:35.102664 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/767d0d22-091d-4768-8fda-f51e6bac2504-kube-api-access-gjrhg" (OuterVolumeSpecName: "kube-api-access-gjrhg") pod "767d0d22-091d-4768-8fda-f51e6bac2504" (UID: "767d0d22-091d-4768-8fda-f51e6bac2504"). InnerVolumeSpecName "kube-api-access-gjrhg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:24:35 crc kubenswrapper[4824]: I1211 02:24:35.112241 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/767d0d22-091d-4768-8fda-f51e6bac2504-scripts" (OuterVolumeSpecName: "scripts") pod "767d0d22-091d-4768-8fda-f51e6bac2504" (UID: "767d0d22-091d-4768-8fda-f51e6bac2504"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:24:35 crc kubenswrapper[4824]: I1211 02:24:35.127283 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/767d0d22-091d-4768-8fda-f51e6bac2504-config-data" (OuterVolumeSpecName: "config-data") pod "767d0d22-091d-4768-8fda-f51e6bac2504" (UID: "767d0d22-091d-4768-8fda-f51e6bac2504"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:24:35 crc kubenswrapper[4824]: I1211 02:24:35.138493 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/767d0d22-091d-4768-8fda-f51e6bac2504-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "767d0d22-091d-4768-8fda-f51e6bac2504" (UID: "767d0d22-091d-4768-8fda-f51e6bac2504"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:24:35 crc kubenswrapper[4824]: I1211 02:24:35.198353 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/767d0d22-091d-4768-8fda-f51e6bac2504-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 02:24:35 crc kubenswrapper[4824]: I1211 02:24:35.198408 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/767d0d22-091d-4768-8fda-f51e6bac2504-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 02:24:35 crc kubenswrapper[4824]: I1211 02:24:35.198428 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gjrhg\" (UniqueName: \"kubernetes.io/projected/767d0d22-091d-4768-8fda-f51e6bac2504-kube-api-access-gjrhg\") on node \"crc\" DevicePath \"\"" Dec 11 02:24:35 crc kubenswrapper[4824]: I1211 02:24:35.198447 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/767d0d22-091d-4768-8fda-f51e6bac2504-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 02:24:35 crc kubenswrapper[4824]: I1211 02:24:35.606403 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-trgbn" event={"ID":"767d0d22-091d-4768-8fda-f51e6bac2504","Type":"ContainerDied","Data":"1d0e48afda179166c270f598cdd15fdf1e51490fd0bcea5c15bd90066fedd5cb"} Dec 11 02:24:35 crc kubenswrapper[4824]: I1211 02:24:35.606514 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-trgbn" Dec 11 02:24:35 crc kubenswrapper[4824]: I1211 02:24:35.607750 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1d0e48afda179166c270f598cdd15fdf1e51490fd0bcea5c15bd90066fedd5cb" Dec 11 02:24:35 crc kubenswrapper[4824]: I1211 02:24:35.799868 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 11 02:24:35 crc kubenswrapper[4824]: I1211 02:24:35.799911 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 11 02:24:35 crc kubenswrapper[4824]: I1211 02:24:35.854302 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 11 02:24:35 crc kubenswrapper[4824]: I1211 02:24:35.865827 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 11 02:24:35 crc kubenswrapper[4824]: I1211 02:24:35.866365 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="cec2d4af-b251-47cf-aa71-b63852b09cce" containerName="nova-scheduler-scheduler" containerID="cri-o://f566f44056ef249cb33952ee8373c0ffc800a9261b52edb6c6987191226aed85" gracePeriod=30 Dec 11 02:24:35 crc kubenswrapper[4824]: I1211 02:24:35.875698 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 11 02:24:35 crc kubenswrapper[4824]: I1211 02:24:35.875943 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="2a3020fe-d2be-4c32-9f69-f5365c4a10c4" containerName="nova-metadata-log" containerID="cri-o://fc00c4365c55304300d9d6ffee4f4185a4318fb2439681904c5bcb773a9482cc" gracePeriod=30 Dec 11 02:24:35 crc kubenswrapper[4824]: I1211 02:24:35.876095 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="2a3020fe-d2be-4c32-9f69-f5365c4a10c4" containerName="nova-metadata-metadata" containerID="cri-o://1a7e5ed0325ecaba26d9852ff51ba1c50eb8facc2eafffc9eb9fb555445e6a28" gracePeriod=30 Dec 11 02:24:36 crc kubenswrapper[4824]: I1211 02:24:36.617583 4824 generic.go:334] "Generic (PLEG): container finished" podID="2a3020fe-d2be-4c32-9f69-f5365c4a10c4" containerID="fc00c4365c55304300d9d6ffee4f4185a4318fb2439681904c5bcb773a9482cc" exitCode=143 Dec 11 02:24:36 crc kubenswrapper[4824]: I1211 02:24:36.617669 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2a3020fe-d2be-4c32-9f69-f5365c4a10c4","Type":"ContainerDied","Data":"fc00c4365c55304300d9d6ffee4f4185a4318fb2439681904c5bcb773a9482cc"} Dec 11 02:24:36 crc kubenswrapper[4824]: I1211 02:24:36.617853 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="f36004e0-4967-40f2-b291-e55fc0bf49b0" containerName="nova-api-log" containerID="cri-o://eae48b788f653486c7102a8163072c1986cb344f908e39f1e8f20f8afac50a79" gracePeriod=30 Dec 11 02:24:36 crc kubenswrapper[4824]: I1211 02:24:36.617949 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="f36004e0-4967-40f2-b291-e55fc0bf49b0" containerName="nova-api-api" containerID="cri-o://e1f5282b07b947a76641d38f716e922be52d7b20881f7735b0a445c8789b887a" gracePeriod=30 Dec 11 02:24:36 crc kubenswrapper[4824]: I1211 02:24:36.623256 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="f36004e0-4967-40f2-b291-e55fc0bf49b0" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.203:8774/\": EOF" Dec 11 02:24:36 crc kubenswrapper[4824]: I1211 02:24:36.623254 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="f36004e0-4967-40f2-b291-e55fc0bf49b0" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.203:8774/\": EOF" Dec 11 02:24:37 crc kubenswrapper[4824]: I1211 02:24:37.297902 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 11 02:24:37 crc kubenswrapper[4824]: I1211 02:24:37.446558 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cec2d4af-b251-47cf-aa71-b63852b09cce-config-data\") pod \"cec2d4af-b251-47cf-aa71-b63852b09cce\" (UID: \"cec2d4af-b251-47cf-aa71-b63852b09cce\") " Dec 11 02:24:37 crc kubenswrapper[4824]: I1211 02:24:37.446620 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cec2d4af-b251-47cf-aa71-b63852b09cce-combined-ca-bundle\") pod \"cec2d4af-b251-47cf-aa71-b63852b09cce\" (UID: \"cec2d4af-b251-47cf-aa71-b63852b09cce\") " Dec 11 02:24:37 crc kubenswrapper[4824]: I1211 02:24:37.446800 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-62jnx\" (UniqueName: \"kubernetes.io/projected/cec2d4af-b251-47cf-aa71-b63852b09cce-kube-api-access-62jnx\") pod \"cec2d4af-b251-47cf-aa71-b63852b09cce\" (UID: \"cec2d4af-b251-47cf-aa71-b63852b09cce\") " Dec 11 02:24:37 crc kubenswrapper[4824]: I1211 02:24:37.453831 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cec2d4af-b251-47cf-aa71-b63852b09cce-kube-api-access-62jnx" (OuterVolumeSpecName: "kube-api-access-62jnx") pod "cec2d4af-b251-47cf-aa71-b63852b09cce" (UID: "cec2d4af-b251-47cf-aa71-b63852b09cce"). InnerVolumeSpecName "kube-api-access-62jnx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:24:37 crc kubenswrapper[4824]: I1211 02:24:37.480667 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cec2d4af-b251-47cf-aa71-b63852b09cce-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cec2d4af-b251-47cf-aa71-b63852b09cce" (UID: "cec2d4af-b251-47cf-aa71-b63852b09cce"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:24:37 crc kubenswrapper[4824]: I1211 02:24:37.495135 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cec2d4af-b251-47cf-aa71-b63852b09cce-config-data" (OuterVolumeSpecName: "config-data") pod "cec2d4af-b251-47cf-aa71-b63852b09cce" (UID: "cec2d4af-b251-47cf-aa71-b63852b09cce"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:24:37 crc kubenswrapper[4824]: I1211 02:24:37.548779 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cec2d4af-b251-47cf-aa71-b63852b09cce-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 02:24:37 crc kubenswrapper[4824]: I1211 02:24:37.548812 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cec2d4af-b251-47cf-aa71-b63852b09cce-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 02:24:37 crc kubenswrapper[4824]: I1211 02:24:37.548822 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-62jnx\" (UniqueName: \"kubernetes.io/projected/cec2d4af-b251-47cf-aa71-b63852b09cce-kube-api-access-62jnx\") on node \"crc\" DevicePath \"\"" Dec 11 02:24:37 crc kubenswrapper[4824]: I1211 02:24:37.631531 4824 generic.go:334] "Generic (PLEG): container finished" podID="f36004e0-4967-40f2-b291-e55fc0bf49b0" containerID="eae48b788f653486c7102a8163072c1986cb344f908e39f1e8f20f8afac50a79" exitCode=143 Dec 11 02:24:37 crc kubenswrapper[4824]: I1211 02:24:37.631591 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f36004e0-4967-40f2-b291-e55fc0bf49b0","Type":"ContainerDied","Data":"eae48b788f653486c7102a8163072c1986cb344f908e39f1e8f20f8afac50a79"} Dec 11 02:24:37 crc kubenswrapper[4824]: I1211 02:24:37.633654 4824 generic.go:334] "Generic (PLEG): container finished" podID="cec2d4af-b251-47cf-aa71-b63852b09cce" containerID="f566f44056ef249cb33952ee8373c0ffc800a9261b52edb6c6987191226aed85" exitCode=0 Dec 11 02:24:37 crc kubenswrapper[4824]: I1211 02:24:37.633681 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"cec2d4af-b251-47cf-aa71-b63852b09cce","Type":"ContainerDied","Data":"f566f44056ef249cb33952ee8373c0ffc800a9261b52edb6c6987191226aed85"} Dec 11 02:24:37 crc kubenswrapper[4824]: I1211 02:24:37.633696 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"cec2d4af-b251-47cf-aa71-b63852b09cce","Type":"ContainerDied","Data":"7ab6590841d97d01d5b7371461fcc8ab0c9361f245ae70c1ea2f51dbb701e579"} Dec 11 02:24:37 crc kubenswrapper[4824]: I1211 02:24:37.633712 4824 scope.go:117] "RemoveContainer" containerID="f566f44056ef249cb33952ee8373c0ffc800a9261b52edb6c6987191226aed85" Dec 11 02:24:37 crc kubenswrapper[4824]: I1211 02:24:37.633798 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 11 02:24:37 crc kubenswrapper[4824]: I1211 02:24:37.679442 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 11 02:24:37 crc kubenswrapper[4824]: I1211 02:24:37.681253 4824 scope.go:117] "RemoveContainer" containerID="f566f44056ef249cb33952ee8373c0ffc800a9261b52edb6c6987191226aed85" Dec 11 02:24:37 crc kubenswrapper[4824]: E1211 02:24:37.683137 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f566f44056ef249cb33952ee8373c0ffc800a9261b52edb6c6987191226aed85\": container with ID starting with f566f44056ef249cb33952ee8373c0ffc800a9261b52edb6c6987191226aed85 not found: ID does not exist" containerID="f566f44056ef249cb33952ee8373c0ffc800a9261b52edb6c6987191226aed85" Dec 11 02:24:37 crc kubenswrapper[4824]: I1211 02:24:37.683174 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f566f44056ef249cb33952ee8373c0ffc800a9261b52edb6c6987191226aed85"} err="failed to get container status \"f566f44056ef249cb33952ee8373c0ffc800a9261b52edb6c6987191226aed85\": rpc error: code = NotFound desc = could not find container \"f566f44056ef249cb33952ee8373c0ffc800a9261b52edb6c6987191226aed85\": container with ID starting with f566f44056ef249cb33952ee8373c0ffc800a9261b52edb6c6987191226aed85 not found: ID does not exist" Dec 11 02:24:37 crc kubenswrapper[4824]: I1211 02:24:37.690211 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 11 02:24:37 crc kubenswrapper[4824]: I1211 02:24:37.703397 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 11 02:24:37 crc kubenswrapper[4824]: E1211 02:24:37.703761 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="767d0d22-091d-4768-8fda-f51e6bac2504" containerName="nova-manage" Dec 11 02:24:37 crc kubenswrapper[4824]: I1211 02:24:37.703781 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="767d0d22-091d-4768-8fda-f51e6bac2504" containerName="nova-manage" Dec 11 02:24:37 crc kubenswrapper[4824]: E1211 02:24:37.703800 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cec2d4af-b251-47cf-aa71-b63852b09cce" containerName="nova-scheduler-scheduler" Dec 11 02:24:37 crc kubenswrapper[4824]: I1211 02:24:37.703807 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="cec2d4af-b251-47cf-aa71-b63852b09cce" containerName="nova-scheduler-scheduler" Dec 11 02:24:37 crc kubenswrapper[4824]: E1211 02:24:37.703816 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa930a8a-6891-4790-9c51-2165f75270c5" containerName="dnsmasq-dns" Dec 11 02:24:37 crc kubenswrapper[4824]: I1211 02:24:37.703823 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa930a8a-6891-4790-9c51-2165f75270c5" containerName="dnsmasq-dns" Dec 11 02:24:37 crc kubenswrapper[4824]: E1211 02:24:37.703852 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa930a8a-6891-4790-9c51-2165f75270c5" containerName="init" Dec 11 02:24:37 crc kubenswrapper[4824]: I1211 02:24:37.703858 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa930a8a-6891-4790-9c51-2165f75270c5" containerName="init" Dec 11 02:24:37 crc kubenswrapper[4824]: I1211 02:24:37.704044 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="767d0d22-091d-4768-8fda-f51e6bac2504" containerName="nova-manage" Dec 11 02:24:37 crc kubenswrapper[4824]: I1211 02:24:37.704064 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="cec2d4af-b251-47cf-aa71-b63852b09cce" containerName="nova-scheduler-scheduler" Dec 11 02:24:37 crc kubenswrapper[4824]: I1211 02:24:37.704074 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa930a8a-6891-4790-9c51-2165f75270c5" containerName="dnsmasq-dns" Dec 11 02:24:37 crc kubenswrapper[4824]: I1211 02:24:37.704702 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 11 02:24:37 crc kubenswrapper[4824]: I1211 02:24:37.706708 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 11 02:24:37 crc kubenswrapper[4824]: I1211 02:24:37.714119 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 11 02:24:37 crc kubenswrapper[4824]: I1211 02:24:37.853884 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc3713cb-287a-4350-a147-e13a5151ac71-config-data\") pod \"nova-scheduler-0\" (UID: \"bc3713cb-287a-4350-a147-e13a5151ac71\") " pod="openstack/nova-scheduler-0" Dec 11 02:24:37 crc kubenswrapper[4824]: I1211 02:24:37.853985 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vnqj5\" (UniqueName: \"kubernetes.io/projected/bc3713cb-287a-4350-a147-e13a5151ac71-kube-api-access-vnqj5\") pod \"nova-scheduler-0\" (UID: \"bc3713cb-287a-4350-a147-e13a5151ac71\") " pod="openstack/nova-scheduler-0" Dec 11 02:24:37 crc kubenswrapper[4824]: I1211 02:24:37.854020 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc3713cb-287a-4350-a147-e13a5151ac71-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"bc3713cb-287a-4350-a147-e13a5151ac71\") " pod="openstack/nova-scheduler-0" Dec 11 02:24:37 crc kubenswrapper[4824]: I1211 02:24:37.956327 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vnqj5\" (UniqueName: \"kubernetes.io/projected/bc3713cb-287a-4350-a147-e13a5151ac71-kube-api-access-vnqj5\") pod \"nova-scheduler-0\" (UID: \"bc3713cb-287a-4350-a147-e13a5151ac71\") " pod="openstack/nova-scheduler-0" Dec 11 02:24:37 crc kubenswrapper[4824]: I1211 02:24:37.956400 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc3713cb-287a-4350-a147-e13a5151ac71-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"bc3713cb-287a-4350-a147-e13a5151ac71\") " pod="openstack/nova-scheduler-0" Dec 11 02:24:37 crc kubenswrapper[4824]: I1211 02:24:37.956575 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc3713cb-287a-4350-a147-e13a5151ac71-config-data\") pod \"nova-scheduler-0\" (UID: \"bc3713cb-287a-4350-a147-e13a5151ac71\") " pod="openstack/nova-scheduler-0" Dec 11 02:24:37 crc kubenswrapper[4824]: I1211 02:24:37.961741 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc3713cb-287a-4350-a147-e13a5151ac71-config-data\") pod \"nova-scheduler-0\" (UID: \"bc3713cb-287a-4350-a147-e13a5151ac71\") " pod="openstack/nova-scheduler-0" Dec 11 02:24:37 crc kubenswrapper[4824]: I1211 02:24:37.962746 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc3713cb-287a-4350-a147-e13a5151ac71-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"bc3713cb-287a-4350-a147-e13a5151ac71\") " pod="openstack/nova-scheduler-0" Dec 11 02:24:37 crc kubenswrapper[4824]: I1211 02:24:37.983229 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vnqj5\" (UniqueName: \"kubernetes.io/projected/bc3713cb-287a-4350-a147-e13a5151ac71-kube-api-access-vnqj5\") pod \"nova-scheduler-0\" (UID: \"bc3713cb-287a-4350-a147-e13a5151ac71\") " pod="openstack/nova-scheduler-0" Dec 11 02:24:38 crc kubenswrapper[4824]: I1211 02:24:38.023016 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 11 02:24:38 crc kubenswrapper[4824]: W1211 02:24:38.528579 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbc3713cb_287a_4350_a147_e13a5151ac71.slice/crio-207f0a8c7946dd00c6fb1ab27b8db9b1e8f763cd278a5b63811cf25557ef6264 WatchSource:0}: Error finding container 207f0a8c7946dd00c6fb1ab27b8db9b1e8f763cd278a5b63811cf25557ef6264: Status 404 returned error can't find the container with id 207f0a8c7946dd00c6fb1ab27b8db9b1e8f763cd278a5b63811cf25557ef6264 Dec 11 02:24:38 crc kubenswrapper[4824]: I1211 02:24:38.532851 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 11 02:24:38 crc kubenswrapper[4824]: I1211 02:24:38.657984 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cec2d4af-b251-47cf-aa71-b63852b09cce" path="/var/lib/kubelet/pods/cec2d4af-b251-47cf-aa71-b63852b09cce/volumes" Dec 11 02:24:38 crc kubenswrapper[4824]: I1211 02:24:38.665733 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"bc3713cb-287a-4350-a147-e13a5151ac71","Type":"ContainerStarted","Data":"207f0a8c7946dd00c6fb1ab27b8db9b1e8f763cd278a5b63811cf25557ef6264"} Dec 11 02:24:39 crc kubenswrapper[4824]: I1211 02:24:39.020270 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="2a3020fe-d2be-4c32-9f69-f5365c4a10c4" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.194:8775/\": read tcp 10.217.0.2:46254->10.217.0.194:8775: read: connection reset by peer" Dec 11 02:24:39 crc kubenswrapper[4824]: I1211 02:24:39.020389 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="2a3020fe-d2be-4c32-9f69-f5365c4a10c4" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.194:8775/\": read tcp 10.217.0.2:46244->10.217.0.194:8775: read: connection reset by peer" Dec 11 02:24:39 crc kubenswrapper[4824]: I1211 02:24:39.635634 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 11 02:24:39 crc kubenswrapper[4824]: I1211 02:24:39.692956 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2a3020fe-d2be-4c32-9f69-f5365c4a10c4-logs\") pod \"2a3020fe-d2be-4c32-9f69-f5365c4a10c4\" (UID: \"2a3020fe-d2be-4c32-9f69-f5365c4a10c4\") " Dec 11 02:24:39 crc kubenswrapper[4824]: I1211 02:24:39.693208 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/2a3020fe-d2be-4c32-9f69-f5365c4a10c4-nova-metadata-tls-certs\") pod \"2a3020fe-d2be-4c32-9f69-f5365c4a10c4\" (UID: \"2a3020fe-d2be-4c32-9f69-f5365c4a10c4\") " Dec 11 02:24:39 crc kubenswrapper[4824]: I1211 02:24:39.693265 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5p9xp\" (UniqueName: \"kubernetes.io/projected/2a3020fe-d2be-4c32-9f69-f5365c4a10c4-kube-api-access-5p9xp\") pod \"2a3020fe-d2be-4c32-9f69-f5365c4a10c4\" (UID: \"2a3020fe-d2be-4c32-9f69-f5365c4a10c4\") " Dec 11 02:24:39 crc kubenswrapper[4824]: I1211 02:24:39.693304 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a3020fe-d2be-4c32-9f69-f5365c4a10c4-combined-ca-bundle\") pod \"2a3020fe-d2be-4c32-9f69-f5365c4a10c4\" (UID: \"2a3020fe-d2be-4c32-9f69-f5365c4a10c4\") " Dec 11 02:24:39 crc kubenswrapper[4824]: I1211 02:24:39.693378 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a3020fe-d2be-4c32-9f69-f5365c4a10c4-config-data\") pod \"2a3020fe-d2be-4c32-9f69-f5365c4a10c4\" (UID: \"2a3020fe-d2be-4c32-9f69-f5365c4a10c4\") " Dec 11 02:24:39 crc kubenswrapper[4824]: I1211 02:24:39.713668 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2a3020fe-d2be-4c32-9f69-f5365c4a10c4-logs" (OuterVolumeSpecName: "logs") pod "2a3020fe-d2be-4c32-9f69-f5365c4a10c4" (UID: "2a3020fe-d2be-4c32-9f69-f5365c4a10c4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:24:39 crc kubenswrapper[4824]: I1211 02:24:39.724032 4824 generic.go:334] "Generic (PLEG): container finished" podID="2a3020fe-d2be-4c32-9f69-f5365c4a10c4" containerID="1a7e5ed0325ecaba26d9852ff51ba1c50eb8facc2eafffc9eb9fb555445e6a28" exitCode=0 Dec 11 02:24:39 crc kubenswrapper[4824]: I1211 02:24:39.724154 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2a3020fe-d2be-4c32-9f69-f5365c4a10c4","Type":"ContainerDied","Data":"1a7e5ed0325ecaba26d9852ff51ba1c50eb8facc2eafffc9eb9fb555445e6a28"} Dec 11 02:24:39 crc kubenswrapper[4824]: I1211 02:24:39.724187 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2a3020fe-d2be-4c32-9f69-f5365c4a10c4","Type":"ContainerDied","Data":"2f4adeabdb58f8f9ce02c9a1035b476dcd67014565ce40d73f7bb961625a3b4a"} Dec 11 02:24:39 crc kubenswrapper[4824]: I1211 02:24:39.724209 4824 scope.go:117] "RemoveContainer" containerID="1a7e5ed0325ecaba26d9852ff51ba1c50eb8facc2eafffc9eb9fb555445e6a28" Dec 11 02:24:39 crc kubenswrapper[4824]: I1211 02:24:39.724340 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 11 02:24:39 crc kubenswrapper[4824]: I1211 02:24:39.730855 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"bc3713cb-287a-4350-a147-e13a5151ac71","Type":"ContainerStarted","Data":"7f0c9dc8d5853655ec1f3f9008dce032df47c301da65ea6c68de3a9dd666b468"} Dec 11 02:24:39 crc kubenswrapper[4824]: I1211 02:24:39.731329 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a3020fe-d2be-4c32-9f69-f5365c4a10c4-kube-api-access-5p9xp" (OuterVolumeSpecName: "kube-api-access-5p9xp") pod "2a3020fe-d2be-4c32-9f69-f5365c4a10c4" (UID: "2a3020fe-d2be-4c32-9f69-f5365c4a10c4"). InnerVolumeSpecName "kube-api-access-5p9xp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:24:39 crc kubenswrapper[4824]: I1211 02:24:39.746405 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a3020fe-d2be-4c32-9f69-f5365c4a10c4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2a3020fe-d2be-4c32-9f69-f5365c4a10c4" (UID: "2a3020fe-d2be-4c32-9f69-f5365c4a10c4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:24:39 crc kubenswrapper[4824]: I1211 02:24:39.762667 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.762647907 podStartE2EDuration="2.762647907s" podCreationTimestamp="2025-12-11 02:24:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:24:39.761022766 +0000 UTC m=+1421.450060145" watchObservedRunningTime="2025-12-11 02:24:39.762647907 +0000 UTC m=+1421.451685296" Dec 11 02:24:39 crc kubenswrapper[4824]: I1211 02:24:39.768792 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a3020fe-d2be-4c32-9f69-f5365c4a10c4-config-data" (OuterVolumeSpecName: "config-data") pod "2a3020fe-d2be-4c32-9f69-f5365c4a10c4" (UID: "2a3020fe-d2be-4c32-9f69-f5365c4a10c4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:24:39 crc kubenswrapper[4824]: I1211 02:24:39.795661 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5p9xp\" (UniqueName: \"kubernetes.io/projected/2a3020fe-d2be-4c32-9f69-f5365c4a10c4-kube-api-access-5p9xp\") on node \"crc\" DevicePath \"\"" Dec 11 02:24:39 crc kubenswrapper[4824]: I1211 02:24:39.795685 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a3020fe-d2be-4c32-9f69-f5365c4a10c4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 02:24:39 crc kubenswrapper[4824]: I1211 02:24:39.795694 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a3020fe-d2be-4c32-9f69-f5365c4a10c4-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 02:24:39 crc kubenswrapper[4824]: I1211 02:24:39.795702 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2a3020fe-d2be-4c32-9f69-f5365c4a10c4-logs\") on node \"crc\" DevicePath \"\"" Dec 11 02:24:39 crc kubenswrapper[4824]: I1211 02:24:39.800266 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a3020fe-d2be-4c32-9f69-f5365c4a10c4-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "2a3020fe-d2be-4c32-9f69-f5365c4a10c4" (UID: "2a3020fe-d2be-4c32-9f69-f5365c4a10c4"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:24:39 crc kubenswrapper[4824]: I1211 02:24:39.843232 4824 scope.go:117] "RemoveContainer" containerID="fc00c4365c55304300d9d6ffee4f4185a4318fb2439681904c5bcb773a9482cc" Dec 11 02:24:39 crc kubenswrapper[4824]: I1211 02:24:39.871053 4824 scope.go:117] "RemoveContainer" containerID="1a7e5ed0325ecaba26d9852ff51ba1c50eb8facc2eafffc9eb9fb555445e6a28" Dec 11 02:24:39 crc kubenswrapper[4824]: E1211 02:24:39.872360 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a7e5ed0325ecaba26d9852ff51ba1c50eb8facc2eafffc9eb9fb555445e6a28\": container with ID starting with 1a7e5ed0325ecaba26d9852ff51ba1c50eb8facc2eafffc9eb9fb555445e6a28 not found: ID does not exist" containerID="1a7e5ed0325ecaba26d9852ff51ba1c50eb8facc2eafffc9eb9fb555445e6a28" Dec 11 02:24:39 crc kubenswrapper[4824]: I1211 02:24:39.872494 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a7e5ed0325ecaba26d9852ff51ba1c50eb8facc2eafffc9eb9fb555445e6a28"} err="failed to get container status \"1a7e5ed0325ecaba26d9852ff51ba1c50eb8facc2eafffc9eb9fb555445e6a28\": rpc error: code = NotFound desc = could not find container \"1a7e5ed0325ecaba26d9852ff51ba1c50eb8facc2eafffc9eb9fb555445e6a28\": container with ID starting with 1a7e5ed0325ecaba26d9852ff51ba1c50eb8facc2eafffc9eb9fb555445e6a28 not found: ID does not exist" Dec 11 02:24:39 crc kubenswrapper[4824]: I1211 02:24:39.872517 4824 scope.go:117] "RemoveContainer" containerID="fc00c4365c55304300d9d6ffee4f4185a4318fb2439681904c5bcb773a9482cc" Dec 11 02:24:39 crc kubenswrapper[4824]: E1211 02:24:39.872999 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc00c4365c55304300d9d6ffee4f4185a4318fb2439681904c5bcb773a9482cc\": container with ID starting with fc00c4365c55304300d9d6ffee4f4185a4318fb2439681904c5bcb773a9482cc not found: ID does not exist" containerID="fc00c4365c55304300d9d6ffee4f4185a4318fb2439681904c5bcb773a9482cc" Dec 11 02:24:39 crc kubenswrapper[4824]: I1211 02:24:39.873036 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc00c4365c55304300d9d6ffee4f4185a4318fb2439681904c5bcb773a9482cc"} err="failed to get container status \"fc00c4365c55304300d9d6ffee4f4185a4318fb2439681904c5bcb773a9482cc\": rpc error: code = NotFound desc = could not find container \"fc00c4365c55304300d9d6ffee4f4185a4318fb2439681904c5bcb773a9482cc\": container with ID starting with fc00c4365c55304300d9d6ffee4f4185a4318fb2439681904c5bcb773a9482cc not found: ID does not exist" Dec 11 02:24:39 crc kubenswrapper[4824]: I1211 02:24:39.897312 4824 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/2a3020fe-d2be-4c32-9f69-f5365c4a10c4-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 11 02:24:40 crc kubenswrapper[4824]: I1211 02:24:40.071294 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 11 02:24:40 crc kubenswrapper[4824]: I1211 02:24:40.087206 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 11 02:24:40 crc kubenswrapper[4824]: I1211 02:24:40.098238 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 11 02:24:40 crc kubenswrapper[4824]: E1211 02:24:40.098732 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a3020fe-d2be-4c32-9f69-f5365c4a10c4" containerName="nova-metadata-log" Dec 11 02:24:40 crc kubenswrapper[4824]: I1211 02:24:40.098750 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a3020fe-d2be-4c32-9f69-f5365c4a10c4" containerName="nova-metadata-log" Dec 11 02:24:40 crc kubenswrapper[4824]: E1211 02:24:40.098785 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a3020fe-d2be-4c32-9f69-f5365c4a10c4" containerName="nova-metadata-metadata" Dec 11 02:24:40 crc kubenswrapper[4824]: I1211 02:24:40.098792 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a3020fe-d2be-4c32-9f69-f5365c4a10c4" containerName="nova-metadata-metadata" Dec 11 02:24:40 crc kubenswrapper[4824]: I1211 02:24:40.098974 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a3020fe-d2be-4c32-9f69-f5365c4a10c4" containerName="nova-metadata-log" Dec 11 02:24:40 crc kubenswrapper[4824]: I1211 02:24:40.098986 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a3020fe-d2be-4c32-9f69-f5365c4a10c4" containerName="nova-metadata-metadata" Dec 11 02:24:40 crc kubenswrapper[4824]: I1211 02:24:40.099891 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 11 02:24:40 crc kubenswrapper[4824]: I1211 02:24:40.101563 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 11 02:24:40 crc kubenswrapper[4824]: I1211 02:24:40.105569 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 11 02:24:40 crc kubenswrapper[4824]: I1211 02:24:40.119910 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 11 02:24:40 crc kubenswrapper[4824]: I1211 02:24:40.202623 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0e1ab61-fe6b-4b2e-aa01-2bf77e119ce4-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e0e1ab61-fe6b-4b2e-aa01-2bf77e119ce4\") " pod="openstack/nova-metadata-0" Dec 11 02:24:40 crc kubenswrapper[4824]: I1211 02:24:40.202711 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e0e1ab61-fe6b-4b2e-aa01-2bf77e119ce4-logs\") pod \"nova-metadata-0\" (UID: \"e0e1ab61-fe6b-4b2e-aa01-2bf77e119ce4\") " pod="openstack/nova-metadata-0" Dec 11 02:24:40 crc kubenswrapper[4824]: I1211 02:24:40.202740 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f8mng\" (UniqueName: \"kubernetes.io/projected/e0e1ab61-fe6b-4b2e-aa01-2bf77e119ce4-kube-api-access-f8mng\") pod \"nova-metadata-0\" (UID: \"e0e1ab61-fe6b-4b2e-aa01-2bf77e119ce4\") " pod="openstack/nova-metadata-0" Dec 11 02:24:40 crc kubenswrapper[4824]: I1211 02:24:40.202781 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0e1ab61-fe6b-4b2e-aa01-2bf77e119ce4-config-data\") pod \"nova-metadata-0\" (UID: \"e0e1ab61-fe6b-4b2e-aa01-2bf77e119ce4\") " pod="openstack/nova-metadata-0" Dec 11 02:24:40 crc kubenswrapper[4824]: I1211 02:24:40.202846 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e0e1ab61-fe6b-4b2e-aa01-2bf77e119ce4-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"e0e1ab61-fe6b-4b2e-aa01-2bf77e119ce4\") " pod="openstack/nova-metadata-0" Dec 11 02:24:40 crc kubenswrapper[4824]: I1211 02:24:40.304347 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e0e1ab61-fe6b-4b2e-aa01-2bf77e119ce4-logs\") pod \"nova-metadata-0\" (UID: \"e0e1ab61-fe6b-4b2e-aa01-2bf77e119ce4\") " pod="openstack/nova-metadata-0" Dec 11 02:24:40 crc kubenswrapper[4824]: I1211 02:24:40.304424 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f8mng\" (UniqueName: \"kubernetes.io/projected/e0e1ab61-fe6b-4b2e-aa01-2bf77e119ce4-kube-api-access-f8mng\") pod \"nova-metadata-0\" (UID: \"e0e1ab61-fe6b-4b2e-aa01-2bf77e119ce4\") " pod="openstack/nova-metadata-0" Dec 11 02:24:40 crc kubenswrapper[4824]: I1211 02:24:40.304514 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0e1ab61-fe6b-4b2e-aa01-2bf77e119ce4-config-data\") pod \"nova-metadata-0\" (UID: \"e0e1ab61-fe6b-4b2e-aa01-2bf77e119ce4\") " pod="openstack/nova-metadata-0" Dec 11 02:24:40 crc kubenswrapper[4824]: I1211 02:24:40.304592 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e0e1ab61-fe6b-4b2e-aa01-2bf77e119ce4-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"e0e1ab61-fe6b-4b2e-aa01-2bf77e119ce4\") " pod="openstack/nova-metadata-0" Dec 11 02:24:40 crc kubenswrapper[4824]: I1211 02:24:40.304706 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0e1ab61-fe6b-4b2e-aa01-2bf77e119ce4-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e0e1ab61-fe6b-4b2e-aa01-2bf77e119ce4\") " pod="openstack/nova-metadata-0" Dec 11 02:24:40 crc kubenswrapper[4824]: I1211 02:24:40.305091 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e0e1ab61-fe6b-4b2e-aa01-2bf77e119ce4-logs\") pod \"nova-metadata-0\" (UID: \"e0e1ab61-fe6b-4b2e-aa01-2bf77e119ce4\") " pod="openstack/nova-metadata-0" Dec 11 02:24:40 crc kubenswrapper[4824]: I1211 02:24:40.309952 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0e1ab61-fe6b-4b2e-aa01-2bf77e119ce4-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e0e1ab61-fe6b-4b2e-aa01-2bf77e119ce4\") " pod="openstack/nova-metadata-0" Dec 11 02:24:40 crc kubenswrapper[4824]: I1211 02:24:40.311015 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e0e1ab61-fe6b-4b2e-aa01-2bf77e119ce4-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"e0e1ab61-fe6b-4b2e-aa01-2bf77e119ce4\") " pod="openstack/nova-metadata-0" Dec 11 02:24:40 crc kubenswrapper[4824]: I1211 02:24:40.312575 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0e1ab61-fe6b-4b2e-aa01-2bf77e119ce4-config-data\") pod \"nova-metadata-0\" (UID: \"e0e1ab61-fe6b-4b2e-aa01-2bf77e119ce4\") " pod="openstack/nova-metadata-0" Dec 11 02:24:40 crc kubenswrapper[4824]: I1211 02:24:40.339612 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f8mng\" (UniqueName: \"kubernetes.io/projected/e0e1ab61-fe6b-4b2e-aa01-2bf77e119ce4-kube-api-access-f8mng\") pod \"nova-metadata-0\" (UID: \"e0e1ab61-fe6b-4b2e-aa01-2bf77e119ce4\") " pod="openstack/nova-metadata-0" Dec 11 02:24:40 crc kubenswrapper[4824]: I1211 02:24:40.420544 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 11 02:24:40 crc kubenswrapper[4824]: I1211 02:24:40.641656 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a3020fe-d2be-4c32-9f69-f5365c4a10c4" path="/var/lib/kubelet/pods/2a3020fe-d2be-4c32-9f69-f5365c4a10c4/volumes" Dec 11 02:24:40 crc kubenswrapper[4824]: I1211 02:24:40.772806 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 11 02:24:41 crc kubenswrapper[4824]: I1211 02:24:41.781449 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e0e1ab61-fe6b-4b2e-aa01-2bf77e119ce4","Type":"ContainerStarted","Data":"34310bde598ebe4aaf36f4deb046e36e92cde1751402cce781219d1c9d2a02d6"} Dec 11 02:24:41 crc kubenswrapper[4824]: I1211 02:24:41.781962 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e0e1ab61-fe6b-4b2e-aa01-2bf77e119ce4","Type":"ContainerStarted","Data":"adb4312e7d83c381c6d1093d97422fdbb8a1e083a91c716f449e8b3ed9ef7281"} Dec 11 02:24:41 crc kubenswrapper[4824]: I1211 02:24:41.781988 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e0e1ab61-fe6b-4b2e-aa01-2bf77e119ce4","Type":"ContainerStarted","Data":"e4a563bf87ec6f36f987ddaacb109a82d91f41137e177807014cb1e38b82ff3d"} Dec 11 02:24:41 crc kubenswrapper[4824]: I1211 02:24:41.815623 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=1.815597902 podStartE2EDuration="1.815597902s" podCreationTimestamp="2025-12-11 02:24:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:24:41.803937209 +0000 UTC m=+1423.492974628" watchObservedRunningTime="2025-12-11 02:24:41.815597902 +0000 UTC m=+1423.504635291" Dec 11 02:24:42 crc kubenswrapper[4824]: I1211 02:24:42.601597 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 11 02:24:42 crc kubenswrapper[4824]: I1211 02:24:42.752946 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f36004e0-4967-40f2-b291-e55fc0bf49b0-internal-tls-certs\") pod \"f36004e0-4967-40f2-b291-e55fc0bf49b0\" (UID: \"f36004e0-4967-40f2-b291-e55fc0bf49b0\") " Dec 11 02:24:42 crc kubenswrapper[4824]: I1211 02:24:42.753054 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f36004e0-4967-40f2-b291-e55fc0bf49b0-public-tls-certs\") pod \"f36004e0-4967-40f2-b291-e55fc0bf49b0\" (UID: \"f36004e0-4967-40f2-b291-e55fc0bf49b0\") " Dec 11 02:24:42 crc kubenswrapper[4824]: I1211 02:24:42.753237 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f36004e0-4967-40f2-b291-e55fc0bf49b0-config-data\") pod \"f36004e0-4967-40f2-b291-e55fc0bf49b0\" (UID: \"f36004e0-4967-40f2-b291-e55fc0bf49b0\") " Dec 11 02:24:42 crc kubenswrapper[4824]: I1211 02:24:42.754096 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2llf8\" (UniqueName: \"kubernetes.io/projected/f36004e0-4967-40f2-b291-e55fc0bf49b0-kube-api-access-2llf8\") pod \"f36004e0-4967-40f2-b291-e55fc0bf49b0\" (UID: \"f36004e0-4967-40f2-b291-e55fc0bf49b0\") " Dec 11 02:24:42 crc kubenswrapper[4824]: I1211 02:24:42.754863 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f36004e0-4967-40f2-b291-e55fc0bf49b0-logs\") pod \"f36004e0-4967-40f2-b291-e55fc0bf49b0\" (UID: \"f36004e0-4967-40f2-b291-e55fc0bf49b0\") " Dec 11 02:24:42 crc kubenswrapper[4824]: I1211 02:24:42.754909 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f36004e0-4967-40f2-b291-e55fc0bf49b0-combined-ca-bundle\") pod \"f36004e0-4967-40f2-b291-e55fc0bf49b0\" (UID: \"f36004e0-4967-40f2-b291-e55fc0bf49b0\") " Dec 11 02:24:42 crc kubenswrapper[4824]: I1211 02:24:42.755321 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f36004e0-4967-40f2-b291-e55fc0bf49b0-logs" (OuterVolumeSpecName: "logs") pod "f36004e0-4967-40f2-b291-e55fc0bf49b0" (UID: "f36004e0-4967-40f2-b291-e55fc0bf49b0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:24:42 crc kubenswrapper[4824]: I1211 02:24:42.755635 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f36004e0-4967-40f2-b291-e55fc0bf49b0-logs\") on node \"crc\" DevicePath \"\"" Dec 11 02:24:42 crc kubenswrapper[4824]: I1211 02:24:42.760614 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f36004e0-4967-40f2-b291-e55fc0bf49b0-kube-api-access-2llf8" (OuterVolumeSpecName: "kube-api-access-2llf8") pod "f36004e0-4967-40f2-b291-e55fc0bf49b0" (UID: "f36004e0-4967-40f2-b291-e55fc0bf49b0"). InnerVolumeSpecName "kube-api-access-2llf8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:24:42 crc kubenswrapper[4824]: I1211 02:24:42.786144 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f36004e0-4967-40f2-b291-e55fc0bf49b0-config-data" (OuterVolumeSpecName: "config-data") pod "f36004e0-4967-40f2-b291-e55fc0bf49b0" (UID: "f36004e0-4967-40f2-b291-e55fc0bf49b0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:24:42 crc kubenswrapper[4824]: I1211 02:24:42.798522 4824 generic.go:334] "Generic (PLEG): container finished" podID="f36004e0-4967-40f2-b291-e55fc0bf49b0" containerID="e1f5282b07b947a76641d38f716e922be52d7b20881f7735b0a445c8789b887a" exitCode=0 Dec 11 02:24:42 crc kubenswrapper[4824]: I1211 02:24:42.798598 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 11 02:24:42 crc kubenswrapper[4824]: I1211 02:24:42.798621 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f36004e0-4967-40f2-b291-e55fc0bf49b0","Type":"ContainerDied","Data":"e1f5282b07b947a76641d38f716e922be52d7b20881f7735b0a445c8789b887a"} Dec 11 02:24:42 crc kubenswrapper[4824]: I1211 02:24:42.798698 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f36004e0-4967-40f2-b291-e55fc0bf49b0","Type":"ContainerDied","Data":"c3b31a326396e3b4ce025aee6a11f15b7107d549e6d85db9a3debaa6b7bbce3a"} Dec 11 02:24:42 crc kubenswrapper[4824]: I1211 02:24:42.798727 4824 scope.go:117] "RemoveContainer" containerID="e1f5282b07b947a76641d38f716e922be52d7b20881f7735b0a445c8789b887a" Dec 11 02:24:42 crc kubenswrapper[4824]: I1211 02:24:42.827596 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f36004e0-4967-40f2-b291-e55fc0bf49b0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f36004e0-4967-40f2-b291-e55fc0bf49b0" (UID: "f36004e0-4967-40f2-b291-e55fc0bf49b0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:24:42 crc kubenswrapper[4824]: I1211 02:24:42.834738 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f36004e0-4967-40f2-b291-e55fc0bf49b0-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "f36004e0-4967-40f2-b291-e55fc0bf49b0" (UID: "f36004e0-4967-40f2-b291-e55fc0bf49b0"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:24:42 crc kubenswrapper[4824]: I1211 02:24:42.855860 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f36004e0-4967-40f2-b291-e55fc0bf49b0-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "f36004e0-4967-40f2-b291-e55fc0bf49b0" (UID: "f36004e0-4967-40f2-b291-e55fc0bf49b0"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:24:42 crc kubenswrapper[4824]: I1211 02:24:42.856776 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f36004e0-4967-40f2-b291-e55fc0bf49b0-internal-tls-certs\") pod \"f36004e0-4967-40f2-b291-e55fc0bf49b0\" (UID: \"f36004e0-4967-40f2-b291-e55fc0bf49b0\") " Dec 11 02:24:42 crc kubenswrapper[4824]: W1211 02:24:42.856839 4824 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/f36004e0-4967-40f2-b291-e55fc0bf49b0/volumes/kubernetes.io~secret/internal-tls-certs Dec 11 02:24:42 crc kubenswrapper[4824]: I1211 02:24:42.856857 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f36004e0-4967-40f2-b291-e55fc0bf49b0-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "f36004e0-4967-40f2-b291-e55fc0bf49b0" (UID: "f36004e0-4967-40f2-b291-e55fc0bf49b0"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:24:42 crc kubenswrapper[4824]: I1211 02:24:42.858612 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f36004e0-4967-40f2-b291-e55fc0bf49b0-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 02:24:42 crc kubenswrapper[4824]: I1211 02:24:42.858640 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2llf8\" (UniqueName: \"kubernetes.io/projected/f36004e0-4967-40f2-b291-e55fc0bf49b0-kube-api-access-2llf8\") on node \"crc\" DevicePath \"\"" Dec 11 02:24:42 crc kubenswrapper[4824]: I1211 02:24:42.858658 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f36004e0-4967-40f2-b291-e55fc0bf49b0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 02:24:42 crc kubenswrapper[4824]: I1211 02:24:42.858674 4824 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f36004e0-4967-40f2-b291-e55fc0bf49b0-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 11 02:24:42 crc kubenswrapper[4824]: I1211 02:24:42.858690 4824 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f36004e0-4967-40f2-b291-e55fc0bf49b0-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 11 02:24:42 crc kubenswrapper[4824]: I1211 02:24:42.953754 4824 scope.go:117] "RemoveContainer" containerID="eae48b788f653486c7102a8163072c1986cb344f908e39f1e8f20f8afac50a79" Dec 11 02:24:42 crc kubenswrapper[4824]: I1211 02:24:42.973626 4824 scope.go:117] "RemoveContainer" containerID="e1f5282b07b947a76641d38f716e922be52d7b20881f7735b0a445c8789b887a" Dec 11 02:24:42 crc kubenswrapper[4824]: E1211 02:24:42.974182 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e1f5282b07b947a76641d38f716e922be52d7b20881f7735b0a445c8789b887a\": container with ID starting with e1f5282b07b947a76641d38f716e922be52d7b20881f7735b0a445c8789b887a not found: ID does not exist" containerID="e1f5282b07b947a76641d38f716e922be52d7b20881f7735b0a445c8789b887a" Dec 11 02:24:42 crc kubenswrapper[4824]: I1211 02:24:42.974238 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1f5282b07b947a76641d38f716e922be52d7b20881f7735b0a445c8789b887a"} err="failed to get container status \"e1f5282b07b947a76641d38f716e922be52d7b20881f7735b0a445c8789b887a\": rpc error: code = NotFound desc = could not find container \"e1f5282b07b947a76641d38f716e922be52d7b20881f7735b0a445c8789b887a\": container with ID starting with e1f5282b07b947a76641d38f716e922be52d7b20881f7735b0a445c8789b887a not found: ID does not exist" Dec 11 02:24:42 crc kubenswrapper[4824]: I1211 02:24:42.974295 4824 scope.go:117] "RemoveContainer" containerID="eae48b788f653486c7102a8163072c1986cb344f908e39f1e8f20f8afac50a79" Dec 11 02:24:42 crc kubenswrapper[4824]: E1211 02:24:42.974975 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eae48b788f653486c7102a8163072c1986cb344f908e39f1e8f20f8afac50a79\": container with ID starting with eae48b788f653486c7102a8163072c1986cb344f908e39f1e8f20f8afac50a79 not found: ID does not exist" containerID="eae48b788f653486c7102a8163072c1986cb344f908e39f1e8f20f8afac50a79" Dec 11 02:24:42 crc kubenswrapper[4824]: I1211 02:24:42.975006 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eae48b788f653486c7102a8163072c1986cb344f908e39f1e8f20f8afac50a79"} err="failed to get container status \"eae48b788f653486c7102a8163072c1986cb344f908e39f1e8f20f8afac50a79\": rpc error: code = NotFound desc = could not find container \"eae48b788f653486c7102a8163072c1986cb344f908e39f1e8f20f8afac50a79\": container with ID starting with eae48b788f653486c7102a8163072c1986cb344f908e39f1e8f20f8afac50a79 not found: ID does not exist" Dec 11 02:24:43 crc kubenswrapper[4824]: I1211 02:24:43.023969 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 11 02:24:43 crc kubenswrapper[4824]: I1211 02:24:43.149849 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 11 02:24:43 crc kubenswrapper[4824]: I1211 02:24:43.176238 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 11 02:24:43 crc kubenswrapper[4824]: I1211 02:24:43.198768 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 11 02:24:43 crc kubenswrapper[4824]: E1211 02:24:43.199319 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f36004e0-4967-40f2-b291-e55fc0bf49b0" containerName="nova-api-log" Dec 11 02:24:43 crc kubenswrapper[4824]: I1211 02:24:43.199344 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f36004e0-4967-40f2-b291-e55fc0bf49b0" containerName="nova-api-log" Dec 11 02:24:43 crc kubenswrapper[4824]: E1211 02:24:43.199371 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f36004e0-4967-40f2-b291-e55fc0bf49b0" containerName="nova-api-api" Dec 11 02:24:43 crc kubenswrapper[4824]: I1211 02:24:43.199381 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f36004e0-4967-40f2-b291-e55fc0bf49b0" containerName="nova-api-api" Dec 11 02:24:43 crc kubenswrapper[4824]: I1211 02:24:43.199593 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f36004e0-4967-40f2-b291-e55fc0bf49b0" containerName="nova-api-log" Dec 11 02:24:43 crc kubenswrapper[4824]: I1211 02:24:43.199624 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f36004e0-4967-40f2-b291-e55fc0bf49b0" containerName="nova-api-api" Dec 11 02:24:43 crc kubenswrapper[4824]: I1211 02:24:43.200915 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 11 02:24:43 crc kubenswrapper[4824]: I1211 02:24:43.203257 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 11 02:24:43 crc kubenswrapper[4824]: I1211 02:24:43.203302 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 11 02:24:43 crc kubenswrapper[4824]: I1211 02:24:43.203624 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 11 02:24:43 crc kubenswrapper[4824]: I1211 02:24:43.215505 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 11 02:24:43 crc kubenswrapper[4824]: E1211 02:24:43.227818 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf36004e0_4967_40f2_b291_e55fc0bf49b0.slice/crio-c3b31a326396e3b4ce025aee6a11f15b7107d549e6d85db9a3debaa6b7bbce3a\": RecentStats: unable to find data in memory cache]" Dec 11 02:24:43 crc kubenswrapper[4824]: I1211 02:24:43.365078 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a36b8e48-3fc4-4132-a824-a0078424150f-logs\") pod \"nova-api-0\" (UID: \"a36b8e48-3fc4-4132-a824-a0078424150f\") " pod="openstack/nova-api-0" Dec 11 02:24:43 crc kubenswrapper[4824]: I1211 02:24:43.366243 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a36b8e48-3fc4-4132-a824-a0078424150f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a36b8e48-3fc4-4132-a824-a0078424150f\") " pod="openstack/nova-api-0" Dec 11 02:24:43 crc kubenswrapper[4824]: I1211 02:24:43.366365 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a36b8e48-3fc4-4132-a824-a0078424150f-public-tls-certs\") pod \"nova-api-0\" (UID: \"a36b8e48-3fc4-4132-a824-a0078424150f\") " pod="openstack/nova-api-0" Dec 11 02:24:43 crc kubenswrapper[4824]: I1211 02:24:43.366438 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a36b8e48-3fc4-4132-a824-a0078424150f-internal-tls-certs\") pod \"nova-api-0\" (UID: \"a36b8e48-3fc4-4132-a824-a0078424150f\") " pod="openstack/nova-api-0" Dec 11 02:24:43 crc kubenswrapper[4824]: I1211 02:24:43.366547 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6pth7\" (UniqueName: \"kubernetes.io/projected/a36b8e48-3fc4-4132-a824-a0078424150f-kube-api-access-6pth7\") pod \"nova-api-0\" (UID: \"a36b8e48-3fc4-4132-a824-a0078424150f\") " pod="openstack/nova-api-0" Dec 11 02:24:43 crc kubenswrapper[4824]: I1211 02:24:43.366701 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a36b8e48-3fc4-4132-a824-a0078424150f-config-data\") pod \"nova-api-0\" (UID: \"a36b8e48-3fc4-4132-a824-a0078424150f\") " pod="openstack/nova-api-0" Dec 11 02:24:43 crc kubenswrapper[4824]: I1211 02:24:43.469602 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a36b8e48-3fc4-4132-a824-a0078424150f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a36b8e48-3fc4-4132-a824-a0078424150f\") " pod="openstack/nova-api-0" Dec 11 02:24:43 crc kubenswrapper[4824]: I1211 02:24:43.469819 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a36b8e48-3fc4-4132-a824-a0078424150f-public-tls-certs\") pod \"nova-api-0\" (UID: \"a36b8e48-3fc4-4132-a824-a0078424150f\") " pod="openstack/nova-api-0" Dec 11 02:24:43 crc kubenswrapper[4824]: I1211 02:24:43.469991 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a36b8e48-3fc4-4132-a824-a0078424150f-internal-tls-certs\") pod \"nova-api-0\" (UID: \"a36b8e48-3fc4-4132-a824-a0078424150f\") " pod="openstack/nova-api-0" Dec 11 02:24:43 crc kubenswrapper[4824]: I1211 02:24:43.470308 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6pth7\" (UniqueName: \"kubernetes.io/projected/a36b8e48-3fc4-4132-a824-a0078424150f-kube-api-access-6pth7\") pod \"nova-api-0\" (UID: \"a36b8e48-3fc4-4132-a824-a0078424150f\") " pod="openstack/nova-api-0" Dec 11 02:24:43 crc kubenswrapper[4824]: I1211 02:24:43.471268 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a36b8e48-3fc4-4132-a824-a0078424150f-config-data\") pod \"nova-api-0\" (UID: \"a36b8e48-3fc4-4132-a824-a0078424150f\") " pod="openstack/nova-api-0" Dec 11 02:24:43 crc kubenswrapper[4824]: I1211 02:24:43.471547 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a36b8e48-3fc4-4132-a824-a0078424150f-logs\") pod \"nova-api-0\" (UID: \"a36b8e48-3fc4-4132-a824-a0078424150f\") " pod="openstack/nova-api-0" Dec 11 02:24:43 crc kubenswrapper[4824]: I1211 02:24:43.473475 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a36b8e48-3fc4-4132-a824-a0078424150f-public-tls-certs\") pod \"nova-api-0\" (UID: \"a36b8e48-3fc4-4132-a824-a0078424150f\") " pod="openstack/nova-api-0" Dec 11 02:24:43 crc kubenswrapper[4824]: I1211 02:24:43.473536 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a36b8e48-3fc4-4132-a824-a0078424150f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a36b8e48-3fc4-4132-a824-a0078424150f\") " pod="openstack/nova-api-0" Dec 11 02:24:43 crc kubenswrapper[4824]: I1211 02:24:43.473601 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a36b8e48-3fc4-4132-a824-a0078424150f-logs\") pod \"nova-api-0\" (UID: \"a36b8e48-3fc4-4132-a824-a0078424150f\") " pod="openstack/nova-api-0" Dec 11 02:24:43 crc kubenswrapper[4824]: I1211 02:24:43.473679 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a36b8e48-3fc4-4132-a824-a0078424150f-internal-tls-certs\") pod \"nova-api-0\" (UID: \"a36b8e48-3fc4-4132-a824-a0078424150f\") " pod="openstack/nova-api-0" Dec 11 02:24:43 crc kubenswrapper[4824]: I1211 02:24:43.477591 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a36b8e48-3fc4-4132-a824-a0078424150f-config-data\") pod \"nova-api-0\" (UID: \"a36b8e48-3fc4-4132-a824-a0078424150f\") " pod="openstack/nova-api-0" Dec 11 02:24:43 crc kubenswrapper[4824]: I1211 02:24:43.488007 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6pth7\" (UniqueName: \"kubernetes.io/projected/a36b8e48-3fc4-4132-a824-a0078424150f-kube-api-access-6pth7\") pod \"nova-api-0\" (UID: \"a36b8e48-3fc4-4132-a824-a0078424150f\") " pod="openstack/nova-api-0" Dec 11 02:24:43 crc kubenswrapper[4824]: I1211 02:24:43.528548 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 11 02:24:44 crc kubenswrapper[4824]: I1211 02:24:44.001399 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 11 02:24:44 crc kubenswrapper[4824]: I1211 02:24:44.676553 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f36004e0-4967-40f2-b291-e55fc0bf49b0" path="/var/lib/kubelet/pods/f36004e0-4967-40f2-b291-e55fc0bf49b0/volumes" Dec 11 02:24:44 crc kubenswrapper[4824]: I1211 02:24:44.822608 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a36b8e48-3fc4-4132-a824-a0078424150f","Type":"ContainerStarted","Data":"3b5eef2a75c41fdf66bafefa0f64e133235ecd031212698a32712559d8a5c939"} Dec 11 02:24:44 crc kubenswrapper[4824]: I1211 02:24:44.822683 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a36b8e48-3fc4-4132-a824-a0078424150f","Type":"ContainerStarted","Data":"1d8c104edb306e698b86ca3c07073023a76f00d09cd0c73d7520145038756094"} Dec 11 02:24:44 crc kubenswrapper[4824]: I1211 02:24:44.822705 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a36b8e48-3fc4-4132-a824-a0078424150f","Type":"ContainerStarted","Data":"8e867d80c8558cc70307e10df70f59c93da5294e2827a27b940edc01619fbdd5"} Dec 11 02:24:44 crc kubenswrapper[4824]: I1211 02:24:44.862448 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=1.862424211 podStartE2EDuration="1.862424211s" podCreationTimestamp="2025-12-11 02:24:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:24:44.848669405 +0000 UTC m=+1426.537706824" watchObservedRunningTime="2025-12-11 02:24:44.862424211 +0000 UTC m=+1426.551461600" Dec 11 02:24:45 crc kubenswrapper[4824]: I1211 02:24:45.422315 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 11 02:24:45 crc kubenswrapper[4824]: I1211 02:24:45.422575 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 11 02:24:48 crc kubenswrapper[4824]: I1211 02:24:48.024886 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 11 02:24:48 crc kubenswrapper[4824]: I1211 02:24:48.081664 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 11 02:24:48 crc kubenswrapper[4824]: I1211 02:24:48.923243 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 11 02:24:50 crc kubenswrapper[4824]: I1211 02:24:50.421645 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 11 02:24:50 crc kubenswrapper[4824]: I1211 02:24:50.421707 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 11 02:24:51 crc kubenswrapper[4824]: I1211 02:24:51.437435 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="e0e1ab61-fe6b-4b2e-aa01-2bf77e119ce4" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.206:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 11 02:24:51 crc kubenswrapper[4824]: I1211 02:24:51.437470 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="e0e1ab61-fe6b-4b2e-aa01-2bf77e119ce4" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.206:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 11 02:24:53 crc kubenswrapper[4824]: I1211 02:24:53.529678 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 11 02:24:53 crc kubenswrapper[4824]: I1211 02:24:53.529998 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 11 02:24:53 crc kubenswrapper[4824]: I1211 02:24:53.787649 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 11 02:24:54 crc kubenswrapper[4824]: I1211 02:24:54.541435 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="a36b8e48-3fc4-4132-a824-a0078424150f" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.207:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 11 02:24:54 crc kubenswrapper[4824]: I1211 02:24:54.541447 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="a36b8e48-3fc4-4132-a824-a0078424150f" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.207:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 11 02:24:56 crc kubenswrapper[4824]: I1211 02:24:56.238163 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-lpxp6"] Dec 11 02:24:56 crc kubenswrapper[4824]: I1211 02:24:56.241180 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lpxp6" Dec 11 02:24:56 crc kubenswrapper[4824]: I1211 02:24:56.251409 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lpxp6"] Dec 11 02:24:56 crc kubenswrapper[4824]: I1211 02:24:56.259908 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f02884bb-bee5-4388-a306-f1280d81dff5-utilities\") pod \"redhat-operators-lpxp6\" (UID: \"f02884bb-bee5-4388-a306-f1280d81dff5\") " pod="openshift-marketplace/redhat-operators-lpxp6" Dec 11 02:24:56 crc kubenswrapper[4824]: I1211 02:24:56.260261 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9d9hv\" (UniqueName: \"kubernetes.io/projected/f02884bb-bee5-4388-a306-f1280d81dff5-kube-api-access-9d9hv\") pod \"redhat-operators-lpxp6\" (UID: \"f02884bb-bee5-4388-a306-f1280d81dff5\") " pod="openshift-marketplace/redhat-operators-lpxp6" Dec 11 02:24:56 crc kubenswrapper[4824]: I1211 02:24:56.260464 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f02884bb-bee5-4388-a306-f1280d81dff5-catalog-content\") pod \"redhat-operators-lpxp6\" (UID: \"f02884bb-bee5-4388-a306-f1280d81dff5\") " pod="openshift-marketplace/redhat-operators-lpxp6" Dec 11 02:24:56 crc kubenswrapper[4824]: I1211 02:24:56.362567 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9d9hv\" (UniqueName: \"kubernetes.io/projected/f02884bb-bee5-4388-a306-f1280d81dff5-kube-api-access-9d9hv\") pod \"redhat-operators-lpxp6\" (UID: \"f02884bb-bee5-4388-a306-f1280d81dff5\") " pod="openshift-marketplace/redhat-operators-lpxp6" Dec 11 02:24:56 crc kubenswrapper[4824]: I1211 02:24:56.362651 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f02884bb-bee5-4388-a306-f1280d81dff5-catalog-content\") pod \"redhat-operators-lpxp6\" (UID: \"f02884bb-bee5-4388-a306-f1280d81dff5\") " pod="openshift-marketplace/redhat-operators-lpxp6" Dec 11 02:24:56 crc kubenswrapper[4824]: I1211 02:24:56.362699 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f02884bb-bee5-4388-a306-f1280d81dff5-utilities\") pod \"redhat-operators-lpxp6\" (UID: \"f02884bb-bee5-4388-a306-f1280d81dff5\") " pod="openshift-marketplace/redhat-operators-lpxp6" Dec 11 02:24:56 crc kubenswrapper[4824]: I1211 02:24:56.363219 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f02884bb-bee5-4388-a306-f1280d81dff5-utilities\") pod \"redhat-operators-lpxp6\" (UID: \"f02884bb-bee5-4388-a306-f1280d81dff5\") " pod="openshift-marketplace/redhat-operators-lpxp6" Dec 11 02:24:56 crc kubenswrapper[4824]: I1211 02:24:56.363445 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f02884bb-bee5-4388-a306-f1280d81dff5-catalog-content\") pod \"redhat-operators-lpxp6\" (UID: \"f02884bb-bee5-4388-a306-f1280d81dff5\") " pod="openshift-marketplace/redhat-operators-lpxp6" Dec 11 02:24:56 crc kubenswrapper[4824]: I1211 02:24:56.385625 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9d9hv\" (UniqueName: \"kubernetes.io/projected/f02884bb-bee5-4388-a306-f1280d81dff5-kube-api-access-9d9hv\") pod \"redhat-operators-lpxp6\" (UID: \"f02884bb-bee5-4388-a306-f1280d81dff5\") " pod="openshift-marketplace/redhat-operators-lpxp6" Dec 11 02:24:56 crc kubenswrapper[4824]: I1211 02:24:56.620938 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lpxp6" Dec 11 02:24:57 crc kubenswrapper[4824]: I1211 02:24:57.280529 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lpxp6"] Dec 11 02:24:57 crc kubenswrapper[4824]: W1211 02:24:57.281554 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf02884bb_bee5_4388_a306_f1280d81dff5.slice/crio-197bdec5b7792f4cebcde5abba592453cc3d9252c0a37d11b56b4ca99e8387f3 WatchSource:0}: Error finding container 197bdec5b7792f4cebcde5abba592453cc3d9252c0a37d11b56b4ca99e8387f3: Status 404 returned error can't find the container with id 197bdec5b7792f4cebcde5abba592453cc3d9252c0a37d11b56b4ca99e8387f3 Dec 11 02:24:57 crc kubenswrapper[4824]: I1211 02:24:57.959064 4824 generic.go:334] "Generic (PLEG): container finished" podID="f02884bb-bee5-4388-a306-f1280d81dff5" containerID="3626f8301752a169225437e9226e614c221966c6f8d00522f45ae09b495f005c" exitCode=0 Dec 11 02:24:57 crc kubenswrapper[4824]: I1211 02:24:57.959191 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lpxp6" event={"ID":"f02884bb-bee5-4388-a306-f1280d81dff5","Type":"ContainerDied","Data":"3626f8301752a169225437e9226e614c221966c6f8d00522f45ae09b495f005c"} Dec 11 02:24:57 crc kubenswrapper[4824]: I1211 02:24:57.959344 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lpxp6" event={"ID":"f02884bb-bee5-4388-a306-f1280d81dff5","Type":"ContainerStarted","Data":"197bdec5b7792f4cebcde5abba592453cc3d9252c0a37d11b56b4ca99e8387f3"} Dec 11 02:24:59 crc kubenswrapper[4824]: I1211 02:24:59.986530 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lpxp6" event={"ID":"f02884bb-bee5-4388-a306-f1280d81dff5","Type":"ContainerStarted","Data":"569cc2dcc127e4ae7953961cb97beacd04a583ddf6882dcc40535cd2902cc19c"} Dec 11 02:25:00 crc kubenswrapper[4824]: I1211 02:25:00.427871 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 11 02:25:00 crc kubenswrapper[4824]: I1211 02:25:00.430450 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 11 02:25:00 crc kubenswrapper[4824]: I1211 02:25:00.437510 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 11 02:25:01 crc kubenswrapper[4824]: I1211 02:25:01.012629 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 11 02:25:02 crc kubenswrapper[4824]: I1211 02:25:02.012065 4824 generic.go:334] "Generic (PLEG): container finished" podID="f02884bb-bee5-4388-a306-f1280d81dff5" containerID="569cc2dcc127e4ae7953961cb97beacd04a583ddf6882dcc40535cd2902cc19c" exitCode=0 Dec 11 02:25:02 crc kubenswrapper[4824]: I1211 02:25:02.012138 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lpxp6" event={"ID":"f02884bb-bee5-4388-a306-f1280d81dff5","Type":"ContainerDied","Data":"569cc2dcc127e4ae7953961cb97beacd04a583ddf6882dcc40535cd2902cc19c"} Dec 11 02:25:03 crc kubenswrapper[4824]: I1211 02:25:03.541584 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 11 02:25:03 crc kubenswrapper[4824]: I1211 02:25:03.542412 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 11 02:25:03 crc kubenswrapper[4824]: I1211 02:25:03.545975 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 11 02:25:03 crc kubenswrapper[4824]: I1211 02:25:03.555663 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 11 02:25:04 crc kubenswrapper[4824]: I1211 02:25:04.046299 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lpxp6" event={"ID":"f02884bb-bee5-4388-a306-f1280d81dff5","Type":"ContainerStarted","Data":"56430a33a172baf82856f335906df090dc7efe4c9ce26ad8e5d83f62bffac02e"} Dec 11 02:25:04 crc kubenswrapper[4824]: I1211 02:25:04.046904 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 11 02:25:04 crc kubenswrapper[4824]: I1211 02:25:04.060530 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 11 02:25:04 crc kubenswrapper[4824]: I1211 02:25:04.075035 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-lpxp6" podStartSLOduration=2.947486741 podStartE2EDuration="8.075020181s" podCreationTimestamp="2025-12-11 02:24:56 +0000 UTC" firstStartedPulling="2025-12-11 02:24:57.962051159 +0000 UTC m=+1439.651088538" lastFinishedPulling="2025-12-11 02:25:03.089584599 +0000 UTC m=+1444.778621978" observedRunningTime="2025-12-11 02:25:04.068470252 +0000 UTC m=+1445.757507681" watchObservedRunningTime="2025-12-11 02:25:04.075020181 +0000 UTC m=+1445.764057560" Dec 11 02:25:06 crc kubenswrapper[4824]: I1211 02:25:06.621721 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-lpxp6" Dec 11 02:25:06 crc kubenswrapper[4824]: I1211 02:25:06.622088 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-lpxp6" Dec 11 02:25:07 crc kubenswrapper[4824]: I1211 02:25:07.712528 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-lpxp6" podUID="f02884bb-bee5-4388-a306-f1280d81dff5" containerName="registry-server" probeResult="failure" output=< Dec 11 02:25:07 crc kubenswrapper[4824]: timeout: failed to connect service ":50051" within 1s Dec 11 02:25:07 crc kubenswrapper[4824]: > Dec 11 02:25:12 crc kubenswrapper[4824]: I1211 02:25:12.387590 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 11 02:25:13 crc kubenswrapper[4824]: I1211 02:25:13.364930 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 11 02:25:16 crc kubenswrapper[4824]: I1211 02:25:16.460136 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="1fd840b6-4ace-45b8-a8d2-38a041110095" containerName="rabbitmq" containerID="cri-o://8dee72102c0022a859a56e54e0f34c0048fded2348fc7ecb0d3de490ad11c4cd" gracePeriod=604796 Dec 11 02:25:16 crc kubenswrapper[4824]: I1211 02:25:16.694413 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-lpxp6" Dec 11 02:25:16 crc kubenswrapper[4824]: I1211 02:25:16.763753 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-lpxp6" Dec 11 02:25:16 crc kubenswrapper[4824]: I1211 02:25:16.929513 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lpxp6"] Dec 11 02:25:17 crc kubenswrapper[4824]: I1211 02:25:17.568620 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="ff37fa2f-dc06-430a-a4fe-af2b379fe563" containerName="rabbitmq" containerID="cri-o://69c1077bbc1e742ec896e90a037d4489755f60af0094fc7227c893dc106fa13b" gracePeriod=604796 Dec 11 02:25:18 crc kubenswrapper[4824]: I1211 02:25:18.192198 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-lpxp6" podUID="f02884bb-bee5-4388-a306-f1280d81dff5" containerName="registry-server" containerID="cri-o://56430a33a172baf82856f335906df090dc7efe4c9ce26ad8e5d83f62bffac02e" gracePeriod=2 Dec 11 02:25:18 crc kubenswrapper[4824]: I1211 02:25:18.702182 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lpxp6" Dec 11 02:25:18 crc kubenswrapper[4824]: I1211 02:25:18.748630 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f02884bb-bee5-4388-a306-f1280d81dff5-catalog-content\") pod \"f02884bb-bee5-4388-a306-f1280d81dff5\" (UID: \"f02884bb-bee5-4388-a306-f1280d81dff5\") " Dec 11 02:25:18 crc kubenswrapper[4824]: I1211 02:25:18.748837 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f02884bb-bee5-4388-a306-f1280d81dff5-utilities\") pod \"f02884bb-bee5-4388-a306-f1280d81dff5\" (UID: \"f02884bb-bee5-4388-a306-f1280d81dff5\") " Dec 11 02:25:18 crc kubenswrapper[4824]: I1211 02:25:18.748893 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9d9hv\" (UniqueName: \"kubernetes.io/projected/f02884bb-bee5-4388-a306-f1280d81dff5-kube-api-access-9d9hv\") pod \"f02884bb-bee5-4388-a306-f1280d81dff5\" (UID: \"f02884bb-bee5-4388-a306-f1280d81dff5\") " Dec 11 02:25:18 crc kubenswrapper[4824]: I1211 02:25:18.753055 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f02884bb-bee5-4388-a306-f1280d81dff5-utilities" (OuterVolumeSpecName: "utilities") pod "f02884bb-bee5-4388-a306-f1280d81dff5" (UID: "f02884bb-bee5-4388-a306-f1280d81dff5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:25:18 crc kubenswrapper[4824]: I1211 02:25:18.761183 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f02884bb-bee5-4388-a306-f1280d81dff5-kube-api-access-9d9hv" (OuterVolumeSpecName: "kube-api-access-9d9hv") pod "f02884bb-bee5-4388-a306-f1280d81dff5" (UID: "f02884bb-bee5-4388-a306-f1280d81dff5"). InnerVolumeSpecName "kube-api-access-9d9hv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:25:18 crc kubenswrapper[4824]: I1211 02:25:18.852327 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f02884bb-bee5-4388-a306-f1280d81dff5-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 02:25:18 crc kubenswrapper[4824]: I1211 02:25:18.852669 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9d9hv\" (UniqueName: \"kubernetes.io/projected/f02884bb-bee5-4388-a306-f1280d81dff5-kube-api-access-9d9hv\") on node \"crc\" DevicePath \"\"" Dec 11 02:25:18 crc kubenswrapper[4824]: I1211 02:25:18.879842 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f02884bb-bee5-4388-a306-f1280d81dff5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f02884bb-bee5-4388-a306-f1280d81dff5" (UID: "f02884bb-bee5-4388-a306-f1280d81dff5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:25:18 crc kubenswrapper[4824]: I1211 02:25:18.954134 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f02884bb-bee5-4388-a306-f1280d81dff5-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 02:25:19 crc kubenswrapper[4824]: I1211 02:25:19.202736 4824 generic.go:334] "Generic (PLEG): container finished" podID="f02884bb-bee5-4388-a306-f1280d81dff5" containerID="56430a33a172baf82856f335906df090dc7efe4c9ce26ad8e5d83f62bffac02e" exitCode=0 Dec 11 02:25:19 crc kubenswrapper[4824]: I1211 02:25:19.202797 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lpxp6" event={"ID":"f02884bb-bee5-4388-a306-f1280d81dff5","Type":"ContainerDied","Data":"56430a33a172baf82856f335906df090dc7efe4c9ce26ad8e5d83f62bffac02e"} Dec 11 02:25:19 crc kubenswrapper[4824]: I1211 02:25:19.202882 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lpxp6" event={"ID":"f02884bb-bee5-4388-a306-f1280d81dff5","Type":"ContainerDied","Data":"197bdec5b7792f4cebcde5abba592453cc3d9252c0a37d11b56b4ca99e8387f3"} Dec 11 02:25:19 crc kubenswrapper[4824]: I1211 02:25:19.202908 4824 scope.go:117] "RemoveContainer" containerID="56430a33a172baf82856f335906df090dc7efe4c9ce26ad8e5d83f62bffac02e" Dec 11 02:25:19 crc kubenswrapper[4824]: I1211 02:25:19.203159 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lpxp6" Dec 11 02:25:19 crc kubenswrapper[4824]: I1211 02:25:19.232963 4824 scope.go:117] "RemoveContainer" containerID="569cc2dcc127e4ae7953961cb97beacd04a583ddf6882dcc40535cd2902cc19c" Dec 11 02:25:19 crc kubenswrapper[4824]: I1211 02:25:19.251167 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lpxp6"] Dec 11 02:25:19 crc kubenswrapper[4824]: I1211 02:25:19.258361 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-lpxp6"] Dec 11 02:25:19 crc kubenswrapper[4824]: I1211 02:25:19.265296 4824 scope.go:117] "RemoveContainer" containerID="3626f8301752a169225437e9226e614c221966c6f8d00522f45ae09b495f005c" Dec 11 02:25:19 crc kubenswrapper[4824]: I1211 02:25:19.357442 4824 scope.go:117] "RemoveContainer" containerID="56430a33a172baf82856f335906df090dc7efe4c9ce26ad8e5d83f62bffac02e" Dec 11 02:25:19 crc kubenswrapper[4824]: E1211 02:25:19.357959 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"56430a33a172baf82856f335906df090dc7efe4c9ce26ad8e5d83f62bffac02e\": container with ID starting with 56430a33a172baf82856f335906df090dc7efe4c9ce26ad8e5d83f62bffac02e not found: ID does not exist" containerID="56430a33a172baf82856f335906df090dc7efe4c9ce26ad8e5d83f62bffac02e" Dec 11 02:25:19 crc kubenswrapper[4824]: I1211 02:25:19.358020 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"56430a33a172baf82856f335906df090dc7efe4c9ce26ad8e5d83f62bffac02e"} err="failed to get container status \"56430a33a172baf82856f335906df090dc7efe4c9ce26ad8e5d83f62bffac02e\": rpc error: code = NotFound desc = could not find container \"56430a33a172baf82856f335906df090dc7efe4c9ce26ad8e5d83f62bffac02e\": container with ID starting with 56430a33a172baf82856f335906df090dc7efe4c9ce26ad8e5d83f62bffac02e not found: ID does not exist" Dec 11 02:25:19 crc kubenswrapper[4824]: I1211 02:25:19.358061 4824 scope.go:117] "RemoveContainer" containerID="569cc2dcc127e4ae7953961cb97beacd04a583ddf6882dcc40535cd2902cc19c" Dec 11 02:25:19 crc kubenswrapper[4824]: E1211 02:25:19.358534 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"569cc2dcc127e4ae7953961cb97beacd04a583ddf6882dcc40535cd2902cc19c\": container with ID starting with 569cc2dcc127e4ae7953961cb97beacd04a583ddf6882dcc40535cd2902cc19c not found: ID does not exist" containerID="569cc2dcc127e4ae7953961cb97beacd04a583ddf6882dcc40535cd2902cc19c" Dec 11 02:25:19 crc kubenswrapper[4824]: I1211 02:25:19.358582 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"569cc2dcc127e4ae7953961cb97beacd04a583ddf6882dcc40535cd2902cc19c"} err="failed to get container status \"569cc2dcc127e4ae7953961cb97beacd04a583ddf6882dcc40535cd2902cc19c\": rpc error: code = NotFound desc = could not find container \"569cc2dcc127e4ae7953961cb97beacd04a583ddf6882dcc40535cd2902cc19c\": container with ID starting with 569cc2dcc127e4ae7953961cb97beacd04a583ddf6882dcc40535cd2902cc19c not found: ID does not exist" Dec 11 02:25:19 crc kubenswrapper[4824]: I1211 02:25:19.358617 4824 scope.go:117] "RemoveContainer" containerID="3626f8301752a169225437e9226e614c221966c6f8d00522f45ae09b495f005c" Dec 11 02:25:19 crc kubenswrapper[4824]: E1211 02:25:19.359215 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3626f8301752a169225437e9226e614c221966c6f8d00522f45ae09b495f005c\": container with ID starting with 3626f8301752a169225437e9226e614c221966c6f8d00522f45ae09b495f005c not found: ID does not exist" containerID="3626f8301752a169225437e9226e614c221966c6f8d00522f45ae09b495f005c" Dec 11 02:25:19 crc kubenswrapper[4824]: I1211 02:25:19.359255 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3626f8301752a169225437e9226e614c221966c6f8d00522f45ae09b495f005c"} err="failed to get container status \"3626f8301752a169225437e9226e614c221966c6f8d00522f45ae09b495f005c\": rpc error: code = NotFound desc = could not find container \"3626f8301752a169225437e9226e614c221966c6f8d00522f45ae09b495f005c\": container with ID starting with 3626f8301752a169225437e9226e614c221966c6f8d00522f45ae09b495f005c not found: ID does not exist" Dec 11 02:25:20 crc kubenswrapper[4824]: I1211 02:25:20.649415 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f02884bb-bee5-4388-a306-f1280d81dff5" path="/var/lib/kubelet/pods/f02884bb-bee5-4388-a306-f1280d81dff5/volumes" Dec 11 02:25:20 crc kubenswrapper[4824]: I1211 02:25:20.746504 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="1fd840b6-4ace-45b8-a8d2-38a041110095" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.97:5671: connect: connection refused" Dec 11 02:25:20 crc kubenswrapper[4824]: I1211 02:25:20.821080 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="ff37fa2f-dc06-430a-a4fe-af2b379fe563" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.98:5671: connect: connection refused" Dec 11 02:25:23 crc kubenswrapper[4824]: I1211 02:25:23.886839 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 11 02:25:23 crc kubenswrapper[4824]: I1211 02:25:23.961916 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dnhpc\" (UniqueName: \"kubernetes.io/projected/1fd840b6-4ace-45b8-a8d2-38a041110095-kube-api-access-dnhpc\") pod \"1fd840b6-4ace-45b8-a8d2-38a041110095\" (UID: \"1fd840b6-4ace-45b8-a8d2-38a041110095\") " Dec 11 02:25:23 crc kubenswrapper[4824]: I1211 02:25:23.961998 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1fd840b6-4ace-45b8-a8d2-38a041110095-config-data\") pod \"1fd840b6-4ace-45b8-a8d2-38a041110095\" (UID: \"1fd840b6-4ace-45b8-a8d2-38a041110095\") " Dec 11 02:25:23 crc kubenswrapper[4824]: I1211 02:25:23.962027 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1fd840b6-4ace-45b8-a8d2-38a041110095-rabbitmq-plugins\") pod \"1fd840b6-4ace-45b8-a8d2-38a041110095\" (UID: \"1fd840b6-4ace-45b8-a8d2-38a041110095\") " Dec 11 02:25:23 crc kubenswrapper[4824]: I1211 02:25:23.962053 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"1fd840b6-4ace-45b8-a8d2-38a041110095\" (UID: \"1fd840b6-4ace-45b8-a8d2-38a041110095\") " Dec 11 02:25:23 crc kubenswrapper[4824]: I1211 02:25:23.962085 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1fd840b6-4ace-45b8-a8d2-38a041110095-erlang-cookie-secret\") pod \"1fd840b6-4ace-45b8-a8d2-38a041110095\" (UID: \"1fd840b6-4ace-45b8-a8d2-38a041110095\") " Dec 11 02:25:23 crc kubenswrapper[4824]: I1211 02:25:23.962129 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1fd840b6-4ace-45b8-a8d2-38a041110095-pod-info\") pod \"1fd840b6-4ace-45b8-a8d2-38a041110095\" (UID: \"1fd840b6-4ace-45b8-a8d2-38a041110095\") " Dec 11 02:25:23 crc kubenswrapper[4824]: I1211 02:25:23.962191 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1fd840b6-4ace-45b8-a8d2-38a041110095-rabbitmq-confd\") pod \"1fd840b6-4ace-45b8-a8d2-38a041110095\" (UID: \"1fd840b6-4ace-45b8-a8d2-38a041110095\") " Dec 11 02:25:23 crc kubenswrapper[4824]: I1211 02:25:23.962239 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1fd840b6-4ace-45b8-a8d2-38a041110095-plugins-conf\") pod \"1fd840b6-4ace-45b8-a8d2-38a041110095\" (UID: \"1fd840b6-4ace-45b8-a8d2-38a041110095\") " Dec 11 02:25:23 crc kubenswrapper[4824]: I1211 02:25:23.962274 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1fd840b6-4ace-45b8-a8d2-38a041110095-server-conf\") pod \"1fd840b6-4ace-45b8-a8d2-38a041110095\" (UID: \"1fd840b6-4ace-45b8-a8d2-38a041110095\") " Dec 11 02:25:23 crc kubenswrapper[4824]: I1211 02:25:23.962316 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1fd840b6-4ace-45b8-a8d2-38a041110095-rabbitmq-erlang-cookie\") pod \"1fd840b6-4ace-45b8-a8d2-38a041110095\" (UID: \"1fd840b6-4ace-45b8-a8d2-38a041110095\") " Dec 11 02:25:23 crc kubenswrapper[4824]: I1211 02:25:23.962393 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1fd840b6-4ace-45b8-a8d2-38a041110095-rabbitmq-tls\") pod \"1fd840b6-4ace-45b8-a8d2-38a041110095\" (UID: \"1fd840b6-4ace-45b8-a8d2-38a041110095\") " Dec 11 02:25:23 crc kubenswrapper[4824]: I1211 02:25:23.962965 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1fd840b6-4ace-45b8-a8d2-38a041110095-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "1fd840b6-4ace-45b8-a8d2-38a041110095" (UID: "1fd840b6-4ace-45b8-a8d2-38a041110095"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:25:23 crc kubenswrapper[4824]: I1211 02:25:23.963257 4824 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1fd840b6-4ace-45b8-a8d2-38a041110095-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 11 02:25:23 crc kubenswrapper[4824]: I1211 02:25:23.963306 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1fd840b6-4ace-45b8-a8d2-38a041110095-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "1fd840b6-4ace-45b8-a8d2-38a041110095" (UID: "1fd840b6-4ace-45b8-a8d2-38a041110095"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:25:23 crc kubenswrapper[4824]: I1211 02:25:23.963604 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1fd840b6-4ace-45b8-a8d2-38a041110095-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "1fd840b6-4ace-45b8-a8d2-38a041110095" (UID: "1fd840b6-4ace-45b8-a8d2-38a041110095"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:25:23 crc kubenswrapper[4824]: I1211 02:25:23.968995 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "persistence") pod "1fd840b6-4ace-45b8-a8d2-38a041110095" (UID: "1fd840b6-4ace-45b8-a8d2-38a041110095"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 11 02:25:23 crc kubenswrapper[4824]: I1211 02:25:23.969680 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/1fd840b6-4ace-45b8-a8d2-38a041110095-pod-info" (OuterVolumeSpecName: "pod-info") pod "1fd840b6-4ace-45b8-a8d2-38a041110095" (UID: "1fd840b6-4ace-45b8-a8d2-38a041110095"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 11 02:25:23 crc kubenswrapper[4824]: I1211 02:25:23.970399 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1fd840b6-4ace-45b8-a8d2-38a041110095-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "1fd840b6-4ace-45b8-a8d2-38a041110095" (UID: "1fd840b6-4ace-45b8-a8d2-38a041110095"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:25:23 crc kubenswrapper[4824]: I1211 02:25:23.975634 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1fd840b6-4ace-45b8-a8d2-38a041110095-kube-api-access-dnhpc" (OuterVolumeSpecName: "kube-api-access-dnhpc") pod "1fd840b6-4ace-45b8-a8d2-38a041110095" (UID: "1fd840b6-4ace-45b8-a8d2-38a041110095"). InnerVolumeSpecName "kube-api-access-dnhpc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:25:23 crc kubenswrapper[4824]: I1211 02:25:23.996465 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1fd840b6-4ace-45b8-a8d2-38a041110095-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "1fd840b6-4ace-45b8-a8d2-38a041110095" (UID: "1fd840b6-4ace-45b8-a8d2-38a041110095"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.020243 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1fd840b6-4ace-45b8-a8d2-38a041110095-config-data" (OuterVolumeSpecName: "config-data") pod "1fd840b6-4ace-45b8-a8d2-38a041110095" (UID: "1fd840b6-4ace-45b8-a8d2-38a041110095"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.064407 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dnhpc\" (UniqueName: \"kubernetes.io/projected/1fd840b6-4ace-45b8-a8d2-38a041110095-kube-api-access-dnhpc\") on node \"crc\" DevicePath \"\"" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.064441 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1fd840b6-4ace-45b8-a8d2-38a041110095-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.064453 4824 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1fd840b6-4ace-45b8-a8d2-38a041110095-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.064473 4824 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.064483 4824 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1fd840b6-4ace-45b8-a8d2-38a041110095-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.064491 4824 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1fd840b6-4ace-45b8-a8d2-38a041110095-pod-info\") on node \"crc\" DevicePath \"\"" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.064500 4824 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1fd840b6-4ace-45b8-a8d2-38a041110095-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.064508 4824 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1fd840b6-4ace-45b8-a8d2-38a041110095-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.070360 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1fd840b6-4ace-45b8-a8d2-38a041110095-server-conf" (OuterVolumeSpecName: "server-conf") pod "1fd840b6-4ace-45b8-a8d2-38a041110095" (UID: "1fd840b6-4ace-45b8-a8d2-38a041110095"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.086531 4824 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.144843 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1fd840b6-4ace-45b8-a8d2-38a041110095-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "1fd840b6-4ace-45b8-a8d2-38a041110095" (UID: "1fd840b6-4ace-45b8-a8d2-38a041110095"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.147661 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.171984 4824 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1fd840b6-4ace-45b8-a8d2-38a041110095-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.172015 4824 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1fd840b6-4ace-45b8-a8d2-38a041110095-server-conf\") on node \"crc\" DevicePath \"\"" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.172024 4824 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.263787 4824 generic.go:334] "Generic (PLEG): container finished" podID="ff37fa2f-dc06-430a-a4fe-af2b379fe563" containerID="69c1077bbc1e742ec896e90a037d4489755f60af0094fc7227c893dc106fa13b" exitCode=0 Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.263849 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.263836 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ff37fa2f-dc06-430a-a4fe-af2b379fe563","Type":"ContainerDied","Data":"69c1077bbc1e742ec896e90a037d4489755f60af0094fc7227c893dc106fa13b"} Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.264179 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ff37fa2f-dc06-430a-a4fe-af2b379fe563","Type":"ContainerDied","Data":"80a3ece43e3b693e934325a4e77de52fbe6239b597684802908dd120ff83d31c"} Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.264216 4824 scope.go:117] "RemoveContainer" containerID="69c1077bbc1e742ec896e90a037d4489755f60af0094fc7227c893dc106fa13b" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.267019 4824 generic.go:334] "Generic (PLEG): container finished" podID="1fd840b6-4ace-45b8-a8d2-38a041110095" containerID="8dee72102c0022a859a56e54e0f34c0048fded2348fc7ecb0d3de490ad11c4cd" exitCode=0 Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.267068 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1fd840b6-4ace-45b8-a8d2-38a041110095","Type":"ContainerDied","Data":"8dee72102c0022a859a56e54e0f34c0048fded2348fc7ecb0d3de490ad11c4cd"} Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.267093 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1fd840b6-4ace-45b8-a8d2-38a041110095","Type":"ContainerDied","Data":"7e13edf4baa20214b9db9d0a84b13be73c96e113fe3b4a0da3a5bcd9654b8599"} Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.267172 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.273474 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7gjz6\" (UniqueName: \"kubernetes.io/projected/ff37fa2f-dc06-430a-a4fe-af2b379fe563-kube-api-access-7gjz6\") pod \"ff37fa2f-dc06-430a-a4fe-af2b379fe563\" (UID: \"ff37fa2f-dc06-430a-a4fe-af2b379fe563\") " Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.273528 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ff37fa2f-dc06-430a-a4fe-af2b379fe563-server-conf\") pod \"ff37fa2f-dc06-430a-a4fe-af2b379fe563\" (UID: \"ff37fa2f-dc06-430a-a4fe-af2b379fe563\") " Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.273563 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ff37fa2f-dc06-430a-a4fe-af2b379fe563-rabbitmq-plugins\") pod \"ff37fa2f-dc06-430a-a4fe-af2b379fe563\" (UID: \"ff37fa2f-dc06-430a-a4fe-af2b379fe563\") " Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.273616 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ff37fa2f-dc06-430a-a4fe-af2b379fe563\" (UID: \"ff37fa2f-dc06-430a-a4fe-af2b379fe563\") " Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.273641 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ff37fa2f-dc06-430a-a4fe-af2b379fe563-pod-info\") pod \"ff37fa2f-dc06-430a-a4fe-af2b379fe563\" (UID: \"ff37fa2f-dc06-430a-a4fe-af2b379fe563\") " Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.273672 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ff37fa2f-dc06-430a-a4fe-af2b379fe563-rabbitmq-confd\") pod \"ff37fa2f-dc06-430a-a4fe-af2b379fe563\" (UID: \"ff37fa2f-dc06-430a-a4fe-af2b379fe563\") " Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.273703 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ff37fa2f-dc06-430a-a4fe-af2b379fe563-erlang-cookie-secret\") pod \"ff37fa2f-dc06-430a-a4fe-af2b379fe563\" (UID: \"ff37fa2f-dc06-430a-a4fe-af2b379fe563\") " Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.273735 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ff37fa2f-dc06-430a-a4fe-af2b379fe563-rabbitmq-erlang-cookie\") pod \"ff37fa2f-dc06-430a-a4fe-af2b379fe563\" (UID: \"ff37fa2f-dc06-430a-a4fe-af2b379fe563\") " Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.273796 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ff37fa2f-dc06-430a-a4fe-af2b379fe563-plugins-conf\") pod \"ff37fa2f-dc06-430a-a4fe-af2b379fe563\" (UID: \"ff37fa2f-dc06-430a-a4fe-af2b379fe563\") " Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.273872 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ff37fa2f-dc06-430a-a4fe-af2b379fe563-config-data\") pod \"ff37fa2f-dc06-430a-a4fe-af2b379fe563\" (UID: \"ff37fa2f-dc06-430a-a4fe-af2b379fe563\") " Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.273935 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ff37fa2f-dc06-430a-a4fe-af2b379fe563-rabbitmq-tls\") pod \"ff37fa2f-dc06-430a-a4fe-af2b379fe563\" (UID: \"ff37fa2f-dc06-430a-a4fe-af2b379fe563\") " Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.276790 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff37fa2f-dc06-430a-a4fe-af2b379fe563-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "ff37fa2f-dc06-430a-a4fe-af2b379fe563" (UID: "ff37fa2f-dc06-430a-a4fe-af2b379fe563"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.277093 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff37fa2f-dc06-430a-a4fe-af2b379fe563-kube-api-access-7gjz6" (OuterVolumeSpecName: "kube-api-access-7gjz6") pod "ff37fa2f-dc06-430a-a4fe-af2b379fe563" (UID: "ff37fa2f-dc06-430a-a4fe-af2b379fe563"). InnerVolumeSpecName "kube-api-access-7gjz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.277944 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff37fa2f-dc06-430a-a4fe-af2b379fe563-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "ff37fa2f-dc06-430a-a4fe-af2b379fe563" (UID: "ff37fa2f-dc06-430a-a4fe-af2b379fe563"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.277976 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff37fa2f-dc06-430a-a4fe-af2b379fe563-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "ff37fa2f-dc06-430a-a4fe-af2b379fe563" (UID: "ff37fa2f-dc06-430a-a4fe-af2b379fe563"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.278934 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff37fa2f-dc06-430a-a4fe-af2b379fe563-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "ff37fa2f-dc06-430a-a4fe-af2b379fe563" (UID: "ff37fa2f-dc06-430a-a4fe-af2b379fe563"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.279951 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "persistence") pod "ff37fa2f-dc06-430a-a4fe-af2b379fe563" (UID: "ff37fa2f-dc06-430a-a4fe-af2b379fe563"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.282366 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff37fa2f-dc06-430a-a4fe-af2b379fe563-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "ff37fa2f-dc06-430a-a4fe-af2b379fe563" (UID: "ff37fa2f-dc06-430a-a4fe-af2b379fe563"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.285213 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/ff37fa2f-dc06-430a-a4fe-af2b379fe563-pod-info" (OuterVolumeSpecName: "pod-info") pod "ff37fa2f-dc06-430a-a4fe-af2b379fe563" (UID: "ff37fa2f-dc06-430a-a4fe-af2b379fe563"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.319069 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff37fa2f-dc06-430a-a4fe-af2b379fe563-config-data" (OuterVolumeSpecName: "config-data") pod "ff37fa2f-dc06-430a-a4fe-af2b379fe563" (UID: "ff37fa2f-dc06-430a-a4fe-af2b379fe563"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.339879 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff37fa2f-dc06-430a-a4fe-af2b379fe563-server-conf" (OuterVolumeSpecName: "server-conf") pod "ff37fa2f-dc06-430a-a4fe-af2b379fe563" (UID: "ff37fa2f-dc06-430a-a4fe-af2b379fe563"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.376278 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ff37fa2f-dc06-430a-a4fe-af2b379fe563-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.376317 4824 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ff37fa2f-dc06-430a-a4fe-af2b379fe563-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.376329 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7gjz6\" (UniqueName: \"kubernetes.io/projected/ff37fa2f-dc06-430a-a4fe-af2b379fe563-kube-api-access-7gjz6\") on node \"crc\" DevicePath \"\"" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.376342 4824 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ff37fa2f-dc06-430a-a4fe-af2b379fe563-server-conf\") on node \"crc\" DevicePath \"\"" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.376356 4824 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ff37fa2f-dc06-430a-a4fe-af2b379fe563-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.376388 4824 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.376399 4824 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ff37fa2f-dc06-430a-a4fe-af2b379fe563-pod-info\") on node \"crc\" DevicePath \"\"" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.376410 4824 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ff37fa2f-dc06-430a-a4fe-af2b379fe563-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.376422 4824 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ff37fa2f-dc06-430a-a4fe-af2b379fe563-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.376433 4824 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ff37fa2f-dc06-430a-a4fe-af2b379fe563-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.380013 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff37fa2f-dc06-430a-a4fe-af2b379fe563-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "ff37fa2f-dc06-430a-a4fe-af2b379fe563" (UID: "ff37fa2f-dc06-430a-a4fe-af2b379fe563"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.401059 4824 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.467728 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.475464 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.477853 4824 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.477874 4824 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ff37fa2f-dc06-430a-a4fe-af2b379fe563-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.482680 4824 scope.go:117] "RemoveContainer" containerID="1d02f24dc933f44c12ca63c4374960b6bf5f14ea8005dd772dbd07f54f6c8f51" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.503044 4824 scope.go:117] "RemoveContainer" containerID="69c1077bbc1e742ec896e90a037d4489755f60af0094fc7227c893dc106fa13b" Dec 11 02:25:24 crc kubenswrapper[4824]: E1211 02:25:24.506769 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69c1077bbc1e742ec896e90a037d4489755f60af0094fc7227c893dc106fa13b\": container with ID starting with 69c1077bbc1e742ec896e90a037d4489755f60af0094fc7227c893dc106fa13b not found: ID does not exist" containerID="69c1077bbc1e742ec896e90a037d4489755f60af0094fc7227c893dc106fa13b" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.506839 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69c1077bbc1e742ec896e90a037d4489755f60af0094fc7227c893dc106fa13b"} err="failed to get container status \"69c1077bbc1e742ec896e90a037d4489755f60af0094fc7227c893dc106fa13b\": rpc error: code = NotFound desc = could not find container \"69c1077bbc1e742ec896e90a037d4489755f60af0094fc7227c893dc106fa13b\": container with ID starting with 69c1077bbc1e742ec896e90a037d4489755f60af0094fc7227c893dc106fa13b not found: ID does not exist" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.506868 4824 scope.go:117] "RemoveContainer" containerID="1d02f24dc933f44c12ca63c4374960b6bf5f14ea8005dd772dbd07f54f6c8f51" Dec 11 02:25:24 crc kubenswrapper[4824]: E1211 02:25:24.507496 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d02f24dc933f44c12ca63c4374960b6bf5f14ea8005dd772dbd07f54f6c8f51\": container with ID starting with 1d02f24dc933f44c12ca63c4374960b6bf5f14ea8005dd772dbd07f54f6c8f51 not found: ID does not exist" containerID="1d02f24dc933f44c12ca63c4374960b6bf5f14ea8005dd772dbd07f54f6c8f51" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.507545 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d02f24dc933f44c12ca63c4374960b6bf5f14ea8005dd772dbd07f54f6c8f51"} err="failed to get container status \"1d02f24dc933f44c12ca63c4374960b6bf5f14ea8005dd772dbd07f54f6c8f51\": rpc error: code = NotFound desc = could not find container \"1d02f24dc933f44c12ca63c4374960b6bf5f14ea8005dd772dbd07f54f6c8f51\": container with ID starting with 1d02f24dc933f44c12ca63c4374960b6bf5f14ea8005dd772dbd07f54f6c8f51 not found: ID does not exist" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.507573 4824 scope.go:117] "RemoveContainer" containerID="8dee72102c0022a859a56e54e0f34c0048fded2348fc7ecb0d3de490ad11c4cd" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.522025 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 11 02:25:24 crc kubenswrapper[4824]: E1211 02:25:24.524830 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1fd840b6-4ace-45b8-a8d2-38a041110095" containerName="rabbitmq" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.524853 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="1fd840b6-4ace-45b8-a8d2-38a041110095" containerName="rabbitmq" Dec 11 02:25:24 crc kubenswrapper[4824]: E1211 02:25:24.524873 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f02884bb-bee5-4388-a306-f1280d81dff5" containerName="registry-server" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.524880 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f02884bb-bee5-4388-a306-f1280d81dff5" containerName="registry-server" Dec 11 02:25:24 crc kubenswrapper[4824]: E1211 02:25:24.524906 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff37fa2f-dc06-430a-a4fe-af2b379fe563" containerName="rabbitmq" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.524913 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff37fa2f-dc06-430a-a4fe-af2b379fe563" containerName="rabbitmq" Dec 11 02:25:24 crc kubenswrapper[4824]: E1211 02:25:24.524927 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff37fa2f-dc06-430a-a4fe-af2b379fe563" containerName="setup-container" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.524933 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff37fa2f-dc06-430a-a4fe-af2b379fe563" containerName="setup-container" Dec 11 02:25:24 crc kubenswrapper[4824]: E1211 02:25:24.524948 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f02884bb-bee5-4388-a306-f1280d81dff5" containerName="extract-content" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.524956 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f02884bb-bee5-4388-a306-f1280d81dff5" containerName="extract-content" Dec 11 02:25:24 crc kubenswrapper[4824]: E1211 02:25:24.524979 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1fd840b6-4ace-45b8-a8d2-38a041110095" containerName="setup-container" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.524985 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="1fd840b6-4ace-45b8-a8d2-38a041110095" containerName="setup-container" Dec 11 02:25:24 crc kubenswrapper[4824]: E1211 02:25:24.524998 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f02884bb-bee5-4388-a306-f1280d81dff5" containerName="extract-utilities" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.525006 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f02884bb-bee5-4388-a306-f1280d81dff5" containerName="extract-utilities" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.525343 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="1fd840b6-4ace-45b8-a8d2-38a041110095" containerName="rabbitmq" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.525355 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f02884bb-bee5-4388-a306-f1280d81dff5" containerName="registry-server" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.525386 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff37fa2f-dc06-430a-a4fe-af2b379fe563" containerName="rabbitmq" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.528198 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.533949 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.535066 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-vtbb9" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.535282 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.535379 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.535479 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.535577 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.535738 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.535833 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.563159 4824 scope.go:117] "RemoveContainer" containerID="690854c03401211031fd0831cdfab50c3cc32ac8bb1ccb2f7ab9e7897c3a6efe" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.579013 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8799d853-613d-4376-a372-7dedd968f264-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"8799d853-613d-4376-a372-7dedd968f264\") " pod="openstack/rabbitmq-server-0" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.579059 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8799d853-613d-4376-a372-7dedd968f264-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"8799d853-613d-4376-a372-7dedd968f264\") " pod="openstack/rabbitmq-server-0" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.579083 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8799d853-613d-4376-a372-7dedd968f264-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"8799d853-613d-4376-a372-7dedd968f264\") " pod="openstack/rabbitmq-server-0" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.579207 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8799d853-613d-4376-a372-7dedd968f264-pod-info\") pod \"rabbitmq-server-0\" (UID: \"8799d853-613d-4376-a372-7dedd968f264\") " pod="openstack/rabbitmq-server-0" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.579258 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8799d853-613d-4376-a372-7dedd968f264-config-data\") pod \"rabbitmq-server-0\" (UID: \"8799d853-613d-4376-a372-7dedd968f264\") " pod="openstack/rabbitmq-server-0" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.579314 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8799d853-613d-4376-a372-7dedd968f264-server-conf\") pod \"rabbitmq-server-0\" (UID: \"8799d853-613d-4376-a372-7dedd968f264\") " pod="openstack/rabbitmq-server-0" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.579365 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sv5lp\" (UniqueName: \"kubernetes.io/projected/8799d853-613d-4376-a372-7dedd968f264-kube-api-access-sv5lp\") pod \"rabbitmq-server-0\" (UID: \"8799d853-613d-4376-a372-7dedd968f264\") " pod="openstack/rabbitmq-server-0" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.579383 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"8799d853-613d-4376-a372-7dedd968f264\") " pod="openstack/rabbitmq-server-0" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.579399 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8799d853-613d-4376-a372-7dedd968f264-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"8799d853-613d-4376-a372-7dedd968f264\") " pod="openstack/rabbitmq-server-0" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.579506 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8799d853-613d-4376-a372-7dedd968f264-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"8799d853-613d-4376-a372-7dedd968f264\") " pod="openstack/rabbitmq-server-0" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.579533 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/8799d853-613d-4376-a372-7dedd968f264-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"8799d853-613d-4376-a372-7dedd968f264\") " pod="openstack/rabbitmq-server-0" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.594045 4824 scope.go:117] "RemoveContainer" containerID="8dee72102c0022a859a56e54e0f34c0048fded2348fc7ecb0d3de490ad11c4cd" Dec 11 02:25:24 crc kubenswrapper[4824]: E1211 02:25:24.602224 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8dee72102c0022a859a56e54e0f34c0048fded2348fc7ecb0d3de490ad11c4cd\": container with ID starting with 8dee72102c0022a859a56e54e0f34c0048fded2348fc7ecb0d3de490ad11c4cd not found: ID does not exist" containerID="8dee72102c0022a859a56e54e0f34c0048fded2348fc7ecb0d3de490ad11c4cd" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.602287 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8dee72102c0022a859a56e54e0f34c0048fded2348fc7ecb0d3de490ad11c4cd"} err="failed to get container status \"8dee72102c0022a859a56e54e0f34c0048fded2348fc7ecb0d3de490ad11c4cd\": rpc error: code = NotFound desc = could not find container \"8dee72102c0022a859a56e54e0f34c0048fded2348fc7ecb0d3de490ad11c4cd\": container with ID starting with 8dee72102c0022a859a56e54e0f34c0048fded2348fc7ecb0d3de490ad11c4cd not found: ID does not exist" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.602315 4824 scope.go:117] "RemoveContainer" containerID="690854c03401211031fd0831cdfab50c3cc32ac8bb1ccb2f7ab9e7897c3a6efe" Dec 11 02:25:24 crc kubenswrapper[4824]: E1211 02:25:24.602741 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"690854c03401211031fd0831cdfab50c3cc32ac8bb1ccb2f7ab9e7897c3a6efe\": container with ID starting with 690854c03401211031fd0831cdfab50c3cc32ac8bb1ccb2f7ab9e7897c3a6efe not found: ID does not exist" containerID="690854c03401211031fd0831cdfab50c3cc32ac8bb1ccb2f7ab9e7897c3a6efe" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.602785 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"690854c03401211031fd0831cdfab50c3cc32ac8bb1ccb2f7ab9e7897c3a6efe"} err="failed to get container status \"690854c03401211031fd0831cdfab50c3cc32ac8bb1ccb2f7ab9e7897c3a6efe\": rpc error: code = NotFound desc = could not find container \"690854c03401211031fd0831cdfab50c3cc32ac8bb1ccb2f7ab9e7897c3a6efe\": container with ID starting with 690854c03401211031fd0831cdfab50c3cc32ac8bb1ccb2f7ab9e7897c3a6efe not found: ID does not exist" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.617735 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.641887 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1fd840b6-4ace-45b8-a8d2-38a041110095" path="/var/lib/kubelet/pods/1fd840b6-4ace-45b8-a8d2-38a041110095/volumes" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.642545 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.651639 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.653856 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.655951 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.656245 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.656960 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.658665 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.658821 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.659149 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-8mmtq" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.659532 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.662784 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.682558 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8799d853-613d-4376-a372-7dedd968f264-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"8799d853-613d-4376-a372-7dedd968f264\") " pod="openstack/rabbitmq-server-0" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.682600 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/89eddaec-a9d6-4b73-b654-2f176d45d162-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"89eddaec-a9d6-4b73-b654-2f176d45d162\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.682619 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/89eddaec-a9d6-4b73-b654-2f176d45d162-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"89eddaec-a9d6-4b73-b654-2f176d45d162\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.682670 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8799d853-613d-4376-a372-7dedd968f264-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"8799d853-613d-4376-a372-7dedd968f264\") " pod="openstack/rabbitmq-server-0" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.682687 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6zrq\" (UniqueName: \"kubernetes.io/projected/89eddaec-a9d6-4b73-b654-2f176d45d162-kube-api-access-v6zrq\") pod \"rabbitmq-cell1-server-0\" (UID: \"89eddaec-a9d6-4b73-b654-2f176d45d162\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.682721 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8799d853-613d-4376-a372-7dedd968f264-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"8799d853-613d-4376-a372-7dedd968f264\") " pod="openstack/rabbitmq-server-0" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.682742 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/89eddaec-a9d6-4b73-b654-2f176d45d162-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"89eddaec-a9d6-4b73-b654-2f176d45d162\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.682839 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/89eddaec-a9d6-4b73-b654-2f176d45d162-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"89eddaec-a9d6-4b73-b654-2f176d45d162\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.682859 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8799d853-613d-4376-a372-7dedd968f264-pod-info\") pod \"rabbitmq-server-0\" (UID: \"8799d853-613d-4376-a372-7dedd968f264\") " pod="openstack/rabbitmq-server-0" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.682915 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/89eddaec-a9d6-4b73-b654-2f176d45d162-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"89eddaec-a9d6-4b73-b654-2f176d45d162\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.682984 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8799d853-613d-4376-a372-7dedd968f264-config-data\") pod \"rabbitmq-server-0\" (UID: \"8799d853-613d-4376-a372-7dedd968f264\") " pod="openstack/rabbitmq-server-0" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.683018 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8799d853-613d-4376-a372-7dedd968f264-server-conf\") pod \"rabbitmq-server-0\" (UID: \"8799d853-613d-4376-a372-7dedd968f264\") " pod="openstack/rabbitmq-server-0" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.683094 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"89eddaec-a9d6-4b73-b654-2f176d45d162\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.683189 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/89eddaec-a9d6-4b73-b654-2f176d45d162-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"89eddaec-a9d6-4b73-b654-2f176d45d162\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.683233 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sv5lp\" (UniqueName: \"kubernetes.io/projected/8799d853-613d-4376-a372-7dedd968f264-kube-api-access-sv5lp\") pod \"rabbitmq-server-0\" (UID: \"8799d853-613d-4376-a372-7dedd968f264\") " pod="openstack/rabbitmq-server-0" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.683254 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"8799d853-613d-4376-a372-7dedd968f264\") " pod="openstack/rabbitmq-server-0" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.683268 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8799d853-613d-4376-a372-7dedd968f264-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"8799d853-613d-4376-a372-7dedd968f264\") " pod="openstack/rabbitmq-server-0" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.683282 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/89eddaec-a9d6-4b73-b654-2f176d45d162-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"89eddaec-a9d6-4b73-b654-2f176d45d162\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.683304 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/89eddaec-a9d6-4b73-b654-2f176d45d162-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"89eddaec-a9d6-4b73-b654-2f176d45d162\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.683331 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8799d853-613d-4376-a372-7dedd968f264-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"8799d853-613d-4376-a372-7dedd968f264\") " pod="openstack/rabbitmq-server-0" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.683376 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/8799d853-613d-4376-a372-7dedd968f264-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"8799d853-613d-4376-a372-7dedd968f264\") " pod="openstack/rabbitmq-server-0" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.683393 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/89eddaec-a9d6-4b73-b654-2f176d45d162-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"89eddaec-a9d6-4b73-b654-2f176d45d162\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.683632 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8799d853-613d-4376-a372-7dedd968f264-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"8799d853-613d-4376-a372-7dedd968f264\") " pod="openstack/rabbitmq-server-0" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.684841 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8799d853-613d-4376-a372-7dedd968f264-config-data\") pod \"rabbitmq-server-0\" (UID: \"8799d853-613d-4376-a372-7dedd968f264\") " pod="openstack/rabbitmq-server-0" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.684988 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8799d853-613d-4376-a372-7dedd968f264-server-conf\") pod \"rabbitmq-server-0\" (UID: \"8799d853-613d-4376-a372-7dedd968f264\") " pod="openstack/rabbitmq-server-0" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.685092 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"8799d853-613d-4376-a372-7dedd968f264\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/rabbitmq-server-0" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.685615 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8799d853-613d-4376-a372-7dedd968f264-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"8799d853-613d-4376-a372-7dedd968f264\") " pod="openstack/rabbitmq-server-0" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.686587 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8799d853-613d-4376-a372-7dedd968f264-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"8799d853-613d-4376-a372-7dedd968f264\") " pod="openstack/rabbitmq-server-0" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.686771 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8799d853-613d-4376-a372-7dedd968f264-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"8799d853-613d-4376-a372-7dedd968f264\") " pod="openstack/rabbitmq-server-0" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.689394 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8799d853-613d-4376-a372-7dedd968f264-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"8799d853-613d-4376-a372-7dedd968f264\") " pod="openstack/rabbitmq-server-0" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.689489 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/8799d853-613d-4376-a372-7dedd968f264-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"8799d853-613d-4376-a372-7dedd968f264\") " pod="openstack/rabbitmq-server-0" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.691630 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8799d853-613d-4376-a372-7dedd968f264-pod-info\") pod \"rabbitmq-server-0\" (UID: \"8799d853-613d-4376-a372-7dedd968f264\") " pod="openstack/rabbitmq-server-0" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.699051 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sv5lp\" (UniqueName: \"kubernetes.io/projected/8799d853-613d-4376-a372-7dedd968f264-kube-api-access-sv5lp\") pod \"rabbitmq-server-0\" (UID: \"8799d853-613d-4376-a372-7dedd968f264\") " pod="openstack/rabbitmq-server-0" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.721276 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"8799d853-613d-4376-a372-7dedd968f264\") " pod="openstack/rabbitmq-server-0" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.786016 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6zrq\" (UniqueName: \"kubernetes.io/projected/89eddaec-a9d6-4b73-b654-2f176d45d162-kube-api-access-v6zrq\") pod \"rabbitmq-cell1-server-0\" (UID: \"89eddaec-a9d6-4b73-b654-2f176d45d162\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.786062 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/89eddaec-a9d6-4b73-b654-2f176d45d162-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"89eddaec-a9d6-4b73-b654-2f176d45d162\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.786862 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/89eddaec-a9d6-4b73-b654-2f176d45d162-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"89eddaec-a9d6-4b73-b654-2f176d45d162\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.786927 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/89eddaec-a9d6-4b73-b654-2f176d45d162-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"89eddaec-a9d6-4b73-b654-2f176d45d162\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.786962 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/89eddaec-a9d6-4b73-b654-2f176d45d162-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"89eddaec-a9d6-4b73-b654-2f176d45d162\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.787020 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"89eddaec-a9d6-4b73-b654-2f176d45d162\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.787038 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/89eddaec-a9d6-4b73-b654-2f176d45d162-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"89eddaec-a9d6-4b73-b654-2f176d45d162\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.787063 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/89eddaec-a9d6-4b73-b654-2f176d45d162-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"89eddaec-a9d6-4b73-b654-2f176d45d162\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.787080 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/89eddaec-a9d6-4b73-b654-2f176d45d162-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"89eddaec-a9d6-4b73-b654-2f176d45d162\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.787122 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/89eddaec-a9d6-4b73-b654-2f176d45d162-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"89eddaec-a9d6-4b73-b654-2f176d45d162\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.787156 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/89eddaec-a9d6-4b73-b654-2f176d45d162-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"89eddaec-a9d6-4b73-b654-2f176d45d162\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.787171 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/89eddaec-a9d6-4b73-b654-2f176d45d162-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"89eddaec-a9d6-4b73-b654-2f176d45d162\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.787472 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/89eddaec-a9d6-4b73-b654-2f176d45d162-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"89eddaec-a9d6-4b73-b654-2f176d45d162\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.789815 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"89eddaec-a9d6-4b73-b654-2f176d45d162\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-cell1-server-0" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.790784 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/89eddaec-a9d6-4b73-b654-2f176d45d162-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"89eddaec-a9d6-4b73-b654-2f176d45d162\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.791242 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/89eddaec-a9d6-4b73-b654-2f176d45d162-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"89eddaec-a9d6-4b73-b654-2f176d45d162\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.795050 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/89eddaec-a9d6-4b73-b654-2f176d45d162-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"89eddaec-a9d6-4b73-b654-2f176d45d162\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.796428 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/89eddaec-a9d6-4b73-b654-2f176d45d162-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"89eddaec-a9d6-4b73-b654-2f176d45d162\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.798272 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/89eddaec-a9d6-4b73-b654-2f176d45d162-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"89eddaec-a9d6-4b73-b654-2f176d45d162\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.803587 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/89eddaec-a9d6-4b73-b654-2f176d45d162-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"89eddaec-a9d6-4b73-b654-2f176d45d162\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.804308 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6zrq\" (UniqueName: \"kubernetes.io/projected/89eddaec-a9d6-4b73-b654-2f176d45d162-kube-api-access-v6zrq\") pod \"rabbitmq-cell1-server-0\" (UID: \"89eddaec-a9d6-4b73-b654-2f176d45d162\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.804530 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/89eddaec-a9d6-4b73-b654-2f176d45d162-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"89eddaec-a9d6-4b73-b654-2f176d45d162\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.818170 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"89eddaec-a9d6-4b73-b654-2f176d45d162\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.870302 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 11 02:25:24 crc kubenswrapper[4824]: I1211 02:25:24.983623 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 11 02:25:25 crc kubenswrapper[4824]: I1211 02:25:25.151606 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 11 02:25:25 crc kubenswrapper[4824]: I1211 02:25:25.279323 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"8799d853-613d-4376-a372-7dedd968f264","Type":"ContainerStarted","Data":"1dac47a5fcad74e1bd5c1a3d34a9b53836450fea505c5543130f4979952945e6"} Dec 11 02:25:25 crc kubenswrapper[4824]: W1211 02:25:25.452328 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod89eddaec_a9d6_4b73_b654_2f176d45d162.slice/crio-1380f271d3e01162a8547d29c35713b831913c50d354caf83bb63c6ddad43d79 WatchSource:0}: Error finding container 1380f271d3e01162a8547d29c35713b831913c50d354caf83bb63c6ddad43d79: Status 404 returned error can't find the container with id 1380f271d3e01162a8547d29c35713b831913c50d354caf83bb63c6ddad43d79 Dec 11 02:25:25 crc kubenswrapper[4824]: I1211 02:25:25.454630 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 11 02:25:26 crc kubenswrapper[4824]: I1211 02:25:26.291734 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"89eddaec-a9d6-4b73-b654-2f176d45d162","Type":"ContainerStarted","Data":"1380f271d3e01162a8547d29c35713b831913c50d354caf83bb63c6ddad43d79"} Dec 11 02:25:26 crc kubenswrapper[4824]: I1211 02:25:26.648927 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff37fa2f-dc06-430a-a4fe-af2b379fe563" path="/var/lib/kubelet/pods/ff37fa2f-dc06-430a-a4fe-af2b379fe563/volumes" Dec 11 02:25:27 crc kubenswrapper[4824]: I1211 02:25:27.306101 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"8799d853-613d-4376-a372-7dedd968f264","Type":"ContainerStarted","Data":"ed43359051393901062422faf9762d9d95dde215edf8efe06d58883210e99376"} Dec 11 02:25:27 crc kubenswrapper[4824]: I1211 02:25:27.308528 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"89eddaec-a9d6-4b73-b654-2f176d45d162","Type":"ContainerStarted","Data":"5ec7fbef36d90f11e756f4556f943f2f69d171390446f82e5a27da5b1f3e4952"} Dec 11 02:25:28 crc kubenswrapper[4824]: I1211 02:25:28.100877 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-s8lft"] Dec 11 02:25:28 crc kubenswrapper[4824]: I1211 02:25:28.105626 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-s8lft" Dec 11 02:25:28 crc kubenswrapper[4824]: I1211 02:25:28.108749 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Dec 11 02:25:28 crc kubenswrapper[4824]: I1211 02:25:28.130238 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-s8lft"] Dec 11 02:25:28 crc kubenswrapper[4824]: I1211 02:25:28.257860 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/74b07b56-f514-4afe-9715-9f42b144f066-ovsdbserver-nb\") pod \"dnsmasq-dns-79bd4cc8c9-s8lft\" (UID: \"74b07b56-f514-4afe-9715-9f42b144f066\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-s8lft" Dec 11 02:25:28 crc kubenswrapper[4824]: I1211 02:25:28.257939 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/74b07b56-f514-4afe-9715-9f42b144f066-dns-swift-storage-0\") pod \"dnsmasq-dns-79bd4cc8c9-s8lft\" (UID: \"74b07b56-f514-4afe-9715-9f42b144f066\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-s8lft" Dec 11 02:25:28 crc kubenswrapper[4824]: I1211 02:25:28.258096 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/74b07b56-f514-4afe-9715-9f42b144f066-dns-svc\") pod \"dnsmasq-dns-79bd4cc8c9-s8lft\" (UID: \"74b07b56-f514-4afe-9715-9f42b144f066\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-s8lft" Dec 11 02:25:28 crc kubenswrapper[4824]: I1211 02:25:28.258250 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/74b07b56-f514-4afe-9715-9f42b144f066-ovsdbserver-sb\") pod \"dnsmasq-dns-79bd4cc8c9-s8lft\" (UID: \"74b07b56-f514-4afe-9715-9f42b144f066\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-s8lft" Dec 11 02:25:28 crc kubenswrapper[4824]: I1211 02:25:28.258292 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h5wfl\" (UniqueName: \"kubernetes.io/projected/74b07b56-f514-4afe-9715-9f42b144f066-kube-api-access-h5wfl\") pod \"dnsmasq-dns-79bd4cc8c9-s8lft\" (UID: \"74b07b56-f514-4afe-9715-9f42b144f066\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-s8lft" Dec 11 02:25:28 crc kubenswrapper[4824]: I1211 02:25:28.258326 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/74b07b56-f514-4afe-9715-9f42b144f066-openstack-edpm-ipam\") pod \"dnsmasq-dns-79bd4cc8c9-s8lft\" (UID: \"74b07b56-f514-4afe-9715-9f42b144f066\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-s8lft" Dec 11 02:25:28 crc kubenswrapper[4824]: I1211 02:25:28.258377 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74b07b56-f514-4afe-9715-9f42b144f066-config\") pod \"dnsmasq-dns-79bd4cc8c9-s8lft\" (UID: \"74b07b56-f514-4afe-9715-9f42b144f066\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-s8lft" Dec 11 02:25:28 crc kubenswrapper[4824]: I1211 02:25:28.359693 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/74b07b56-f514-4afe-9715-9f42b144f066-dns-swift-storage-0\") pod \"dnsmasq-dns-79bd4cc8c9-s8lft\" (UID: \"74b07b56-f514-4afe-9715-9f42b144f066\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-s8lft" Dec 11 02:25:28 crc kubenswrapper[4824]: I1211 02:25:28.359761 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/74b07b56-f514-4afe-9715-9f42b144f066-dns-svc\") pod \"dnsmasq-dns-79bd4cc8c9-s8lft\" (UID: \"74b07b56-f514-4afe-9715-9f42b144f066\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-s8lft" Dec 11 02:25:28 crc kubenswrapper[4824]: I1211 02:25:28.359817 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/74b07b56-f514-4afe-9715-9f42b144f066-ovsdbserver-sb\") pod \"dnsmasq-dns-79bd4cc8c9-s8lft\" (UID: \"74b07b56-f514-4afe-9715-9f42b144f066\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-s8lft" Dec 11 02:25:28 crc kubenswrapper[4824]: I1211 02:25:28.359839 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h5wfl\" (UniqueName: \"kubernetes.io/projected/74b07b56-f514-4afe-9715-9f42b144f066-kube-api-access-h5wfl\") pod \"dnsmasq-dns-79bd4cc8c9-s8lft\" (UID: \"74b07b56-f514-4afe-9715-9f42b144f066\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-s8lft" Dec 11 02:25:28 crc kubenswrapper[4824]: I1211 02:25:28.359867 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/74b07b56-f514-4afe-9715-9f42b144f066-openstack-edpm-ipam\") pod \"dnsmasq-dns-79bd4cc8c9-s8lft\" (UID: \"74b07b56-f514-4afe-9715-9f42b144f066\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-s8lft" Dec 11 02:25:28 crc kubenswrapper[4824]: I1211 02:25:28.359900 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74b07b56-f514-4afe-9715-9f42b144f066-config\") pod \"dnsmasq-dns-79bd4cc8c9-s8lft\" (UID: \"74b07b56-f514-4afe-9715-9f42b144f066\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-s8lft" Dec 11 02:25:28 crc kubenswrapper[4824]: I1211 02:25:28.359994 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/74b07b56-f514-4afe-9715-9f42b144f066-ovsdbserver-nb\") pod \"dnsmasq-dns-79bd4cc8c9-s8lft\" (UID: \"74b07b56-f514-4afe-9715-9f42b144f066\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-s8lft" Dec 11 02:25:28 crc kubenswrapper[4824]: I1211 02:25:28.361298 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74b07b56-f514-4afe-9715-9f42b144f066-config\") pod \"dnsmasq-dns-79bd4cc8c9-s8lft\" (UID: \"74b07b56-f514-4afe-9715-9f42b144f066\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-s8lft" Dec 11 02:25:28 crc kubenswrapper[4824]: I1211 02:25:28.361518 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/74b07b56-f514-4afe-9715-9f42b144f066-ovsdbserver-sb\") pod \"dnsmasq-dns-79bd4cc8c9-s8lft\" (UID: \"74b07b56-f514-4afe-9715-9f42b144f066\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-s8lft" Dec 11 02:25:28 crc kubenswrapper[4824]: I1211 02:25:28.361533 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/74b07b56-f514-4afe-9715-9f42b144f066-openstack-edpm-ipam\") pod \"dnsmasq-dns-79bd4cc8c9-s8lft\" (UID: \"74b07b56-f514-4afe-9715-9f42b144f066\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-s8lft" Dec 11 02:25:28 crc kubenswrapper[4824]: I1211 02:25:28.361591 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/74b07b56-f514-4afe-9715-9f42b144f066-ovsdbserver-nb\") pod \"dnsmasq-dns-79bd4cc8c9-s8lft\" (UID: \"74b07b56-f514-4afe-9715-9f42b144f066\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-s8lft" Dec 11 02:25:28 crc kubenswrapper[4824]: I1211 02:25:28.361931 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/74b07b56-f514-4afe-9715-9f42b144f066-dns-svc\") pod \"dnsmasq-dns-79bd4cc8c9-s8lft\" (UID: \"74b07b56-f514-4afe-9715-9f42b144f066\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-s8lft" Dec 11 02:25:28 crc kubenswrapper[4824]: I1211 02:25:28.362175 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/74b07b56-f514-4afe-9715-9f42b144f066-dns-swift-storage-0\") pod \"dnsmasq-dns-79bd4cc8c9-s8lft\" (UID: \"74b07b56-f514-4afe-9715-9f42b144f066\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-s8lft" Dec 11 02:25:28 crc kubenswrapper[4824]: I1211 02:25:28.378070 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h5wfl\" (UniqueName: \"kubernetes.io/projected/74b07b56-f514-4afe-9715-9f42b144f066-kube-api-access-h5wfl\") pod \"dnsmasq-dns-79bd4cc8c9-s8lft\" (UID: \"74b07b56-f514-4afe-9715-9f42b144f066\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-s8lft" Dec 11 02:25:28 crc kubenswrapper[4824]: I1211 02:25:28.444018 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-s8lft" Dec 11 02:25:29 crc kubenswrapper[4824]: I1211 02:25:29.022437 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-s8lft"] Dec 11 02:25:29 crc kubenswrapper[4824]: I1211 02:25:29.330347 4824 generic.go:334] "Generic (PLEG): container finished" podID="74b07b56-f514-4afe-9715-9f42b144f066" containerID="95cefc4c5cdf946297d7b1ed190c5ddc45f510eefea53ed37ce92e5daee34aa5" exitCode=0 Dec 11 02:25:29 crc kubenswrapper[4824]: I1211 02:25:29.330377 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-s8lft" event={"ID":"74b07b56-f514-4afe-9715-9f42b144f066","Type":"ContainerDied","Data":"95cefc4c5cdf946297d7b1ed190c5ddc45f510eefea53ed37ce92e5daee34aa5"} Dec 11 02:25:29 crc kubenswrapper[4824]: I1211 02:25:29.330746 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-s8lft" event={"ID":"74b07b56-f514-4afe-9715-9f42b144f066","Type":"ContainerStarted","Data":"383d0eefac77c72252ab4edc2dc9f634ae1298032c1b6c9284614fb4384ece73"} Dec 11 02:25:30 crc kubenswrapper[4824]: I1211 02:25:30.345192 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-s8lft" event={"ID":"74b07b56-f514-4afe-9715-9f42b144f066","Type":"ContainerStarted","Data":"3a4eac77f7a127683859cee1c67d38bf38413baa0c24d41f15af3eb0b9adaa95"} Dec 11 02:25:30 crc kubenswrapper[4824]: I1211 02:25:30.345820 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-79bd4cc8c9-s8lft" Dec 11 02:25:30 crc kubenswrapper[4824]: I1211 02:25:30.382725 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-79bd4cc8c9-s8lft" podStartSLOduration=2.382693349 podStartE2EDuration="2.382693349s" podCreationTimestamp="2025-12-11 02:25:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:25:30.377312222 +0000 UTC m=+1472.066349681" watchObservedRunningTime="2025-12-11 02:25:30.382693349 +0000 UTC m=+1472.071730768" Dec 11 02:25:37 crc kubenswrapper[4824]: I1211 02:25:37.280653 4824 patch_prober.go:28] interesting pod/machine-config-daemon-gx6xt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 02:25:37 crc kubenswrapper[4824]: I1211 02:25:37.281392 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 02:25:38 crc kubenswrapper[4824]: I1211 02:25:38.446519 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-79bd4cc8c9-s8lft" Dec 11 02:25:38 crc kubenswrapper[4824]: I1211 02:25:38.595750 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-n9sd5"] Dec 11 02:25:38 crc kubenswrapper[4824]: I1211 02:25:38.596273 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-89c5cd4d5-n9sd5" podUID="c358674b-3741-4d28-a629-029d94cdc3d4" containerName="dnsmasq-dns" containerID="cri-o://bd89a5a921ae19774166b4f494c2bb2e0ce856de55580b2d902ebdb5c24f7d9e" gracePeriod=10 Dec 11 02:25:38 crc kubenswrapper[4824]: I1211 02:25:38.735383 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-55478c4467-mptxc"] Dec 11 02:25:38 crc kubenswrapper[4824]: I1211 02:25:38.736940 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55478c4467-mptxc" Dec 11 02:25:38 crc kubenswrapper[4824]: I1211 02:25:38.753007 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55478c4467-mptxc"] Dec 11 02:25:38 crc kubenswrapper[4824]: I1211 02:25:38.825391 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/71c2c849-fc4f-44ac-92d4-75848ea56874-ovsdbserver-nb\") pod \"dnsmasq-dns-55478c4467-mptxc\" (UID: \"71c2c849-fc4f-44ac-92d4-75848ea56874\") " pod="openstack/dnsmasq-dns-55478c4467-mptxc" Dec 11 02:25:38 crc kubenswrapper[4824]: I1211 02:25:38.825470 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/71c2c849-fc4f-44ac-92d4-75848ea56874-dns-swift-storage-0\") pod \"dnsmasq-dns-55478c4467-mptxc\" (UID: \"71c2c849-fc4f-44ac-92d4-75848ea56874\") " pod="openstack/dnsmasq-dns-55478c4467-mptxc" Dec 11 02:25:38 crc kubenswrapper[4824]: I1211 02:25:38.825495 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/71c2c849-fc4f-44ac-92d4-75848ea56874-openstack-edpm-ipam\") pod \"dnsmasq-dns-55478c4467-mptxc\" (UID: \"71c2c849-fc4f-44ac-92d4-75848ea56874\") " pod="openstack/dnsmasq-dns-55478c4467-mptxc" Dec 11 02:25:38 crc kubenswrapper[4824]: I1211 02:25:38.825547 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/71c2c849-fc4f-44ac-92d4-75848ea56874-ovsdbserver-sb\") pod \"dnsmasq-dns-55478c4467-mptxc\" (UID: \"71c2c849-fc4f-44ac-92d4-75848ea56874\") " pod="openstack/dnsmasq-dns-55478c4467-mptxc" Dec 11 02:25:38 crc kubenswrapper[4824]: I1211 02:25:38.825564 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/71c2c849-fc4f-44ac-92d4-75848ea56874-dns-svc\") pod \"dnsmasq-dns-55478c4467-mptxc\" (UID: \"71c2c849-fc4f-44ac-92d4-75848ea56874\") " pod="openstack/dnsmasq-dns-55478c4467-mptxc" Dec 11 02:25:38 crc kubenswrapper[4824]: I1211 02:25:38.825625 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9bzsk\" (UniqueName: \"kubernetes.io/projected/71c2c849-fc4f-44ac-92d4-75848ea56874-kube-api-access-9bzsk\") pod \"dnsmasq-dns-55478c4467-mptxc\" (UID: \"71c2c849-fc4f-44ac-92d4-75848ea56874\") " pod="openstack/dnsmasq-dns-55478c4467-mptxc" Dec 11 02:25:38 crc kubenswrapper[4824]: I1211 02:25:38.825663 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/71c2c849-fc4f-44ac-92d4-75848ea56874-config\") pod \"dnsmasq-dns-55478c4467-mptxc\" (UID: \"71c2c849-fc4f-44ac-92d4-75848ea56874\") " pod="openstack/dnsmasq-dns-55478c4467-mptxc" Dec 11 02:25:38 crc kubenswrapper[4824]: I1211 02:25:38.927251 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/71c2c849-fc4f-44ac-92d4-75848ea56874-ovsdbserver-sb\") pod \"dnsmasq-dns-55478c4467-mptxc\" (UID: \"71c2c849-fc4f-44ac-92d4-75848ea56874\") " pod="openstack/dnsmasq-dns-55478c4467-mptxc" Dec 11 02:25:38 crc kubenswrapper[4824]: I1211 02:25:38.927639 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/71c2c849-fc4f-44ac-92d4-75848ea56874-dns-svc\") pod \"dnsmasq-dns-55478c4467-mptxc\" (UID: \"71c2c849-fc4f-44ac-92d4-75848ea56874\") " pod="openstack/dnsmasq-dns-55478c4467-mptxc" Dec 11 02:25:38 crc kubenswrapper[4824]: I1211 02:25:38.927677 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9bzsk\" (UniqueName: \"kubernetes.io/projected/71c2c849-fc4f-44ac-92d4-75848ea56874-kube-api-access-9bzsk\") pod \"dnsmasq-dns-55478c4467-mptxc\" (UID: \"71c2c849-fc4f-44ac-92d4-75848ea56874\") " pod="openstack/dnsmasq-dns-55478c4467-mptxc" Dec 11 02:25:38 crc kubenswrapper[4824]: I1211 02:25:38.927715 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/71c2c849-fc4f-44ac-92d4-75848ea56874-config\") pod \"dnsmasq-dns-55478c4467-mptxc\" (UID: \"71c2c849-fc4f-44ac-92d4-75848ea56874\") " pod="openstack/dnsmasq-dns-55478c4467-mptxc" Dec 11 02:25:38 crc kubenswrapper[4824]: I1211 02:25:38.928073 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/71c2c849-fc4f-44ac-92d4-75848ea56874-ovsdbserver-sb\") pod \"dnsmasq-dns-55478c4467-mptxc\" (UID: \"71c2c849-fc4f-44ac-92d4-75848ea56874\") " pod="openstack/dnsmasq-dns-55478c4467-mptxc" Dec 11 02:25:38 crc kubenswrapper[4824]: I1211 02:25:38.928154 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/71c2c849-fc4f-44ac-92d4-75848ea56874-ovsdbserver-nb\") pod \"dnsmasq-dns-55478c4467-mptxc\" (UID: \"71c2c849-fc4f-44ac-92d4-75848ea56874\") " pod="openstack/dnsmasq-dns-55478c4467-mptxc" Dec 11 02:25:38 crc kubenswrapper[4824]: I1211 02:25:38.928213 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/71c2c849-fc4f-44ac-92d4-75848ea56874-dns-swift-storage-0\") pod \"dnsmasq-dns-55478c4467-mptxc\" (UID: \"71c2c849-fc4f-44ac-92d4-75848ea56874\") " pod="openstack/dnsmasq-dns-55478c4467-mptxc" Dec 11 02:25:38 crc kubenswrapper[4824]: I1211 02:25:38.928238 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/71c2c849-fc4f-44ac-92d4-75848ea56874-openstack-edpm-ipam\") pod \"dnsmasq-dns-55478c4467-mptxc\" (UID: \"71c2c849-fc4f-44ac-92d4-75848ea56874\") " pod="openstack/dnsmasq-dns-55478c4467-mptxc" Dec 11 02:25:38 crc kubenswrapper[4824]: I1211 02:25:38.928306 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/71c2c849-fc4f-44ac-92d4-75848ea56874-dns-svc\") pod \"dnsmasq-dns-55478c4467-mptxc\" (UID: \"71c2c849-fc4f-44ac-92d4-75848ea56874\") " pod="openstack/dnsmasq-dns-55478c4467-mptxc" Dec 11 02:25:38 crc kubenswrapper[4824]: I1211 02:25:38.928633 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/71c2c849-fc4f-44ac-92d4-75848ea56874-config\") pod \"dnsmasq-dns-55478c4467-mptxc\" (UID: \"71c2c849-fc4f-44ac-92d4-75848ea56874\") " pod="openstack/dnsmasq-dns-55478c4467-mptxc" Dec 11 02:25:38 crc kubenswrapper[4824]: I1211 02:25:38.928853 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/71c2c849-fc4f-44ac-92d4-75848ea56874-ovsdbserver-nb\") pod \"dnsmasq-dns-55478c4467-mptxc\" (UID: \"71c2c849-fc4f-44ac-92d4-75848ea56874\") " pod="openstack/dnsmasq-dns-55478c4467-mptxc" Dec 11 02:25:38 crc kubenswrapper[4824]: I1211 02:25:38.928859 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/71c2c849-fc4f-44ac-92d4-75848ea56874-openstack-edpm-ipam\") pod \"dnsmasq-dns-55478c4467-mptxc\" (UID: \"71c2c849-fc4f-44ac-92d4-75848ea56874\") " pod="openstack/dnsmasq-dns-55478c4467-mptxc" Dec 11 02:25:38 crc kubenswrapper[4824]: I1211 02:25:38.929833 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/71c2c849-fc4f-44ac-92d4-75848ea56874-dns-swift-storage-0\") pod \"dnsmasq-dns-55478c4467-mptxc\" (UID: \"71c2c849-fc4f-44ac-92d4-75848ea56874\") " pod="openstack/dnsmasq-dns-55478c4467-mptxc" Dec 11 02:25:38 crc kubenswrapper[4824]: I1211 02:25:38.947812 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9bzsk\" (UniqueName: \"kubernetes.io/projected/71c2c849-fc4f-44ac-92d4-75848ea56874-kube-api-access-9bzsk\") pod \"dnsmasq-dns-55478c4467-mptxc\" (UID: \"71c2c849-fc4f-44ac-92d4-75848ea56874\") " pod="openstack/dnsmasq-dns-55478c4467-mptxc" Dec 11 02:25:39 crc kubenswrapper[4824]: I1211 02:25:39.031674 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-n9sd5" Dec 11 02:25:39 crc kubenswrapper[4824]: I1211 02:25:39.080464 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55478c4467-mptxc" Dec 11 02:25:39 crc kubenswrapper[4824]: I1211 02:25:39.130993 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c358674b-3741-4d28-a629-029d94cdc3d4-dns-swift-storage-0\") pod \"c358674b-3741-4d28-a629-029d94cdc3d4\" (UID: \"c358674b-3741-4d28-a629-029d94cdc3d4\") " Dec 11 02:25:39 crc kubenswrapper[4824]: I1211 02:25:39.131193 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nmnsp\" (UniqueName: \"kubernetes.io/projected/c358674b-3741-4d28-a629-029d94cdc3d4-kube-api-access-nmnsp\") pod \"c358674b-3741-4d28-a629-029d94cdc3d4\" (UID: \"c358674b-3741-4d28-a629-029d94cdc3d4\") " Dec 11 02:25:39 crc kubenswrapper[4824]: I1211 02:25:39.131321 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c358674b-3741-4d28-a629-029d94cdc3d4-config\") pod \"c358674b-3741-4d28-a629-029d94cdc3d4\" (UID: \"c358674b-3741-4d28-a629-029d94cdc3d4\") " Dec 11 02:25:39 crc kubenswrapper[4824]: I1211 02:25:39.131459 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c358674b-3741-4d28-a629-029d94cdc3d4-ovsdbserver-sb\") pod \"c358674b-3741-4d28-a629-029d94cdc3d4\" (UID: \"c358674b-3741-4d28-a629-029d94cdc3d4\") " Dec 11 02:25:39 crc kubenswrapper[4824]: I1211 02:25:39.131529 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c358674b-3741-4d28-a629-029d94cdc3d4-dns-svc\") pod \"c358674b-3741-4d28-a629-029d94cdc3d4\" (UID: \"c358674b-3741-4d28-a629-029d94cdc3d4\") " Dec 11 02:25:39 crc kubenswrapper[4824]: I1211 02:25:39.131573 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c358674b-3741-4d28-a629-029d94cdc3d4-ovsdbserver-nb\") pod \"c358674b-3741-4d28-a629-029d94cdc3d4\" (UID: \"c358674b-3741-4d28-a629-029d94cdc3d4\") " Dec 11 02:25:39 crc kubenswrapper[4824]: I1211 02:25:39.143217 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c358674b-3741-4d28-a629-029d94cdc3d4-kube-api-access-nmnsp" (OuterVolumeSpecName: "kube-api-access-nmnsp") pod "c358674b-3741-4d28-a629-029d94cdc3d4" (UID: "c358674b-3741-4d28-a629-029d94cdc3d4"). InnerVolumeSpecName "kube-api-access-nmnsp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:25:39 crc kubenswrapper[4824]: I1211 02:25:39.180353 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c358674b-3741-4d28-a629-029d94cdc3d4-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c358674b-3741-4d28-a629-029d94cdc3d4" (UID: "c358674b-3741-4d28-a629-029d94cdc3d4"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:25:39 crc kubenswrapper[4824]: I1211 02:25:39.183053 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c358674b-3741-4d28-a629-029d94cdc3d4-config" (OuterVolumeSpecName: "config") pod "c358674b-3741-4d28-a629-029d94cdc3d4" (UID: "c358674b-3741-4d28-a629-029d94cdc3d4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:25:39 crc kubenswrapper[4824]: I1211 02:25:39.185397 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c358674b-3741-4d28-a629-029d94cdc3d4-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "c358674b-3741-4d28-a629-029d94cdc3d4" (UID: "c358674b-3741-4d28-a629-029d94cdc3d4"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:25:39 crc kubenswrapper[4824]: I1211 02:25:39.191708 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c358674b-3741-4d28-a629-029d94cdc3d4-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c358674b-3741-4d28-a629-029d94cdc3d4" (UID: "c358674b-3741-4d28-a629-029d94cdc3d4"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:25:39 crc kubenswrapper[4824]: I1211 02:25:39.202760 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c358674b-3741-4d28-a629-029d94cdc3d4-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c358674b-3741-4d28-a629-029d94cdc3d4" (UID: "c358674b-3741-4d28-a629-029d94cdc3d4"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:25:39 crc kubenswrapper[4824]: I1211 02:25:39.232935 4824 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c358674b-3741-4d28-a629-029d94cdc3d4-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 11 02:25:39 crc kubenswrapper[4824]: I1211 02:25:39.233208 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nmnsp\" (UniqueName: \"kubernetes.io/projected/c358674b-3741-4d28-a629-029d94cdc3d4-kube-api-access-nmnsp\") on node \"crc\" DevicePath \"\"" Dec 11 02:25:39 crc kubenswrapper[4824]: I1211 02:25:39.233281 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c358674b-3741-4d28-a629-029d94cdc3d4-config\") on node \"crc\" DevicePath \"\"" Dec 11 02:25:39 crc kubenswrapper[4824]: I1211 02:25:39.233344 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c358674b-3741-4d28-a629-029d94cdc3d4-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 11 02:25:39 crc kubenswrapper[4824]: I1211 02:25:39.233401 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c358674b-3741-4d28-a629-029d94cdc3d4-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 11 02:25:39 crc kubenswrapper[4824]: I1211 02:25:39.233454 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c358674b-3741-4d28-a629-029d94cdc3d4-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 11 02:25:39 crc kubenswrapper[4824]: I1211 02:25:39.481364 4824 generic.go:334] "Generic (PLEG): container finished" podID="c358674b-3741-4d28-a629-029d94cdc3d4" containerID="bd89a5a921ae19774166b4f494c2bb2e0ce856de55580b2d902ebdb5c24f7d9e" exitCode=0 Dec 11 02:25:39 crc kubenswrapper[4824]: I1211 02:25:39.481430 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-n9sd5" event={"ID":"c358674b-3741-4d28-a629-029d94cdc3d4","Type":"ContainerDied","Data":"bd89a5a921ae19774166b4f494c2bb2e0ce856de55580b2d902ebdb5c24f7d9e"} Dec 11 02:25:39 crc kubenswrapper[4824]: I1211 02:25:39.481486 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-n9sd5" event={"ID":"c358674b-3741-4d28-a629-029d94cdc3d4","Type":"ContainerDied","Data":"353d3c2d26c496b382830a8aeaa639e60dcca462352b24b7bf941ab2a4d0d71e"} Dec 11 02:25:39 crc kubenswrapper[4824]: I1211 02:25:39.481521 4824 scope.go:117] "RemoveContainer" containerID="bd89a5a921ae19774166b4f494c2bb2e0ce856de55580b2d902ebdb5c24f7d9e" Dec 11 02:25:39 crc kubenswrapper[4824]: I1211 02:25:39.481788 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-n9sd5" Dec 11 02:25:39 crc kubenswrapper[4824]: I1211 02:25:39.520248 4824 scope.go:117] "RemoveContainer" containerID="67044c3ed0dfd1e496a077f8d2c586480c8b3e34954ce85885b81960285fd53d" Dec 11 02:25:39 crc kubenswrapper[4824]: I1211 02:25:39.527387 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-n9sd5"] Dec 11 02:25:39 crc kubenswrapper[4824]: I1211 02:25:39.536949 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-n9sd5"] Dec 11 02:25:39 crc kubenswrapper[4824]: I1211 02:25:39.545522 4824 scope.go:117] "RemoveContainer" containerID="bd89a5a921ae19774166b4f494c2bb2e0ce856de55580b2d902ebdb5c24f7d9e" Dec 11 02:25:39 crc kubenswrapper[4824]: E1211 02:25:39.547442 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd89a5a921ae19774166b4f494c2bb2e0ce856de55580b2d902ebdb5c24f7d9e\": container with ID starting with bd89a5a921ae19774166b4f494c2bb2e0ce856de55580b2d902ebdb5c24f7d9e not found: ID does not exist" containerID="bd89a5a921ae19774166b4f494c2bb2e0ce856de55580b2d902ebdb5c24f7d9e" Dec 11 02:25:39 crc kubenswrapper[4824]: I1211 02:25:39.547475 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd89a5a921ae19774166b4f494c2bb2e0ce856de55580b2d902ebdb5c24f7d9e"} err="failed to get container status \"bd89a5a921ae19774166b4f494c2bb2e0ce856de55580b2d902ebdb5c24f7d9e\": rpc error: code = NotFound desc = could not find container \"bd89a5a921ae19774166b4f494c2bb2e0ce856de55580b2d902ebdb5c24f7d9e\": container with ID starting with bd89a5a921ae19774166b4f494c2bb2e0ce856de55580b2d902ebdb5c24f7d9e not found: ID does not exist" Dec 11 02:25:39 crc kubenswrapper[4824]: I1211 02:25:39.547564 4824 scope.go:117] "RemoveContainer" containerID="67044c3ed0dfd1e496a077f8d2c586480c8b3e34954ce85885b81960285fd53d" Dec 11 02:25:39 crc kubenswrapper[4824]: E1211 02:25:39.547919 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"67044c3ed0dfd1e496a077f8d2c586480c8b3e34954ce85885b81960285fd53d\": container with ID starting with 67044c3ed0dfd1e496a077f8d2c586480c8b3e34954ce85885b81960285fd53d not found: ID does not exist" containerID="67044c3ed0dfd1e496a077f8d2c586480c8b3e34954ce85885b81960285fd53d" Dec 11 02:25:39 crc kubenswrapper[4824]: I1211 02:25:39.547936 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"67044c3ed0dfd1e496a077f8d2c586480c8b3e34954ce85885b81960285fd53d"} err="failed to get container status \"67044c3ed0dfd1e496a077f8d2c586480c8b3e34954ce85885b81960285fd53d\": rpc error: code = NotFound desc = could not find container \"67044c3ed0dfd1e496a077f8d2c586480c8b3e34954ce85885b81960285fd53d\": container with ID starting with 67044c3ed0dfd1e496a077f8d2c586480c8b3e34954ce85885b81960285fd53d not found: ID does not exist" Dec 11 02:25:39 crc kubenswrapper[4824]: I1211 02:25:39.576160 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55478c4467-mptxc"] Dec 11 02:25:40 crc kubenswrapper[4824]: I1211 02:25:40.496418 4824 generic.go:334] "Generic (PLEG): container finished" podID="71c2c849-fc4f-44ac-92d4-75848ea56874" containerID="6dc24b011e8efb4bafcc8e94b8869176f88a544fab5b403c3ec31f9072c17ed2" exitCode=0 Dec 11 02:25:40 crc kubenswrapper[4824]: I1211 02:25:40.496506 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55478c4467-mptxc" event={"ID":"71c2c849-fc4f-44ac-92d4-75848ea56874","Type":"ContainerDied","Data":"6dc24b011e8efb4bafcc8e94b8869176f88a544fab5b403c3ec31f9072c17ed2"} Dec 11 02:25:40 crc kubenswrapper[4824]: I1211 02:25:40.496775 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55478c4467-mptxc" event={"ID":"71c2c849-fc4f-44ac-92d4-75848ea56874","Type":"ContainerStarted","Data":"f73564600366602fe61c0844ac722e68ed1a7d0dce87fd6960e2d59ab6d1f247"} Dec 11 02:25:40 crc kubenswrapper[4824]: I1211 02:25:40.647592 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c358674b-3741-4d28-a629-029d94cdc3d4" path="/var/lib/kubelet/pods/c358674b-3741-4d28-a629-029d94cdc3d4/volumes" Dec 11 02:25:41 crc kubenswrapper[4824]: I1211 02:25:41.512467 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55478c4467-mptxc" event={"ID":"71c2c849-fc4f-44ac-92d4-75848ea56874","Type":"ContainerStarted","Data":"17c843c56aaa6913fa6dcfe830cae6a3c4d34df467dd0694403142f77639ce8a"} Dec 11 02:25:41 crc kubenswrapper[4824]: I1211 02:25:41.512871 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-55478c4467-mptxc" Dec 11 02:25:41 crc kubenswrapper[4824]: I1211 02:25:41.549935 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-55478c4467-mptxc" podStartSLOduration=3.549905605 podStartE2EDuration="3.549905605s" podCreationTimestamp="2025-12-11 02:25:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:25:41.535702077 +0000 UTC m=+1483.224739496" watchObservedRunningTime="2025-12-11 02:25:41.549905605 +0000 UTC m=+1483.238943014" Dec 11 02:25:43 crc kubenswrapper[4824]: I1211 02:25:43.891778 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-89c5cd4d5-n9sd5" podUID="c358674b-3741-4d28-a629-029d94cdc3d4" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.201:5353: i/o timeout" Dec 11 02:25:49 crc kubenswrapper[4824]: I1211 02:25:49.083388 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-55478c4467-mptxc" Dec 11 02:25:49 crc kubenswrapper[4824]: I1211 02:25:49.205943 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-s8lft"] Dec 11 02:25:49 crc kubenswrapper[4824]: I1211 02:25:49.206674 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-79bd4cc8c9-s8lft" podUID="74b07b56-f514-4afe-9715-9f42b144f066" containerName="dnsmasq-dns" containerID="cri-o://3a4eac77f7a127683859cee1c67d38bf38413baa0c24d41f15af3eb0b9adaa95" gracePeriod=10 Dec 11 02:25:49 crc kubenswrapper[4824]: I1211 02:25:49.658970 4824 generic.go:334] "Generic (PLEG): container finished" podID="74b07b56-f514-4afe-9715-9f42b144f066" containerID="3a4eac77f7a127683859cee1c67d38bf38413baa0c24d41f15af3eb0b9adaa95" exitCode=0 Dec 11 02:25:49 crc kubenswrapper[4824]: I1211 02:25:49.659012 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-s8lft" event={"ID":"74b07b56-f514-4afe-9715-9f42b144f066","Type":"ContainerDied","Data":"3a4eac77f7a127683859cee1c67d38bf38413baa0c24d41f15af3eb0b9adaa95"} Dec 11 02:25:49 crc kubenswrapper[4824]: I1211 02:25:49.659037 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-s8lft" event={"ID":"74b07b56-f514-4afe-9715-9f42b144f066","Type":"ContainerDied","Data":"383d0eefac77c72252ab4edc2dc9f634ae1298032c1b6c9284614fb4384ece73"} Dec 11 02:25:49 crc kubenswrapper[4824]: I1211 02:25:49.659047 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="383d0eefac77c72252ab4edc2dc9f634ae1298032c1b6c9284614fb4384ece73" Dec 11 02:25:49 crc kubenswrapper[4824]: I1211 02:25:49.730845 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-s8lft" Dec 11 02:25:49 crc kubenswrapper[4824]: I1211 02:25:49.887852 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/74b07b56-f514-4afe-9715-9f42b144f066-dns-svc\") pod \"74b07b56-f514-4afe-9715-9f42b144f066\" (UID: \"74b07b56-f514-4afe-9715-9f42b144f066\") " Dec 11 02:25:49 crc kubenswrapper[4824]: I1211 02:25:49.887913 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/74b07b56-f514-4afe-9715-9f42b144f066-openstack-edpm-ipam\") pod \"74b07b56-f514-4afe-9715-9f42b144f066\" (UID: \"74b07b56-f514-4afe-9715-9f42b144f066\") " Dec 11 02:25:49 crc kubenswrapper[4824]: I1211 02:25:49.887957 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/74b07b56-f514-4afe-9715-9f42b144f066-ovsdbserver-nb\") pod \"74b07b56-f514-4afe-9715-9f42b144f066\" (UID: \"74b07b56-f514-4afe-9715-9f42b144f066\") " Dec 11 02:25:49 crc kubenswrapper[4824]: I1211 02:25:49.887983 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/74b07b56-f514-4afe-9715-9f42b144f066-dns-swift-storage-0\") pod \"74b07b56-f514-4afe-9715-9f42b144f066\" (UID: \"74b07b56-f514-4afe-9715-9f42b144f066\") " Dec 11 02:25:49 crc kubenswrapper[4824]: I1211 02:25:49.888067 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h5wfl\" (UniqueName: \"kubernetes.io/projected/74b07b56-f514-4afe-9715-9f42b144f066-kube-api-access-h5wfl\") pod \"74b07b56-f514-4afe-9715-9f42b144f066\" (UID: \"74b07b56-f514-4afe-9715-9f42b144f066\") " Dec 11 02:25:49 crc kubenswrapper[4824]: I1211 02:25:49.888106 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/74b07b56-f514-4afe-9715-9f42b144f066-ovsdbserver-sb\") pod \"74b07b56-f514-4afe-9715-9f42b144f066\" (UID: \"74b07b56-f514-4afe-9715-9f42b144f066\") " Dec 11 02:25:49 crc kubenswrapper[4824]: I1211 02:25:49.888145 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74b07b56-f514-4afe-9715-9f42b144f066-config\") pod \"74b07b56-f514-4afe-9715-9f42b144f066\" (UID: \"74b07b56-f514-4afe-9715-9f42b144f066\") " Dec 11 02:25:49 crc kubenswrapper[4824]: I1211 02:25:49.899425 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74b07b56-f514-4afe-9715-9f42b144f066-kube-api-access-h5wfl" (OuterVolumeSpecName: "kube-api-access-h5wfl") pod "74b07b56-f514-4afe-9715-9f42b144f066" (UID: "74b07b56-f514-4afe-9715-9f42b144f066"). InnerVolumeSpecName "kube-api-access-h5wfl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:25:49 crc kubenswrapper[4824]: I1211 02:25:49.934636 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/74b07b56-f514-4afe-9715-9f42b144f066-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "74b07b56-f514-4afe-9715-9f42b144f066" (UID: "74b07b56-f514-4afe-9715-9f42b144f066"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:25:49 crc kubenswrapper[4824]: I1211 02:25:49.936539 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/74b07b56-f514-4afe-9715-9f42b144f066-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "74b07b56-f514-4afe-9715-9f42b144f066" (UID: "74b07b56-f514-4afe-9715-9f42b144f066"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:25:49 crc kubenswrapper[4824]: I1211 02:25:49.956702 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/74b07b56-f514-4afe-9715-9f42b144f066-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "74b07b56-f514-4afe-9715-9f42b144f066" (UID: "74b07b56-f514-4afe-9715-9f42b144f066"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:25:49 crc kubenswrapper[4824]: I1211 02:25:49.965295 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/74b07b56-f514-4afe-9715-9f42b144f066-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "74b07b56-f514-4afe-9715-9f42b144f066" (UID: "74b07b56-f514-4afe-9715-9f42b144f066"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:25:49 crc kubenswrapper[4824]: I1211 02:25:49.966850 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/74b07b56-f514-4afe-9715-9f42b144f066-config" (OuterVolumeSpecName: "config") pod "74b07b56-f514-4afe-9715-9f42b144f066" (UID: "74b07b56-f514-4afe-9715-9f42b144f066"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:25:49 crc kubenswrapper[4824]: I1211 02:25:49.975285 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/74b07b56-f514-4afe-9715-9f42b144f066-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "74b07b56-f514-4afe-9715-9f42b144f066" (UID: "74b07b56-f514-4afe-9715-9f42b144f066"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:25:49 crc kubenswrapper[4824]: I1211 02:25:49.990938 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/74b07b56-f514-4afe-9715-9f42b144f066-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 11 02:25:49 crc kubenswrapper[4824]: I1211 02:25:49.990977 4824 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/74b07b56-f514-4afe-9715-9f42b144f066-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 11 02:25:49 crc kubenswrapper[4824]: I1211 02:25:49.990994 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/74b07b56-f514-4afe-9715-9f42b144f066-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 11 02:25:49 crc kubenswrapper[4824]: I1211 02:25:49.991008 4824 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/74b07b56-f514-4afe-9715-9f42b144f066-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 11 02:25:49 crc kubenswrapper[4824]: I1211 02:25:49.991020 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h5wfl\" (UniqueName: \"kubernetes.io/projected/74b07b56-f514-4afe-9715-9f42b144f066-kube-api-access-h5wfl\") on node \"crc\" DevicePath \"\"" Dec 11 02:25:49 crc kubenswrapper[4824]: I1211 02:25:49.991031 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/74b07b56-f514-4afe-9715-9f42b144f066-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 11 02:25:49 crc kubenswrapper[4824]: I1211 02:25:49.991043 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74b07b56-f514-4afe-9715-9f42b144f066-config\") on node \"crc\" DevicePath \"\"" Dec 11 02:25:50 crc kubenswrapper[4824]: I1211 02:25:50.671549 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-s8lft" Dec 11 02:25:50 crc kubenswrapper[4824]: I1211 02:25:50.698497 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-s8lft"] Dec 11 02:25:50 crc kubenswrapper[4824]: I1211 02:25:50.708724 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-s8lft"] Dec 11 02:25:52 crc kubenswrapper[4824]: I1211 02:25:52.642938 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74b07b56-f514-4afe-9715-9f42b144f066" path="/var/lib/kubelet/pods/74b07b56-f514-4afe-9715-9f42b144f066/volumes" Dec 11 02:25:59 crc kubenswrapper[4824]: I1211 02:25:59.801286 4824 generic.go:334] "Generic (PLEG): container finished" podID="89eddaec-a9d6-4b73-b654-2f176d45d162" containerID="5ec7fbef36d90f11e756f4556f943f2f69d171390446f82e5a27da5b1f3e4952" exitCode=0 Dec 11 02:25:59 crc kubenswrapper[4824]: I1211 02:25:59.801351 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"89eddaec-a9d6-4b73-b654-2f176d45d162","Type":"ContainerDied","Data":"5ec7fbef36d90f11e756f4556f943f2f69d171390446f82e5a27da5b1f3e4952"} Dec 11 02:26:00 crc kubenswrapper[4824]: I1211 02:26:00.814896 4824 generic.go:334] "Generic (PLEG): container finished" podID="8799d853-613d-4376-a372-7dedd968f264" containerID="ed43359051393901062422faf9762d9d95dde215edf8efe06d58883210e99376" exitCode=0 Dec 11 02:26:00 crc kubenswrapper[4824]: I1211 02:26:00.815673 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"8799d853-613d-4376-a372-7dedd968f264","Type":"ContainerDied","Data":"ed43359051393901062422faf9762d9d95dde215edf8efe06d58883210e99376"} Dec 11 02:26:00 crc kubenswrapper[4824]: I1211 02:26:00.821726 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"89eddaec-a9d6-4b73-b654-2f176d45d162","Type":"ContainerStarted","Data":"0c9120507c9d9d0975531c3712963dbb7a37ceb06bffd5a5f2cb491756a8a371"} Dec 11 02:26:00 crc kubenswrapper[4824]: I1211 02:26:00.821985 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 11 02:26:00 crc kubenswrapper[4824]: I1211 02:26:00.900581 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=36.900558012 podStartE2EDuration="36.900558012s" podCreationTimestamp="2025-12-11 02:25:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:26:00.891252187 +0000 UTC m=+1502.580289606" watchObservedRunningTime="2025-12-11 02:26:00.900558012 +0000 UTC m=+1502.589595401" Dec 11 02:26:01 crc kubenswrapper[4824]: I1211 02:26:01.838626 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"8799d853-613d-4376-a372-7dedd968f264","Type":"ContainerStarted","Data":"0831c3f0d4ffde893027137b2f6660f7904cb1454f59940c1e2203b3400facbe"} Dec 11 02:26:01 crc kubenswrapper[4824]: I1211 02:26:01.839356 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 11 02:26:01 crc kubenswrapper[4824]: I1211 02:26:01.877976 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.877959454 podStartE2EDuration="37.877959454s" podCreationTimestamp="2025-12-11 02:25:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 02:26:01.864749854 +0000 UTC m=+1503.553787233" watchObservedRunningTime="2025-12-11 02:26:01.877959454 +0000 UTC m=+1503.566996833" Dec 11 02:26:02 crc kubenswrapper[4824]: I1211 02:26:01.999413 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nl76s"] Dec 11 02:26:02 crc kubenswrapper[4824]: E1211 02:26:01.999823 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c358674b-3741-4d28-a629-029d94cdc3d4" containerName="init" Dec 11 02:26:02 crc kubenswrapper[4824]: I1211 02:26:01.999836 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="c358674b-3741-4d28-a629-029d94cdc3d4" containerName="init" Dec 11 02:26:02 crc kubenswrapper[4824]: E1211 02:26:01.999851 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74b07b56-f514-4afe-9715-9f42b144f066" containerName="init" Dec 11 02:26:02 crc kubenswrapper[4824]: I1211 02:26:01.999859 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="74b07b56-f514-4afe-9715-9f42b144f066" containerName="init" Dec 11 02:26:02 crc kubenswrapper[4824]: E1211 02:26:01.999869 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c358674b-3741-4d28-a629-029d94cdc3d4" containerName="dnsmasq-dns" Dec 11 02:26:02 crc kubenswrapper[4824]: I1211 02:26:01.999875 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="c358674b-3741-4d28-a629-029d94cdc3d4" containerName="dnsmasq-dns" Dec 11 02:26:02 crc kubenswrapper[4824]: E1211 02:26:01.999887 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74b07b56-f514-4afe-9715-9f42b144f066" containerName="dnsmasq-dns" Dec 11 02:26:02 crc kubenswrapper[4824]: I1211 02:26:01.999893 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="74b07b56-f514-4afe-9715-9f42b144f066" containerName="dnsmasq-dns" Dec 11 02:26:02 crc kubenswrapper[4824]: I1211 02:26:02.000084 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="c358674b-3741-4d28-a629-029d94cdc3d4" containerName="dnsmasq-dns" Dec 11 02:26:02 crc kubenswrapper[4824]: I1211 02:26:02.000100 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="74b07b56-f514-4afe-9715-9f42b144f066" containerName="dnsmasq-dns" Dec 11 02:26:02 crc kubenswrapper[4824]: I1211 02:26:02.000701 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nl76s" Dec 11 02:26:02 crc kubenswrapper[4824]: I1211 02:26:02.006329 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 11 02:26:02 crc kubenswrapper[4824]: I1211 02:26:02.006726 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 11 02:26:02 crc kubenswrapper[4824]: I1211 02:26:02.006921 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 11 02:26:02 crc kubenswrapper[4824]: I1211 02:26:02.007961 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-hs2fv" Dec 11 02:26:02 crc kubenswrapper[4824]: I1211 02:26:02.036022 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nl76s"] Dec 11 02:26:02 crc kubenswrapper[4824]: I1211 02:26:02.150666 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sms25\" (UniqueName: \"kubernetes.io/projected/09a89381-9670-4aff-965b-bd6d0d3e5ae7-kube-api-access-sms25\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-nl76s\" (UID: \"09a89381-9670-4aff-965b-bd6d0d3e5ae7\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nl76s" Dec 11 02:26:02 crc kubenswrapper[4824]: I1211 02:26:02.150716 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/09a89381-9670-4aff-965b-bd6d0d3e5ae7-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-nl76s\" (UID: \"09a89381-9670-4aff-965b-bd6d0d3e5ae7\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nl76s" Dec 11 02:26:02 crc kubenswrapper[4824]: I1211 02:26:02.150772 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09a89381-9670-4aff-965b-bd6d0d3e5ae7-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-nl76s\" (UID: \"09a89381-9670-4aff-965b-bd6d0d3e5ae7\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nl76s" Dec 11 02:26:02 crc kubenswrapper[4824]: I1211 02:26:02.150854 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/09a89381-9670-4aff-965b-bd6d0d3e5ae7-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-nl76s\" (UID: \"09a89381-9670-4aff-965b-bd6d0d3e5ae7\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nl76s" Dec 11 02:26:02 crc kubenswrapper[4824]: I1211 02:26:02.252838 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/09a89381-9670-4aff-965b-bd6d0d3e5ae7-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-nl76s\" (UID: \"09a89381-9670-4aff-965b-bd6d0d3e5ae7\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nl76s" Dec 11 02:26:02 crc kubenswrapper[4824]: I1211 02:26:02.253355 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sms25\" (UniqueName: \"kubernetes.io/projected/09a89381-9670-4aff-965b-bd6d0d3e5ae7-kube-api-access-sms25\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-nl76s\" (UID: \"09a89381-9670-4aff-965b-bd6d0d3e5ae7\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nl76s" Dec 11 02:26:02 crc kubenswrapper[4824]: I1211 02:26:02.253560 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/09a89381-9670-4aff-965b-bd6d0d3e5ae7-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-nl76s\" (UID: \"09a89381-9670-4aff-965b-bd6d0d3e5ae7\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nl76s" Dec 11 02:26:02 crc kubenswrapper[4824]: I1211 02:26:02.253772 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09a89381-9670-4aff-965b-bd6d0d3e5ae7-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-nl76s\" (UID: \"09a89381-9670-4aff-965b-bd6d0d3e5ae7\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nl76s" Dec 11 02:26:02 crc kubenswrapper[4824]: I1211 02:26:02.262039 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/09a89381-9670-4aff-965b-bd6d0d3e5ae7-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-nl76s\" (UID: \"09a89381-9670-4aff-965b-bd6d0d3e5ae7\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nl76s" Dec 11 02:26:02 crc kubenswrapper[4824]: I1211 02:26:02.270619 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09a89381-9670-4aff-965b-bd6d0d3e5ae7-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-nl76s\" (UID: \"09a89381-9670-4aff-965b-bd6d0d3e5ae7\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nl76s" Dec 11 02:26:02 crc kubenswrapper[4824]: I1211 02:26:02.271779 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sms25\" (UniqueName: \"kubernetes.io/projected/09a89381-9670-4aff-965b-bd6d0d3e5ae7-kube-api-access-sms25\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-nl76s\" (UID: \"09a89381-9670-4aff-965b-bd6d0d3e5ae7\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nl76s" Dec 11 02:26:02 crc kubenswrapper[4824]: I1211 02:26:02.279995 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/09a89381-9670-4aff-965b-bd6d0d3e5ae7-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-nl76s\" (UID: \"09a89381-9670-4aff-965b-bd6d0d3e5ae7\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nl76s" Dec 11 02:26:02 crc kubenswrapper[4824]: I1211 02:26:02.323766 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nl76s" Dec 11 02:26:02 crc kubenswrapper[4824]: I1211 02:26:02.942026 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nl76s"] Dec 11 02:26:03 crc kubenswrapper[4824]: I1211 02:26:03.855038 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nl76s" event={"ID":"09a89381-9670-4aff-965b-bd6d0d3e5ae7","Type":"ContainerStarted","Data":"f7de801e1fb814245b291ffd1ad9feb08c18c9125e2de786b31035106f144a6f"} Dec 11 02:26:07 crc kubenswrapper[4824]: I1211 02:26:07.250886 4824 patch_prober.go:28] interesting pod/machine-config-daemon-gx6xt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 02:26:07 crc kubenswrapper[4824]: I1211 02:26:07.251494 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 02:26:14 crc kubenswrapper[4824]: I1211 02:26:14.890281 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 11 02:26:14 crc kubenswrapper[4824]: I1211 02:26:14.992173 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 11 02:26:21 crc kubenswrapper[4824]: I1211 02:26:21.035067 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nl76s" event={"ID":"09a89381-9670-4aff-965b-bd6d0d3e5ae7","Type":"ContainerStarted","Data":"9fde52ad2a5e578c84367ac0cc3e05de161a7bcfa62fede73c29eef3f12f5617"} Dec 11 02:26:21 crc kubenswrapper[4824]: I1211 02:26:21.055153 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nl76s" podStartSLOduration=3.13248465 podStartE2EDuration="20.055133573s" podCreationTimestamp="2025-12-11 02:26:01 +0000 UTC" firstStartedPulling="2025-12-11 02:26:02.935207364 +0000 UTC m=+1504.624244743" lastFinishedPulling="2025-12-11 02:26:19.857856277 +0000 UTC m=+1521.546893666" observedRunningTime="2025-12-11 02:26:21.052175752 +0000 UTC m=+1522.741213171" watchObservedRunningTime="2025-12-11 02:26:21.055133573 +0000 UTC m=+1522.744170962" Dec 11 02:26:31 crc kubenswrapper[4824]: I1211 02:26:31.149939 4824 generic.go:334] "Generic (PLEG): container finished" podID="09a89381-9670-4aff-965b-bd6d0d3e5ae7" containerID="9fde52ad2a5e578c84367ac0cc3e05de161a7bcfa62fede73c29eef3f12f5617" exitCode=0 Dec 11 02:26:31 crc kubenswrapper[4824]: I1211 02:26:31.150030 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nl76s" event={"ID":"09a89381-9670-4aff-965b-bd6d0d3e5ae7","Type":"ContainerDied","Data":"9fde52ad2a5e578c84367ac0cc3e05de161a7bcfa62fede73c29eef3f12f5617"} Dec 11 02:26:32 crc kubenswrapper[4824]: I1211 02:26:32.712894 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nl76s" Dec 11 02:26:32 crc kubenswrapper[4824]: I1211 02:26:32.729531 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/09a89381-9670-4aff-965b-bd6d0d3e5ae7-ssh-key\") pod \"09a89381-9670-4aff-965b-bd6d0d3e5ae7\" (UID: \"09a89381-9670-4aff-965b-bd6d0d3e5ae7\") " Dec 11 02:26:32 crc kubenswrapper[4824]: I1211 02:26:32.729594 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sms25\" (UniqueName: \"kubernetes.io/projected/09a89381-9670-4aff-965b-bd6d0d3e5ae7-kube-api-access-sms25\") pod \"09a89381-9670-4aff-965b-bd6d0d3e5ae7\" (UID: \"09a89381-9670-4aff-965b-bd6d0d3e5ae7\") " Dec 11 02:26:32 crc kubenswrapper[4824]: I1211 02:26:32.729722 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09a89381-9670-4aff-965b-bd6d0d3e5ae7-repo-setup-combined-ca-bundle\") pod \"09a89381-9670-4aff-965b-bd6d0d3e5ae7\" (UID: \"09a89381-9670-4aff-965b-bd6d0d3e5ae7\") " Dec 11 02:26:32 crc kubenswrapper[4824]: I1211 02:26:32.729743 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/09a89381-9670-4aff-965b-bd6d0d3e5ae7-inventory\") pod \"09a89381-9670-4aff-965b-bd6d0d3e5ae7\" (UID: \"09a89381-9670-4aff-965b-bd6d0d3e5ae7\") " Dec 11 02:26:32 crc kubenswrapper[4824]: I1211 02:26:32.741794 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09a89381-9670-4aff-965b-bd6d0d3e5ae7-kube-api-access-sms25" (OuterVolumeSpecName: "kube-api-access-sms25") pod "09a89381-9670-4aff-965b-bd6d0d3e5ae7" (UID: "09a89381-9670-4aff-965b-bd6d0d3e5ae7"). InnerVolumeSpecName "kube-api-access-sms25". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:26:32 crc kubenswrapper[4824]: I1211 02:26:32.751904 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09a89381-9670-4aff-965b-bd6d0d3e5ae7-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "09a89381-9670-4aff-965b-bd6d0d3e5ae7" (UID: "09a89381-9670-4aff-965b-bd6d0d3e5ae7"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:26:32 crc kubenswrapper[4824]: I1211 02:26:32.781790 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09a89381-9670-4aff-965b-bd6d0d3e5ae7-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "09a89381-9670-4aff-965b-bd6d0d3e5ae7" (UID: "09a89381-9670-4aff-965b-bd6d0d3e5ae7"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:26:32 crc kubenswrapper[4824]: I1211 02:26:32.781836 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09a89381-9670-4aff-965b-bd6d0d3e5ae7-inventory" (OuterVolumeSpecName: "inventory") pod "09a89381-9670-4aff-965b-bd6d0d3e5ae7" (UID: "09a89381-9670-4aff-965b-bd6d0d3e5ae7"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:26:32 crc kubenswrapper[4824]: I1211 02:26:32.832018 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/09a89381-9670-4aff-965b-bd6d0d3e5ae7-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 11 02:26:32 crc kubenswrapper[4824]: I1211 02:26:32.832060 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sms25\" (UniqueName: \"kubernetes.io/projected/09a89381-9670-4aff-965b-bd6d0d3e5ae7-kube-api-access-sms25\") on node \"crc\" DevicePath \"\"" Dec 11 02:26:32 crc kubenswrapper[4824]: I1211 02:26:32.832079 4824 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09a89381-9670-4aff-965b-bd6d0d3e5ae7-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 02:26:32 crc kubenswrapper[4824]: I1211 02:26:32.832096 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/09a89381-9670-4aff-965b-bd6d0d3e5ae7-inventory\") on node \"crc\" DevicePath \"\"" Dec 11 02:26:33 crc kubenswrapper[4824]: I1211 02:26:33.182596 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nl76s" event={"ID":"09a89381-9670-4aff-965b-bd6d0d3e5ae7","Type":"ContainerDied","Data":"f7de801e1fb814245b291ffd1ad9feb08c18c9125e2de786b31035106f144a6f"} Dec 11 02:26:33 crc kubenswrapper[4824]: I1211 02:26:33.182661 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f7de801e1fb814245b291ffd1ad9feb08c18c9125e2de786b31035106f144a6f" Dec 11 02:26:33 crc kubenswrapper[4824]: I1211 02:26:33.182745 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nl76s" Dec 11 02:26:33 crc kubenswrapper[4824]: I1211 02:26:33.320772 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-4292s"] Dec 11 02:26:33 crc kubenswrapper[4824]: E1211 02:26:33.321263 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09a89381-9670-4aff-965b-bd6d0d3e5ae7" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 11 02:26:33 crc kubenswrapper[4824]: I1211 02:26:33.321285 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="09a89381-9670-4aff-965b-bd6d0d3e5ae7" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 11 02:26:33 crc kubenswrapper[4824]: I1211 02:26:33.321533 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="09a89381-9670-4aff-965b-bd6d0d3e5ae7" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 11 02:26:33 crc kubenswrapper[4824]: I1211 02:26:33.322318 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-4292s" Dec 11 02:26:33 crc kubenswrapper[4824]: I1211 02:26:33.329967 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 11 02:26:33 crc kubenswrapper[4824]: I1211 02:26:33.330016 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 11 02:26:33 crc kubenswrapper[4824]: I1211 02:26:33.330066 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 11 02:26:33 crc kubenswrapper[4824]: I1211 02:26:33.330128 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-hs2fv" Dec 11 02:26:33 crc kubenswrapper[4824]: I1211 02:26:33.338972 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-4292s"] Dec 11 02:26:33 crc kubenswrapper[4824]: I1211 02:26:33.342466 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/17d30a7d-e050-463d-ba60-bc8ae69cb21e-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-4292s\" (UID: \"17d30a7d-e050-463d-ba60-bc8ae69cb21e\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-4292s" Dec 11 02:26:33 crc kubenswrapper[4824]: I1211 02:26:33.342522 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/17d30a7d-e050-463d-ba60-bc8ae69cb21e-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-4292s\" (UID: \"17d30a7d-e050-463d-ba60-bc8ae69cb21e\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-4292s" Dec 11 02:26:33 crc kubenswrapper[4824]: I1211 02:26:33.342607 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5zgv2\" (UniqueName: \"kubernetes.io/projected/17d30a7d-e050-463d-ba60-bc8ae69cb21e-kube-api-access-5zgv2\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-4292s\" (UID: \"17d30a7d-e050-463d-ba60-bc8ae69cb21e\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-4292s" Dec 11 02:26:33 crc kubenswrapper[4824]: I1211 02:26:33.444503 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5zgv2\" (UniqueName: \"kubernetes.io/projected/17d30a7d-e050-463d-ba60-bc8ae69cb21e-kube-api-access-5zgv2\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-4292s\" (UID: \"17d30a7d-e050-463d-ba60-bc8ae69cb21e\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-4292s" Dec 11 02:26:33 crc kubenswrapper[4824]: I1211 02:26:33.444660 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/17d30a7d-e050-463d-ba60-bc8ae69cb21e-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-4292s\" (UID: \"17d30a7d-e050-463d-ba60-bc8ae69cb21e\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-4292s" Dec 11 02:26:33 crc kubenswrapper[4824]: I1211 02:26:33.444702 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/17d30a7d-e050-463d-ba60-bc8ae69cb21e-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-4292s\" (UID: \"17d30a7d-e050-463d-ba60-bc8ae69cb21e\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-4292s" Dec 11 02:26:33 crc kubenswrapper[4824]: I1211 02:26:33.453085 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/17d30a7d-e050-463d-ba60-bc8ae69cb21e-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-4292s\" (UID: \"17d30a7d-e050-463d-ba60-bc8ae69cb21e\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-4292s" Dec 11 02:26:33 crc kubenswrapper[4824]: I1211 02:26:33.453811 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/17d30a7d-e050-463d-ba60-bc8ae69cb21e-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-4292s\" (UID: \"17d30a7d-e050-463d-ba60-bc8ae69cb21e\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-4292s" Dec 11 02:26:33 crc kubenswrapper[4824]: I1211 02:26:33.484256 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5zgv2\" (UniqueName: \"kubernetes.io/projected/17d30a7d-e050-463d-ba60-bc8ae69cb21e-kube-api-access-5zgv2\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-4292s\" (UID: \"17d30a7d-e050-463d-ba60-bc8ae69cb21e\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-4292s" Dec 11 02:26:33 crc kubenswrapper[4824]: I1211 02:26:33.640195 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-4292s" Dec 11 02:26:34 crc kubenswrapper[4824]: I1211 02:26:34.297665 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-4292s"] Dec 11 02:26:35 crc kubenswrapper[4824]: I1211 02:26:35.219291 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-4292s" event={"ID":"17d30a7d-e050-463d-ba60-bc8ae69cb21e","Type":"ContainerStarted","Data":"bf1c8fc8209da2f0967ddcd4d5bad13b1eac00d7e33afcb2fd3cb31d0abe7752"} Dec 11 02:26:35 crc kubenswrapper[4824]: I1211 02:26:35.219670 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-4292s" event={"ID":"17d30a7d-e050-463d-ba60-bc8ae69cb21e","Type":"ContainerStarted","Data":"e1c5731cfdb58078fecc881fda57eb100bafd42a82d29260cbcdd64da24b2fd5"} Dec 11 02:26:35 crc kubenswrapper[4824]: I1211 02:26:35.250535 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-4292s" podStartSLOduration=2.078172654 podStartE2EDuration="2.250510342s" podCreationTimestamp="2025-12-11 02:26:33 +0000 UTC" firstStartedPulling="2025-12-11 02:26:34.294280587 +0000 UTC m=+1535.983317976" lastFinishedPulling="2025-12-11 02:26:34.466618255 +0000 UTC m=+1536.155655664" observedRunningTime="2025-12-11 02:26:35.238802483 +0000 UTC m=+1536.927839942" watchObservedRunningTime="2025-12-11 02:26:35.250510342 +0000 UTC m=+1536.939547761" Dec 11 02:26:37 crc kubenswrapper[4824]: I1211 02:26:37.250961 4824 patch_prober.go:28] interesting pod/machine-config-daemon-gx6xt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 02:26:37 crc kubenswrapper[4824]: I1211 02:26:37.251391 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 02:26:37 crc kubenswrapper[4824]: I1211 02:26:37.251457 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" Dec 11 02:26:37 crc kubenswrapper[4824]: I1211 02:26:37.252594 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1d7148bf050b9a450d49eb09a5933cb6d8dd6490c731a4e84f5c37876b3c58a6"} pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 11 02:26:37 crc kubenswrapper[4824]: I1211 02:26:37.252708 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" containerName="machine-config-daemon" containerID="cri-o://1d7148bf050b9a450d49eb09a5933cb6d8dd6490c731a4e84f5c37876b3c58a6" gracePeriod=600 Dec 11 02:26:37 crc kubenswrapper[4824]: E1211 02:26:37.383557 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 02:26:38 crc kubenswrapper[4824]: I1211 02:26:38.260443 4824 generic.go:334] "Generic (PLEG): container finished" podID="17d30a7d-e050-463d-ba60-bc8ae69cb21e" containerID="bf1c8fc8209da2f0967ddcd4d5bad13b1eac00d7e33afcb2fd3cb31d0abe7752" exitCode=0 Dec 11 02:26:38 crc kubenswrapper[4824]: I1211 02:26:38.260572 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-4292s" event={"ID":"17d30a7d-e050-463d-ba60-bc8ae69cb21e","Type":"ContainerDied","Data":"bf1c8fc8209da2f0967ddcd4d5bad13b1eac00d7e33afcb2fd3cb31d0abe7752"} Dec 11 02:26:38 crc kubenswrapper[4824]: I1211 02:26:38.264679 4824 generic.go:334] "Generic (PLEG): container finished" podID="44b156e4-64a4-4d45-aa5e-9b10a862faed" containerID="1d7148bf050b9a450d49eb09a5933cb6d8dd6490c731a4e84f5c37876b3c58a6" exitCode=0 Dec 11 02:26:38 crc kubenswrapper[4824]: I1211 02:26:38.264731 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" event={"ID":"44b156e4-64a4-4d45-aa5e-9b10a862faed","Type":"ContainerDied","Data":"1d7148bf050b9a450d49eb09a5933cb6d8dd6490c731a4e84f5c37876b3c58a6"} Dec 11 02:26:38 crc kubenswrapper[4824]: I1211 02:26:38.264809 4824 scope.go:117] "RemoveContainer" containerID="d40ff17e029de5775ac2507f904eb2259da25d15d3aa3e3a0e4c811c4717820d" Dec 11 02:26:38 crc kubenswrapper[4824]: I1211 02:26:38.265575 4824 scope.go:117] "RemoveContainer" containerID="1d7148bf050b9a450d49eb09a5933cb6d8dd6490c731a4e84f5c37876b3c58a6" Dec 11 02:26:38 crc kubenswrapper[4824]: E1211 02:26:38.266070 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 02:26:39 crc kubenswrapper[4824]: I1211 02:26:39.802521 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-4292s" Dec 11 02:26:39 crc kubenswrapper[4824]: I1211 02:26:39.921069 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/17d30a7d-e050-463d-ba60-bc8ae69cb21e-inventory\") pod \"17d30a7d-e050-463d-ba60-bc8ae69cb21e\" (UID: \"17d30a7d-e050-463d-ba60-bc8ae69cb21e\") " Dec 11 02:26:39 crc kubenswrapper[4824]: I1211 02:26:39.921741 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5zgv2\" (UniqueName: \"kubernetes.io/projected/17d30a7d-e050-463d-ba60-bc8ae69cb21e-kube-api-access-5zgv2\") pod \"17d30a7d-e050-463d-ba60-bc8ae69cb21e\" (UID: \"17d30a7d-e050-463d-ba60-bc8ae69cb21e\") " Dec 11 02:26:39 crc kubenswrapper[4824]: I1211 02:26:39.921975 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/17d30a7d-e050-463d-ba60-bc8ae69cb21e-ssh-key\") pod \"17d30a7d-e050-463d-ba60-bc8ae69cb21e\" (UID: \"17d30a7d-e050-463d-ba60-bc8ae69cb21e\") " Dec 11 02:26:39 crc kubenswrapper[4824]: I1211 02:26:39.930043 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17d30a7d-e050-463d-ba60-bc8ae69cb21e-kube-api-access-5zgv2" (OuterVolumeSpecName: "kube-api-access-5zgv2") pod "17d30a7d-e050-463d-ba60-bc8ae69cb21e" (UID: "17d30a7d-e050-463d-ba60-bc8ae69cb21e"). InnerVolumeSpecName "kube-api-access-5zgv2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:26:39 crc kubenswrapper[4824]: I1211 02:26:39.978454 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17d30a7d-e050-463d-ba60-bc8ae69cb21e-inventory" (OuterVolumeSpecName: "inventory") pod "17d30a7d-e050-463d-ba60-bc8ae69cb21e" (UID: "17d30a7d-e050-463d-ba60-bc8ae69cb21e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:26:39 crc kubenswrapper[4824]: I1211 02:26:39.981969 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17d30a7d-e050-463d-ba60-bc8ae69cb21e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "17d30a7d-e050-463d-ba60-bc8ae69cb21e" (UID: "17d30a7d-e050-463d-ba60-bc8ae69cb21e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:26:40 crc kubenswrapper[4824]: I1211 02:26:40.024903 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/17d30a7d-e050-463d-ba60-bc8ae69cb21e-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 11 02:26:40 crc kubenswrapper[4824]: I1211 02:26:40.024954 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/17d30a7d-e050-463d-ba60-bc8ae69cb21e-inventory\") on node \"crc\" DevicePath \"\"" Dec 11 02:26:40 crc kubenswrapper[4824]: I1211 02:26:40.024975 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5zgv2\" (UniqueName: \"kubernetes.io/projected/17d30a7d-e050-463d-ba60-bc8ae69cb21e-kube-api-access-5zgv2\") on node \"crc\" DevicePath \"\"" Dec 11 02:26:40 crc kubenswrapper[4824]: I1211 02:26:40.307873 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-4292s" event={"ID":"17d30a7d-e050-463d-ba60-bc8ae69cb21e","Type":"ContainerDied","Data":"e1c5731cfdb58078fecc881fda57eb100bafd42a82d29260cbcdd64da24b2fd5"} Dec 11 02:26:40 crc kubenswrapper[4824]: I1211 02:26:40.307936 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e1c5731cfdb58078fecc881fda57eb100bafd42a82d29260cbcdd64da24b2fd5" Dec 11 02:26:40 crc kubenswrapper[4824]: I1211 02:26:40.307943 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-4292s" Dec 11 02:26:40 crc kubenswrapper[4824]: I1211 02:26:40.390172 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bsknw"] Dec 11 02:26:40 crc kubenswrapper[4824]: E1211 02:26:40.391092 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17d30a7d-e050-463d-ba60-bc8ae69cb21e" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 11 02:26:40 crc kubenswrapper[4824]: I1211 02:26:40.391287 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="17d30a7d-e050-463d-ba60-bc8ae69cb21e" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 11 02:26:40 crc kubenswrapper[4824]: I1211 02:26:40.391834 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="17d30a7d-e050-463d-ba60-bc8ae69cb21e" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 11 02:26:40 crc kubenswrapper[4824]: I1211 02:26:40.393443 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bsknw" Dec 11 02:26:40 crc kubenswrapper[4824]: I1211 02:26:40.397359 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 11 02:26:40 crc kubenswrapper[4824]: I1211 02:26:40.397793 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 11 02:26:40 crc kubenswrapper[4824]: I1211 02:26:40.397965 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 11 02:26:40 crc kubenswrapper[4824]: I1211 02:26:40.398228 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-hs2fv" Dec 11 02:26:40 crc kubenswrapper[4824]: I1211 02:26:40.408208 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bsknw"] Dec 11 02:26:40 crc kubenswrapper[4824]: I1211 02:26:40.434244 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1bf6cf36-d7eb-42be-aeee-c6824de8b6ba-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bsknw\" (UID: \"1bf6cf36-d7eb-42be-aeee-c6824de8b6ba\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bsknw" Dec 11 02:26:40 crc kubenswrapper[4824]: I1211 02:26:40.434303 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1bf6cf36-d7eb-42be-aeee-c6824de8b6ba-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bsknw\" (UID: \"1bf6cf36-d7eb-42be-aeee-c6824de8b6ba\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bsknw" Dec 11 02:26:40 crc kubenswrapper[4824]: I1211 02:26:40.434480 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bf6cf36-d7eb-42be-aeee-c6824de8b6ba-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bsknw\" (UID: \"1bf6cf36-d7eb-42be-aeee-c6824de8b6ba\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bsknw" Dec 11 02:26:40 crc kubenswrapper[4824]: I1211 02:26:40.434664 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mbzqq\" (UniqueName: \"kubernetes.io/projected/1bf6cf36-d7eb-42be-aeee-c6824de8b6ba-kube-api-access-mbzqq\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bsknw\" (UID: \"1bf6cf36-d7eb-42be-aeee-c6824de8b6ba\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bsknw" Dec 11 02:26:40 crc kubenswrapper[4824]: I1211 02:26:40.536720 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bf6cf36-d7eb-42be-aeee-c6824de8b6ba-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bsknw\" (UID: \"1bf6cf36-d7eb-42be-aeee-c6824de8b6ba\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bsknw" Dec 11 02:26:40 crc kubenswrapper[4824]: I1211 02:26:40.536801 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mbzqq\" (UniqueName: \"kubernetes.io/projected/1bf6cf36-d7eb-42be-aeee-c6824de8b6ba-kube-api-access-mbzqq\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bsknw\" (UID: \"1bf6cf36-d7eb-42be-aeee-c6824de8b6ba\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bsknw" Dec 11 02:26:40 crc kubenswrapper[4824]: I1211 02:26:40.536910 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1bf6cf36-d7eb-42be-aeee-c6824de8b6ba-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bsknw\" (UID: \"1bf6cf36-d7eb-42be-aeee-c6824de8b6ba\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bsknw" Dec 11 02:26:40 crc kubenswrapper[4824]: I1211 02:26:40.536945 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1bf6cf36-d7eb-42be-aeee-c6824de8b6ba-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bsknw\" (UID: \"1bf6cf36-d7eb-42be-aeee-c6824de8b6ba\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bsknw" Dec 11 02:26:40 crc kubenswrapper[4824]: I1211 02:26:40.540985 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1bf6cf36-d7eb-42be-aeee-c6824de8b6ba-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bsknw\" (UID: \"1bf6cf36-d7eb-42be-aeee-c6824de8b6ba\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bsknw" Dec 11 02:26:40 crc kubenswrapper[4824]: I1211 02:26:40.541203 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1bf6cf36-d7eb-42be-aeee-c6824de8b6ba-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bsknw\" (UID: \"1bf6cf36-d7eb-42be-aeee-c6824de8b6ba\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bsknw" Dec 11 02:26:40 crc kubenswrapper[4824]: I1211 02:26:40.541522 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bf6cf36-d7eb-42be-aeee-c6824de8b6ba-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bsknw\" (UID: \"1bf6cf36-d7eb-42be-aeee-c6824de8b6ba\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bsknw" Dec 11 02:26:40 crc kubenswrapper[4824]: I1211 02:26:40.554181 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mbzqq\" (UniqueName: \"kubernetes.io/projected/1bf6cf36-d7eb-42be-aeee-c6824de8b6ba-kube-api-access-mbzqq\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bsknw\" (UID: \"1bf6cf36-d7eb-42be-aeee-c6824de8b6ba\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bsknw" Dec 11 02:26:40 crc kubenswrapper[4824]: I1211 02:26:40.732242 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bsknw" Dec 11 02:26:41 crc kubenswrapper[4824]: I1211 02:26:41.349185 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bsknw"] Dec 11 02:26:42 crc kubenswrapper[4824]: I1211 02:26:42.337721 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bsknw" event={"ID":"1bf6cf36-d7eb-42be-aeee-c6824de8b6ba","Type":"ContainerStarted","Data":"ec51036ea596e760fa5982054877bd1b1ff5c19fcb0273dc04bf784fe3fd55bd"} Dec 11 02:26:42 crc kubenswrapper[4824]: I1211 02:26:42.338066 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bsknw" event={"ID":"1bf6cf36-d7eb-42be-aeee-c6824de8b6ba","Type":"ContainerStarted","Data":"b4d83fe70f0ef64308d132ca05182325d140cbddf622d0ed6381e85016a924b0"} Dec 11 02:26:42 crc kubenswrapper[4824]: I1211 02:26:42.370813 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bsknw" podStartSLOduration=2.17954472 podStartE2EDuration="2.370780724s" podCreationTimestamp="2025-12-11 02:26:40 +0000 UTC" firstStartedPulling="2025-12-11 02:26:41.354453909 +0000 UTC m=+1543.043491308" lastFinishedPulling="2025-12-11 02:26:41.545689893 +0000 UTC m=+1543.234727312" observedRunningTime="2025-12-11 02:26:42.361158482 +0000 UTC m=+1544.050195921" watchObservedRunningTime="2025-12-11 02:26:42.370780724 +0000 UTC m=+1544.059818143" Dec 11 02:26:52 crc kubenswrapper[4824]: I1211 02:26:52.632647 4824 scope.go:117] "RemoveContainer" containerID="1d7148bf050b9a450d49eb09a5933cb6d8dd6490c731a4e84f5c37876b3c58a6" Dec 11 02:26:52 crc kubenswrapper[4824]: E1211 02:26:52.633642 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 02:27:04 crc kubenswrapper[4824]: I1211 02:27:04.633268 4824 scope.go:117] "RemoveContainer" containerID="1d7148bf050b9a450d49eb09a5933cb6d8dd6490c731a4e84f5c37876b3c58a6" Dec 11 02:27:04 crc kubenswrapper[4824]: E1211 02:27:04.634612 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 02:27:18 crc kubenswrapper[4824]: I1211 02:27:18.647946 4824 scope.go:117] "RemoveContainer" containerID="1d7148bf050b9a450d49eb09a5933cb6d8dd6490c731a4e84f5c37876b3c58a6" Dec 11 02:27:18 crc kubenswrapper[4824]: E1211 02:27:18.649417 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 02:27:32 crc kubenswrapper[4824]: I1211 02:27:32.633663 4824 scope.go:117] "RemoveContainer" containerID="1d7148bf050b9a450d49eb09a5933cb6d8dd6490c731a4e84f5c37876b3c58a6" Dec 11 02:27:32 crc kubenswrapper[4824]: E1211 02:27:32.635192 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 02:27:44 crc kubenswrapper[4824]: I1211 02:27:44.632877 4824 scope.go:117] "RemoveContainer" containerID="1d7148bf050b9a450d49eb09a5933cb6d8dd6490c731a4e84f5c37876b3c58a6" Dec 11 02:27:44 crc kubenswrapper[4824]: E1211 02:27:44.633760 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 02:27:57 crc kubenswrapper[4824]: I1211 02:27:57.633550 4824 scope.go:117] "RemoveContainer" containerID="1d7148bf050b9a450d49eb09a5933cb6d8dd6490c731a4e84f5c37876b3c58a6" Dec 11 02:27:57 crc kubenswrapper[4824]: E1211 02:27:57.634630 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 02:28:02 crc kubenswrapper[4824]: I1211 02:28:02.081245 4824 scope.go:117] "RemoveContainer" containerID="ed28097c0c3c423f890fb5bc1a668d40016ed48cdda8c3ac88e9a5c4ad01814d" Dec 11 02:28:02 crc kubenswrapper[4824]: I1211 02:28:02.148056 4824 scope.go:117] "RemoveContainer" containerID="cbe3be8fd77c4538b4c67e74d833cf6343d64462c38ce95f2eb81652eb1887a2" Dec 11 02:28:02 crc kubenswrapper[4824]: I1211 02:28:02.193185 4824 scope.go:117] "RemoveContainer" containerID="3ad722591b153d578fa20c1ff377018b683efae930cebb1a49b0e62ef80c0355" Dec 11 02:28:12 crc kubenswrapper[4824]: I1211 02:28:12.632408 4824 scope.go:117] "RemoveContainer" containerID="1d7148bf050b9a450d49eb09a5933cb6d8dd6490c731a4e84f5c37876b3c58a6" Dec 11 02:28:12 crc kubenswrapper[4824]: E1211 02:28:12.635898 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 02:28:24 crc kubenswrapper[4824]: I1211 02:28:24.632658 4824 scope.go:117] "RemoveContainer" containerID="1d7148bf050b9a450d49eb09a5933cb6d8dd6490c731a4e84f5c37876b3c58a6" Dec 11 02:28:24 crc kubenswrapper[4824]: E1211 02:28:24.633561 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 02:28:36 crc kubenswrapper[4824]: I1211 02:28:36.632384 4824 scope.go:117] "RemoveContainer" containerID="1d7148bf050b9a450d49eb09a5933cb6d8dd6490c731a4e84f5c37876b3c58a6" Dec 11 02:28:36 crc kubenswrapper[4824]: E1211 02:28:36.633278 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 02:28:50 crc kubenswrapper[4824]: I1211 02:28:50.633803 4824 scope.go:117] "RemoveContainer" containerID="1d7148bf050b9a450d49eb09a5933cb6d8dd6490c731a4e84f5c37876b3c58a6" Dec 11 02:28:50 crc kubenswrapper[4824]: E1211 02:28:50.634942 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 02:29:02 crc kubenswrapper[4824]: I1211 02:29:02.296923 4824 scope.go:117] "RemoveContainer" containerID="e52b4e9af95e8babd1001beb6863b9e7df734a7ad25d0f34810a66eacde1e066" Dec 11 02:29:02 crc kubenswrapper[4824]: I1211 02:29:02.328959 4824 scope.go:117] "RemoveContainer" containerID="ed1dc24efac1fb0032a295396091f46900b22703e02321b6a39c6e82be838f84" Dec 11 02:29:02 crc kubenswrapper[4824]: I1211 02:29:02.360456 4824 scope.go:117] "RemoveContainer" containerID="6a65d85a224360efdc1c1ccd3fb02c778ea2c034d15593e608fd2424c05be1ed" Dec 11 02:29:02 crc kubenswrapper[4824]: I1211 02:29:02.389602 4824 scope.go:117] "RemoveContainer" containerID="5c21f9201fcef1fe44ad0114460c7e9d419836c72431082208bb4d6c4a529d23" Dec 11 02:29:02 crc kubenswrapper[4824]: I1211 02:29:02.633438 4824 scope.go:117] "RemoveContainer" containerID="1d7148bf050b9a450d49eb09a5933cb6d8dd6490c731a4e84f5c37876b3c58a6" Dec 11 02:29:02 crc kubenswrapper[4824]: E1211 02:29:02.633986 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 02:29:17 crc kubenswrapper[4824]: I1211 02:29:17.633656 4824 scope.go:117] "RemoveContainer" containerID="1d7148bf050b9a450d49eb09a5933cb6d8dd6490c731a4e84f5c37876b3c58a6" Dec 11 02:29:17 crc kubenswrapper[4824]: E1211 02:29:17.634900 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 02:29:31 crc kubenswrapper[4824]: I1211 02:29:31.633434 4824 scope.go:117] "RemoveContainer" containerID="1d7148bf050b9a450d49eb09a5933cb6d8dd6490c731a4e84f5c37876b3c58a6" Dec 11 02:29:31 crc kubenswrapper[4824]: E1211 02:29:31.634717 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 02:29:42 crc kubenswrapper[4824]: I1211 02:29:42.633256 4824 scope.go:117] "RemoveContainer" containerID="1d7148bf050b9a450d49eb09a5933cb6d8dd6490c731a4e84f5c37876b3c58a6" Dec 11 02:29:42 crc kubenswrapper[4824]: E1211 02:29:42.633983 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 02:29:53 crc kubenswrapper[4824]: I1211 02:29:53.633384 4824 scope.go:117] "RemoveContainer" containerID="1d7148bf050b9a450d49eb09a5933cb6d8dd6490c731a4e84f5c37876b3c58a6" Dec 11 02:29:53 crc kubenswrapper[4824]: E1211 02:29:53.634314 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 02:30:00 crc kubenswrapper[4824]: I1211 02:30:00.168749 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423670-szq5v"] Dec 11 02:30:00 crc kubenswrapper[4824]: I1211 02:30:00.171528 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423670-szq5v" Dec 11 02:30:00 crc kubenswrapper[4824]: I1211 02:30:00.175736 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 11 02:30:00 crc kubenswrapper[4824]: I1211 02:30:00.178538 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 11 02:30:00 crc kubenswrapper[4824]: I1211 02:30:00.195308 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423670-szq5v"] Dec 11 02:30:00 crc kubenswrapper[4824]: I1211 02:30:00.202777 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0104cb73-80d6-4553-9bab-1ead3adeaca2-secret-volume\") pod \"collect-profiles-29423670-szq5v\" (UID: \"0104cb73-80d6-4553-9bab-1ead3adeaca2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423670-szq5v" Dec 11 02:30:00 crc kubenswrapper[4824]: I1211 02:30:00.202837 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0104cb73-80d6-4553-9bab-1ead3adeaca2-config-volume\") pod \"collect-profiles-29423670-szq5v\" (UID: \"0104cb73-80d6-4553-9bab-1ead3adeaca2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423670-szq5v" Dec 11 02:30:00 crc kubenswrapper[4824]: I1211 02:30:00.202893 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q8mbp\" (UniqueName: \"kubernetes.io/projected/0104cb73-80d6-4553-9bab-1ead3adeaca2-kube-api-access-q8mbp\") pod \"collect-profiles-29423670-szq5v\" (UID: \"0104cb73-80d6-4553-9bab-1ead3adeaca2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423670-szq5v" Dec 11 02:30:00 crc kubenswrapper[4824]: I1211 02:30:00.304757 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0104cb73-80d6-4553-9bab-1ead3adeaca2-secret-volume\") pod \"collect-profiles-29423670-szq5v\" (UID: \"0104cb73-80d6-4553-9bab-1ead3adeaca2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423670-szq5v" Dec 11 02:30:00 crc kubenswrapper[4824]: I1211 02:30:00.304834 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0104cb73-80d6-4553-9bab-1ead3adeaca2-config-volume\") pod \"collect-profiles-29423670-szq5v\" (UID: \"0104cb73-80d6-4553-9bab-1ead3adeaca2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423670-szq5v" Dec 11 02:30:00 crc kubenswrapper[4824]: I1211 02:30:00.304876 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q8mbp\" (UniqueName: \"kubernetes.io/projected/0104cb73-80d6-4553-9bab-1ead3adeaca2-kube-api-access-q8mbp\") pod \"collect-profiles-29423670-szq5v\" (UID: \"0104cb73-80d6-4553-9bab-1ead3adeaca2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423670-szq5v" Dec 11 02:30:00 crc kubenswrapper[4824]: I1211 02:30:00.306603 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0104cb73-80d6-4553-9bab-1ead3adeaca2-config-volume\") pod \"collect-profiles-29423670-szq5v\" (UID: \"0104cb73-80d6-4553-9bab-1ead3adeaca2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423670-szq5v" Dec 11 02:30:00 crc kubenswrapper[4824]: I1211 02:30:00.320987 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0104cb73-80d6-4553-9bab-1ead3adeaca2-secret-volume\") pod \"collect-profiles-29423670-szq5v\" (UID: \"0104cb73-80d6-4553-9bab-1ead3adeaca2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423670-szq5v" Dec 11 02:30:00 crc kubenswrapper[4824]: I1211 02:30:00.322940 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q8mbp\" (UniqueName: \"kubernetes.io/projected/0104cb73-80d6-4553-9bab-1ead3adeaca2-kube-api-access-q8mbp\") pod \"collect-profiles-29423670-szq5v\" (UID: \"0104cb73-80d6-4553-9bab-1ead3adeaca2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423670-szq5v" Dec 11 02:30:00 crc kubenswrapper[4824]: I1211 02:30:00.501235 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423670-szq5v" Dec 11 02:30:01 crc kubenswrapper[4824]: I1211 02:30:01.053531 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423670-szq5v"] Dec 11 02:30:01 crc kubenswrapper[4824]: I1211 02:30:01.972691 4824 generic.go:334] "Generic (PLEG): container finished" podID="0104cb73-80d6-4553-9bab-1ead3adeaca2" containerID="3e60c204f2b971f9917d13f901f25bd11bc76d14ad92661319c0c2c3553c70cf" exitCode=0 Dec 11 02:30:01 crc kubenswrapper[4824]: I1211 02:30:01.972747 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423670-szq5v" event={"ID":"0104cb73-80d6-4553-9bab-1ead3adeaca2","Type":"ContainerDied","Data":"3e60c204f2b971f9917d13f901f25bd11bc76d14ad92661319c0c2c3553c70cf"} Dec 11 02:30:01 crc kubenswrapper[4824]: I1211 02:30:01.973055 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423670-szq5v" event={"ID":"0104cb73-80d6-4553-9bab-1ead3adeaca2","Type":"ContainerStarted","Data":"c8a5fcfccfe6d33402c55d91dbcfaedbda839f8ccee938dc0bfc84c6a106e0bd"} Dec 11 02:30:02 crc kubenswrapper[4824]: I1211 02:30:02.461492 4824 scope.go:117] "RemoveContainer" containerID="ae6988e689b7a435f7dff39aa726ebdb43772d5aad1cd52cf4fb6b62c24ae678" Dec 11 02:30:02 crc kubenswrapper[4824]: I1211 02:30:02.494632 4824 scope.go:117] "RemoveContainer" containerID="16899e5bc595474bd65445e7d16197f13d56f2ae311171be359b4486c19514f6" Dec 11 02:30:02 crc kubenswrapper[4824]: I1211 02:30:02.527671 4824 scope.go:117] "RemoveContainer" containerID="1f88f5ae1bf5cbbcf93220c49f7f08aa3db2f1f5e7c46ed76b5a5919ea60fd63" Dec 11 02:30:02 crc kubenswrapper[4824]: I1211 02:30:02.565438 4824 scope.go:117] "RemoveContainer" containerID="381b142df922b142dcfbebe34bef199920c42f16e2dba9b9cc688f157ba4e937" Dec 11 02:30:03 crc kubenswrapper[4824]: I1211 02:30:03.442064 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423670-szq5v" Dec 11 02:30:03 crc kubenswrapper[4824]: I1211 02:30:03.596184 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0104cb73-80d6-4553-9bab-1ead3adeaca2-secret-volume\") pod \"0104cb73-80d6-4553-9bab-1ead3adeaca2\" (UID: \"0104cb73-80d6-4553-9bab-1ead3adeaca2\") " Dec 11 02:30:03 crc kubenswrapper[4824]: I1211 02:30:03.596482 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q8mbp\" (UniqueName: \"kubernetes.io/projected/0104cb73-80d6-4553-9bab-1ead3adeaca2-kube-api-access-q8mbp\") pod \"0104cb73-80d6-4553-9bab-1ead3adeaca2\" (UID: \"0104cb73-80d6-4553-9bab-1ead3adeaca2\") " Dec 11 02:30:03 crc kubenswrapper[4824]: I1211 02:30:03.596555 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0104cb73-80d6-4553-9bab-1ead3adeaca2-config-volume\") pod \"0104cb73-80d6-4553-9bab-1ead3adeaca2\" (UID: \"0104cb73-80d6-4553-9bab-1ead3adeaca2\") " Dec 11 02:30:03 crc kubenswrapper[4824]: I1211 02:30:03.597826 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0104cb73-80d6-4553-9bab-1ead3adeaca2-config-volume" (OuterVolumeSpecName: "config-volume") pod "0104cb73-80d6-4553-9bab-1ead3adeaca2" (UID: "0104cb73-80d6-4553-9bab-1ead3adeaca2"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:30:03 crc kubenswrapper[4824]: I1211 02:30:03.606411 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0104cb73-80d6-4553-9bab-1ead3adeaca2-kube-api-access-q8mbp" (OuterVolumeSpecName: "kube-api-access-q8mbp") pod "0104cb73-80d6-4553-9bab-1ead3adeaca2" (UID: "0104cb73-80d6-4553-9bab-1ead3adeaca2"). InnerVolumeSpecName "kube-api-access-q8mbp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:30:03 crc kubenswrapper[4824]: I1211 02:30:03.607462 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0104cb73-80d6-4553-9bab-1ead3adeaca2-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "0104cb73-80d6-4553-9bab-1ead3adeaca2" (UID: "0104cb73-80d6-4553-9bab-1ead3adeaca2"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:30:03 crc kubenswrapper[4824]: I1211 02:30:03.699737 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q8mbp\" (UniqueName: \"kubernetes.io/projected/0104cb73-80d6-4553-9bab-1ead3adeaca2-kube-api-access-q8mbp\") on node \"crc\" DevicePath \"\"" Dec 11 02:30:03 crc kubenswrapper[4824]: I1211 02:30:03.699792 4824 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0104cb73-80d6-4553-9bab-1ead3adeaca2-config-volume\") on node \"crc\" DevicePath \"\"" Dec 11 02:30:03 crc kubenswrapper[4824]: I1211 02:30:03.699812 4824 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0104cb73-80d6-4553-9bab-1ead3adeaca2-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 11 02:30:04 crc kubenswrapper[4824]: I1211 02:30:04.006041 4824 generic.go:334] "Generic (PLEG): container finished" podID="1bf6cf36-d7eb-42be-aeee-c6824de8b6ba" containerID="ec51036ea596e760fa5982054877bd1b1ff5c19fcb0273dc04bf784fe3fd55bd" exitCode=0 Dec 11 02:30:04 crc kubenswrapper[4824]: I1211 02:30:04.006192 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bsknw" event={"ID":"1bf6cf36-d7eb-42be-aeee-c6824de8b6ba","Type":"ContainerDied","Data":"ec51036ea596e760fa5982054877bd1b1ff5c19fcb0273dc04bf784fe3fd55bd"} Dec 11 02:30:04 crc kubenswrapper[4824]: I1211 02:30:04.009271 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423670-szq5v" event={"ID":"0104cb73-80d6-4553-9bab-1ead3adeaca2","Type":"ContainerDied","Data":"c8a5fcfccfe6d33402c55d91dbcfaedbda839f8ccee938dc0bfc84c6a106e0bd"} Dec 11 02:30:04 crc kubenswrapper[4824]: I1211 02:30:04.009329 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c8a5fcfccfe6d33402c55d91dbcfaedbda839f8ccee938dc0bfc84c6a106e0bd" Dec 11 02:30:04 crc kubenswrapper[4824]: I1211 02:30:04.009344 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423670-szq5v" Dec 11 02:30:04 crc kubenswrapper[4824]: I1211 02:30:04.632637 4824 scope.go:117] "RemoveContainer" containerID="1d7148bf050b9a450d49eb09a5933cb6d8dd6490c731a4e84f5c37876b3c58a6" Dec 11 02:30:04 crc kubenswrapper[4824]: E1211 02:30:04.633536 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 02:30:05 crc kubenswrapper[4824]: I1211 02:30:05.663363 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bsknw" Dec 11 02:30:05 crc kubenswrapper[4824]: I1211 02:30:05.750140 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mbzqq\" (UniqueName: \"kubernetes.io/projected/1bf6cf36-d7eb-42be-aeee-c6824de8b6ba-kube-api-access-mbzqq\") pod \"1bf6cf36-d7eb-42be-aeee-c6824de8b6ba\" (UID: \"1bf6cf36-d7eb-42be-aeee-c6824de8b6ba\") " Dec 11 02:30:05 crc kubenswrapper[4824]: I1211 02:30:05.750247 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bf6cf36-d7eb-42be-aeee-c6824de8b6ba-bootstrap-combined-ca-bundle\") pod \"1bf6cf36-d7eb-42be-aeee-c6824de8b6ba\" (UID: \"1bf6cf36-d7eb-42be-aeee-c6824de8b6ba\") " Dec 11 02:30:05 crc kubenswrapper[4824]: I1211 02:30:05.750299 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1bf6cf36-d7eb-42be-aeee-c6824de8b6ba-inventory\") pod \"1bf6cf36-d7eb-42be-aeee-c6824de8b6ba\" (UID: \"1bf6cf36-d7eb-42be-aeee-c6824de8b6ba\") " Dec 11 02:30:05 crc kubenswrapper[4824]: I1211 02:30:05.750328 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1bf6cf36-d7eb-42be-aeee-c6824de8b6ba-ssh-key\") pod \"1bf6cf36-d7eb-42be-aeee-c6824de8b6ba\" (UID: \"1bf6cf36-d7eb-42be-aeee-c6824de8b6ba\") " Dec 11 02:30:05 crc kubenswrapper[4824]: I1211 02:30:05.757714 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf6cf36-d7eb-42be-aeee-c6824de8b6ba-kube-api-access-mbzqq" (OuterVolumeSpecName: "kube-api-access-mbzqq") pod "1bf6cf36-d7eb-42be-aeee-c6824de8b6ba" (UID: "1bf6cf36-d7eb-42be-aeee-c6824de8b6ba"). InnerVolumeSpecName "kube-api-access-mbzqq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:30:05 crc kubenswrapper[4824]: I1211 02:30:05.773484 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf6cf36-d7eb-42be-aeee-c6824de8b6ba-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "1bf6cf36-d7eb-42be-aeee-c6824de8b6ba" (UID: "1bf6cf36-d7eb-42be-aeee-c6824de8b6ba"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:30:05 crc kubenswrapper[4824]: I1211 02:30:05.819892 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf6cf36-d7eb-42be-aeee-c6824de8b6ba-inventory" (OuterVolumeSpecName: "inventory") pod "1bf6cf36-d7eb-42be-aeee-c6824de8b6ba" (UID: "1bf6cf36-d7eb-42be-aeee-c6824de8b6ba"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:30:05 crc kubenswrapper[4824]: I1211 02:30:05.828291 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf6cf36-d7eb-42be-aeee-c6824de8b6ba-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "1bf6cf36-d7eb-42be-aeee-c6824de8b6ba" (UID: "1bf6cf36-d7eb-42be-aeee-c6824de8b6ba"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:30:05 crc kubenswrapper[4824]: I1211 02:30:05.852322 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mbzqq\" (UniqueName: \"kubernetes.io/projected/1bf6cf36-d7eb-42be-aeee-c6824de8b6ba-kube-api-access-mbzqq\") on node \"crc\" DevicePath \"\"" Dec 11 02:30:05 crc kubenswrapper[4824]: I1211 02:30:05.852354 4824 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bf6cf36-d7eb-42be-aeee-c6824de8b6ba-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 02:30:05 crc kubenswrapper[4824]: I1211 02:30:05.852367 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1bf6cf36-d7eb-42be-aeee-c6824de8b6ba-inventory\") on node \"crc\" DevicePath \"\"" Dec 11 02:30:05 crc kubenswrapper[4824]: I1211 02:30:05.852379 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1bf6cf36-d7eb-42be-aeee-c6824de8b6ba-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 11 02:30:06 crc kubenswrapper[4824]: I1211 02:30:06.035054 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bsknw" event={"ID":"1bf6cf36-d7eb-42be-aeee-c6824de8b6ba","Type":"ContainerDied","Data":"b4d83fe70f0ef64308d132ca05182325d140cbddf622d0ed6381e85016a924b0"} Dec 11 02:30:06 crc kubenswrapper[4824]: I1211 02:30:06.035148 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b4d83fe70f0ef64308d132ca05182325d140cbddf622d0ed6381e85016a924b0" Dec 11 02:30:06 crc kubenswrapper[4824]: I1211 02:30:06.035512 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bsknw" Dec 11 02:30:06 crc kubenswrapper[4824]: I1211 02:30:06.154765 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-47r5p"] Dec 11 02:30:06 crc kubenswrapper[4824]: E1211 02:30:06.155515 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bf6cf36-d7eb-42be-aeee-c6824de8b6ba" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 11 02:30:06 crc kubenswrapper[4824]: I1211 02:30:06.155549 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bf6cf36-d7eb-42be-aeee-c6824de8b6ba" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 11 02:30:06 crc kubenswrapper[4824]: E1211 02:30:06.155577 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0104cb73-80d6-4553-9bab-1ead3adeaca2" containerName="collect-profiles" Dec 11 02:30:06 crc kubenswrapper[4824]: I1211 02:30:06.155590 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="0104cb73-80d6-4553-9bab-1ead3adeaca2" containerName="collect-profiles" Dec 11 02:30:06 crc kubenswrapper[4824]: I1211 02:30:06.155947 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bf6cf36-d7eb-42be-aeee-c6824de8b6ba" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 11 02:30:06 crc kubenswrapper[4824]: I1211 02:30:06.155987 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="0104cb73-80d6-4553-9bab-1ead3adeaca2" containerName="collect-profiles" Dec 11 02:30:06 crc kubenswrapper[4824]: I1211 02:30:06.156950 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-47r5p" Dec 11 02:30:06 crc kubenswrapper[4824]: I1211 02:30:06.160077 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-hs2fv" Dec 11 02:30:06 crc kubenswrapper[4824]: I1211 02:30:06.160086 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 11 02:30:06 crc kubenswrapper[4824]: I1211 02:30:06.160536 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 11 02:30:06 crc kubenswrapper[4824]: I1211 02:30:06.163181 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 11 02:30:06 crc kubenswrapper[4824]: I1211 02:30:06.172949 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-47r5p"] Dec 11 02:30:06 crc kubenswrapper[4824]: I1211 02:30:06.260327 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6c6c5ca9-ed78-4f51-a88b-12ec3df9efa2-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-47r5p\" (UID: \"6c6c5ca9-ed78-4f51-a88b-12ec3df9efa2\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-47r5p" Dec 11 02:30:06 crc kubenswrapper[4824]: I1211 02:30:06.260736 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5nkbz\" (UniqueName: \"kubernetes.io/projected/6c6c5ca9-ed78-4f51-a88b-12ec3df9efa2-kube-api-access-5nkbz\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-47r5p\" (UID: \"6c6c5ca9-ed78-4f51-a88b-12ec3df9efa2\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-47r5p" Dec 11 02:30:06 crc kubenswrapper[4824]: I1211 02:30:06.260959 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6c6c5ca9-ed78-4f51-a88b-12ec3df9efa2-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-47r5p\" (UID: \"6c6c5ca9-ed78-4f51-a88b-12ec3df9efa2\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-47r5p" Dec 11 02:30:06 crc kubenswrapper[4824]: I1211 02:30:06.362542 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5nkbz\" (UniqueName: \"kubernetes.io/projected/6c6c5ca9-ed78-4f51-a88b-12ec3df9efa2-kube-api-access-5nkbz\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-47r5p\" (UID: \"6c6c5ca9-ed78-4f51-a88b-12ec3df9efa2\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-47r5p" Dec 11 02:30:06 crc kubenswrapper[4824]: I1211 02:30:06.362648 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6c6c5ca9-ed78-4f51-a88b-12ec3df9efa2-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-47r5p\" (UID: \"6c6c5ca9-ed78-4f51-a88b-12ec3df9efa2\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-47r5p" Dec 11 02:30:06 crc kubenswrapper[4824]: I1211 02:30:06.362946 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6c6c5ca9-ed78-4f51-a88b-12ec3df9efa2-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-47r5p\" (UID: \"6c6c5ca9-ed78-4f51-a88b-12ec3df9efa2\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-47r5p" Dec 11 02:30:06 crc kubenswrapper[4824]: I1211 02:30:06.371272 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6c6c5ca9-ed78-4f51-a88b-12ec3df9efa2-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-47r5p\" (UID: \"6c6c5ca9-ed78-4f51-a88b-12ec3df9efa2\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-47r5p" Dec 11 02:30:06 crc kubenswrapper[4824]: I1211 02:30:06.372296 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6c6c5ca9-ed78-4f51-a88b-12ec3df9efa2-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-47r5p\" (UID: \"6c6c5ca9-ed78-4f51-a88b-12ec3df9efa2\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-47r5p" Dec 11 02:30:06 crc kubenswrapper[4824]: I1211 02:30:06.397285 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5nkbz\" (UniqueName: \"kubernetes.io/projected/6c6c5ca9-ed78-4f51-a88b-12ec3df9efa2-kube-api-access-5nkbz\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-47r5p\" (UID: \"6c6c5ca9-ed78-4f51-a88b-12ec3df9efa2\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-47r5p" Dec 11 02:30:06 crc kubenswrapper[4824]: I1211 02:30:06.500386 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-47r5p" Dec 11 02:30:07 crc kubenswrapper[4824]: I1211 02:30:07.186047 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-47r5p"] Dec 11 02:30:07 crc kubenswrapper[4824]: W1211 02:30:07.189053 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6c6c5ca9_ed78_4f51_a88b_12ec3df9efa2.slice/crio-4903a7a26efbabb12df19d0f5c74ebeac9787500cb4b681c105e2350a916d8a5 WatchSource:0}: Error finding container 4903a7a26efbabb12df19d0f5c74ebeac9787500cb4b681c105e2350a916d8a5: Status 404 returned error can't find the container with id 4903a7a26efbabb12df19d0f5c74ebeac9787500cb4b681c105e2350a916d8a5 Dec 11 02:30:07 crc kubenswrapper[4824]: I1211 02:30:07.192322 4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 11 02:30:08 crc kubenswrapper[4824]: I1211 02:30:08.057831 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-47r5p" event={"ID":"6c6c5ca9-ed78-4f51-a88b-12ec3df9efa2","Type":"ContainerStarted","Data":"60b69e58d8875b485075e1a5106b3b96a87d36c4b6de86042451507f94cb7261"} Dec 11 02:30:08 crc kubenswrapper[4824]: I1211 02:30:08.058148 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-47r5p" event={"ID":"6c6c5ca9-ed78-4f51-a88b-12ec3df9efa2","Type":"ContainerStarted","Data":"4903a7a26efbabb12df19d0f5c74ebeac9787500cb4b681c105e2350a916d8a5"} Dec 11 02:30:09 crc kubenswrapper[4824]: I1211 02:30:09.109604 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-47r5p" podStartSLOduration=2.913136873 podStartE2EDuration="3.10957356s" podCreationTimestamp="2025-12-11 02:30:06 +0000 UTC" firstStartedPulling="2025-12-11 02:30:07.1920017 +0000 UTC m=+1748.881039089" lastFinishedPulling="2025-12-11 02:30:07.388438387 +0000 UTC m=+1749.077475776" observedRunningTime="2025-12-11 02:30:09.093928226 +0000 UTC m=+1750.782965675" watchObservedRunningTime="2025-12-11 02:30:09.10957356 +0000 UTC m=+1750.798610979" Dec 11 02:30:18 crc kubenswrapper[4824]: I1211 02:30:18.648234 4824 scope.go:117] "RemoveContainer" containerID="1d7148bf050b9a450d49eb09a5933cb6d8dd6490c731a4e84f5c37876b3c58a6" Dec 11 02:30:18 crc kubenswrapper[4824]: E1211 02:30:18.648838 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 02:30:29 crc kubenswrapper[4824]: I1211 02:30:29.633258 4824 scope.go:117] "RemoveContainer" containerID="1d7148bf050b9a450d49eb09a5933cb6d8dd6490c731a4e84f5c37876b3c58a6" Dec 11 02:30:29 crc kubenswrapper[4824]: E1211 02:30:29.634554 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 02:30:43 crc kubenswrapper[4824]: I1211 02:30:43.632727 4824 scope.go:117] "RemoveContainer" containerID="1d7148bf050b9a450d49eb09a5933cb6d8dd6490c731a4e84f5c37876b3c58a6" Dec 11 02:30:43 crc kubenswrapper[4824]: E1211 02:30:43.633542 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 02:30:50 crc kubenswrapper[4824]: I1211 02:30:50.057874 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-4d06-account-create-update-59g6f"] Dec 11 02:30:50 crc kubenswrapper[4824]: I1211 02:30:50.077726 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-zwsrx"] Dec 11 02:30:50 crc kubenswrapper[4824]: I1211 02:30:50.093092 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-zwsrx"] Dec 11 02:30:50 crc kubenswrapper[4824]: I1211 02:30:50.103777 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-4d06-account-create-update-59g6f"] Dec 11 02:30:50 crc kubenswrapper[4824]: I1211 02:30:50.657334 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2549107d-73ea-4e06-94a5-ff7a1f6b30b6" path="/var/lib/kubelet/pods/2549107d-73ea-4e06-94a5-ff7a1f6b30b6/volumes" Dec 11 02:30:50 crc kubenswrapper[4824]: I1211 02:30:50.658585 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cf6d5db8-d6c0-4e6e-b8d1-22b94c561060" path="/var/lib/kubelet/pods/cf6d5db8-d6c0-4e6e-b8d1-22b94c561060/volumes" Dec 11 02:30:51 crc kubenswrapper[4824]: I1211 02:30:51.050346 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-5da1-account-create-update-g5bs8"] Dec 11 02:30:51 crc kubenswrapper[4824]: I1211 02:30:51.063708 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-cjh7n"] Dec 11 02:30:51 crc kubenswrapper[4824]: I1211 02:30:51.073246 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-7znqz"] Dec 11 02:30:51 crc kubenswrapper[4824]: I1211 02:30:51.082973 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-7znqz"] Dec 11 02:30:51 crc kubenswrapper[4824]: I1211 02:30:51.094357 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-5da1-account-create-update-g5bs8"] Dec 11 02:30:51 crc kubenswrapper[4824]: I1211 02:30:51.112825 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-d963-account-create-update-k6d7n"] Dec 11 02:30:51 crc kubenswrapper[4824]: I1211 02:30:51.124476 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-cjh7n"] Dec 11 02:30:51 crc kubenswrapper[4824]: I1211 02:30:51.134938 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-d963-account-create-update-k6d7n"] Dec 11 02:30:52 crc kubenswrapper[4824]: I1211 02:30:52.643373 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="77484236-1254-407f-950e-2a4861f10ead" path="/var/lib/kubelet/pods/77484236-1254-407f-950e-2a4861f10ead/volumes" Dec 11 02:30:52 crc kubenswrapper[4824]: I1211 02:30:52.644986 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c259c18-68cf-4a86-9870-4f481ba4b12a" path="/var/lib/kubelet/pods/7c259c18-68cf-4a86-9870-4f481ba4b12a/volumes" Dec 11 02:30:52 crc kubenswrapper[4824]: I1211 02:30:52.645576 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d62e4855-c5c9-4d8b-ab6d-7695bbbe3f6b" path="/var/lib/kubelet/pods/d62e4855-c5c9-4d8b-ab6d-7695bbbe3f6b/volumes" Dec 11 02:30:52 crc kubenswrapper[4824]: I1211 02:30:52.646147 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7b35324-f201-4847-9e8e-062029d43cee" path="/var/lib/kubelet/pods/d7b35324-f201-4847-9e8e-062029d43cee/volumes" Dec 11 02:30:54 crc kubenswrapper[4824]: I1211 02:30:54.633099 4824 scope.go:117] "RemoveContainer" containerID="1d7148bf050b9a450d49eb09a5933cb6d8dd6490c731a4e84f5c37876b3c58a6" Dec 11 02:30:54 crc kubenswrapper[4824]: E1211 02:30:54.634232 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 02:31:02 crc kubenswrapper[4824]: I1211 02:31:02.636373 4824 scope.go:117] "RemoveContainer" containerID="d2188a3d6e6bd3e7da5caae01449214156d57fdd87701c6fdbe7d78f10cd6e90" Dec 11 02:31:02 crc kubenswrapper[4824]: I1211 02:31:02.664595 4824 scope.go:117] "RemoveContainer" containerID="e8f49b6cf914032bf3ef3528ec6072ebe413fa5371ea0fffa28848847f52611c" Dec 11 02:31:02 crc kubenswrapper[4824]: I1211 02:31:02.755792 4824 scope.go:117] "RemoveContainer" containerID="56177de03d14f23a4b5e3f313a8924b97023cd805443017f164b97c499007198" Dec 11 02:31:02 crc kubenswrapper[4824]: I1211 02:31:02.796279 4824 scope.go:117] "RemoveContainer" containerID="0943d6e1ee6b3e01f1a85b58a6a6a5d5fda8468e382b9355445cb109fc0a87a1" Dec 11 02:31:02 crc kubenswrapper[4824]: I1211 02:31:02.863889 4824 scope.go:117] "RemoveContainer" containerID="6ad4a4876b1daca5369d39a7eece8a9decffb4d62224968b7a5f400151acfd56" Dec 11 02:31:02 crc kubenswrapper[4824]: I1211 02:31:02.916692 4824 scope.go:117] "RemoveContainer" containerID="5fc6bc4a209be71b47ee41b3aff839a4ccef0e1483672ccc600c113e0e00f29e" Dec 11 02:31:07 crc kubenswrapper[4824]: I1211 02:31:07.632698 4824 scope.go:117] "RemoveContainer" containerID="1d7148bf050b9a450d49eb09a5933cb6d8dd6490c731a4e84f5c37876b3c58a6" Dec 11 02:31:07 crc kubenswrapper[4824]: E1211 02:31:07.633519 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 02:31:18 crc kubenswrapper[4824]: I1211 02:31:18.047875 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-4ctff"] Dec 11 02:31:18 crc kubenswrapper[4824]: I1211 02:31:18.056243 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-4ctff"] Dec 11 02:31:18 crc kubenswrapper[4824]: I1211 02:31:18.649102 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b4a47f6-97f5-4395-b82a-166b934174b5" path="/var/lib/kubelet/pods/3b4a47f6-97f5-4395-b82a-166b934174b5/volumes" Dec 11 02:31:22 crc kubenswrapper[4824]: I1211 02:31:22.633592 4824 scope.go:117] "RemoveContainer" containerID="1d7148bf050b9a450d49eb09a5933cb6d8dd6490c731a4e84f5c37876b3c58a6" Dec 11 02:31:22 crc kubenswrapper[4824]: E1211 02:31:22.634822 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 02:31:25 crc kubenswrapper[4824]: I1211 02:31:25.032496 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-bce9-account-create-update-qnmss"] Dec 11 02:31:25 crc kubenswrapper[4824]: I1211 02:31:25.067605 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-5a29-account-create-update-sb4qh"] Dec 11 02:31:25 crc kubenswrapper[4824]: I1211 02:31:25.075553 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-qwpmm"] Dec 11 02:31:25 crc kubenswrapper[4824]: I1211 02:31:25.082317 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-xg279"] Dec 11 02:31:25 crc kubenswrapper[4824]: I1211 02:31:25.089085 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-qwpmm"] Dec 11 02:31:25 crc kubenswrapper[4824]: I1211 02:31:25.096241 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-bce9-account-create-update-qnmss"] Dec 11 02:31:25 crc kubenswrapper[4824]: I1211 02:31:25.102794 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-xg279"] Dec 11 02:31:25 crc kubenswrapper[4824]: I1211 02:31:25.109531 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-5a29-account-create-update-sb4qh"] Dec 11 02:31:26 crc kubenswrapper[4824]: I1211 02:31:26.652926 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20766632-27ff-42ab-9b3c-28aa8b8912bf" path="/var/lib/kubelet/pods/20766632-27ff-42ab-9b3c-28aa8b8912bf/volumes" Dec 11 02:31:26 crc kubenswrapper[4824]: I1211 02:31:26.654107 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57686bda-9117-48cc-b29d-f5ff374937c9" path="/var/lib/kubelet/pods/57686bda-9117-48cc-b29d-f5ff374937c9/volumes" Dec 11 02:31:26 crc kubenswrapper[4824]: I1211 02:31:26.655223 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="793e155e-73e9-46d5-9de6-a4aebf3ca2b4" path="/var/lib/kubelet/pods/793e155e-73e9-46d5-9de6-a4aebf3ca2b4/volumes" Dec 11 02:31:26 crc kubenswrapper[4824]: I1211 02:31:26.656357 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d9a4a8f-d58e-4a90-9c33-3ffa6a7f5c36" path="/var/lib/kubelet/pods/9d9a4a8f-d58e-4a90-9c33-3ffa6a7f5c36/volumes" Dec 11 02:31:29 crc kubenswrapper[4824]: I1211 02:31:29.035842 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-70b2-account-create-update-fzkfv"] Dec 11 02:31:29 crc kubenswrapper[4824]: I1211 02:31:29.046475 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-70b2-account-create-update-fzkfv"] Dec 11 02:31:29 crc kubenswrapper[4824]: I1211 02:31:29.056886 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-9hwb8"] Dec 11 02:31:29 crc kubenswrapper[4824]: I1211 02:31:29.068189 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-9hwb8"] Dec 11 02:31:30 crc kubenswrapper[4824]: I1211 02:31:30.652080 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2bcb322d-a528-4d31-b639-367f5f7f3dd1" path="/var/lib/kubelet/pods/2bcb322d-a528-4d31-b639-367f5f7f3dd1/volumes" Dec 11 02:31:30 crc kubenswrapper[4824]: I1211 02:31:30.653937 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ab730c6-95e4-438f-90d5-04812d26384f" path="/var/lib/kubelet/pods/4ab730c6-95e4-438f-90d5-04812d26384f/volumes" Dec 11 02:31:33 crc kubenswrapper[4824]: I1211 02:31:33.055446 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-wdzl5"] Dec 11 02:31:33 crc kubenswrapper[4824]: I1211 02:31:33.074140 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-wdzl5"] Dec 11 02:31:34 crc kubenswrapper[4824]: I1211 02:31:34.651192 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="922cc36d-5fd8-4cba-b6e3-5b16e6cb7531" path="/var/lib/kubelet/pods/922cc36d-5fd8-4cba-b6e3-5b16e6cb7531/volumes" Dec 11 02:31:37 crc kubenswrapper[4824]: I1211 02:31:37.633177 4824 scope.go:117] "RemoveContainer" containerID="1d7148bf050b9a450d49eb09a5933cb6d8dd6490c731a4e84f5c37876b3c58a6" Dec 11 02:31:38 crc kubenswrapper[4824]: I1211 02:31:38.287977 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" event={"ID":"44b156e4-64a4-4d45-aa5e-9b10a862faed","Type":"ContainerStarted","Data":"80aa24371216c8cd3cae76744f65819f6cda361ceaf0439f8ef7d77b23dd30ff"} Dec 11 02:32:03 crc kubenswrapper[4824]: I1211 02:32:03.146702 4824 scope.go:117] "RemoveContainer" containerID="b8e0d5fc75fb9a79e6229c9786439bf7ba7382203aec403b4081a22d8bdf8d57" Dec 11 02:32:03 crc kubenswrapper[4824]: I1211 02:32:03.182574 4824 scope.go:117] "RemoveContainer" containerID="0496ed10e37ac8801f313f31fb43102a538a86330c9c8f2f332bd62537eacb66" Dec 11 02:32:03 crc kubenswrapper[4824]: I1211 02:32:03.256701 4824 scope.go:117] "RemoveContainer" containerID="3a4eac77f7a127683859cee1c67d38bf38413baa0c24d41f15af3eb0b9adaa95" Dec 11 02:32:03 crc kubenswrapper[4824]: I1211 02:32:03.299969 4824 scope.go:117] "RemoveContainer" containerID="4e15e7888c04c76b15f70f53b7ac680863bb27fccb96052650dd3487bf114c16" Dec 11 02:32:03 crc kubenswrapper[4824]: I1211 02:32:03.335769 4824 scope.go:117] "RemoveContainer" containerID="49694cc653f8c737dd7fc310645d2f7ea51a57f908795caeb53db67a847171bd" Dec 11 02:32:03 crc kubenswrapper[4824]: I1211 02:32:03.386185 4824 scope.go:117] "RemoveContainer" containerID="03ee038a6bf69c20fa36ea52fe0d3bec95eb293c2cc767dbb0d52a4f515cb1e2" Dec 11 02:32:03 crc kubenswrapper[4824]: I1211 02:32:03.447789 4824 scope.go:117] "RemoveContainer" containerID="4b2deaf1f85e37a911132c3ac53a75fb38c3326d25e7be47b5415a082094bb05" Dec 11 02:32:03 crc kubenswrapper[4824]: I1211 02:32:03.493765 4824 scope.go:117] "RemoveContainer" containerID="95cefc4c5cdf946297d7b1ed190c5ddc45f510eefea53ed37ce92e5daee34aa5" Dec 11 02:32:03 crc kubenswrapper[4824]: I1211 02:32:03.519953 4824 scope.go:117] "RemoveContainer" containerID="6f33dbd6ce00a9d9546c84dba259368ec1dc821d57f101e08545e1808635435b" Dec 11 02:32:03 crc kubenswrapper[4824]: I1211 02:32:03.547067 4824 scope.go:117] "RemoveContainer" containerID="13a2e99f97a6fd8fd4b48427476f27d577b17501906194f19debf85ca7f2b4c4" Dec 11 02:32:03 crc kubenswrapper[4824]: I1211 02:32:03.612424 4824 generic.go:334] "Generic (PLEG): container finished" podID="6c6c5ca9-ed78-4f51-a88b-12ec3df9efa2" containerID="60b69e58d8875b485075e1a5106b3b96a87d36c4b6de86042451507f94cb7261" exitCode=0 Dec 11 02:32:03 crc kubenswrapper[4824]: I1211 02:32:03.612471 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-47r5p" event={"ID":"6c6c5ca9-ed78-4f51-a88b-12ec3df9efa2","Type":"ContainerDied","Data":"60b69e58d8875b485075e1a5106b3b96a87d36c4b6de86042451507f94cb7261"} Dec 11 02:32:04 crc kubenswrapper[4824]: I1211 02:32:04.069284 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-nm5bg"] Dec 11 02:32:04 crc kubenswrapper[4824]: I1211 02:32:04.079686 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-nm5bg"] Dec 11 02:32:04 crc kubenswrapper[4824]: I1211 02:32:04.660194 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="07b76bc0-c213-45d8-bebd-06b3e70322ab" path="/var/lib/kubelet/pods/07b76bc0-c213-45d8-bebd-06b3e70322ab/volumes" Dec 11 02:32:05 crc kubenswrapper[4824]: I1211 02:32:05.042966 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-lpwl8"] Dec 11 02:32:05 crc kubenswrapper[4824]: I1211 02:32:05.056180 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-lpwl8"] Dec 11 02:32:05 crc kubenswrapper[4824]: I1211 02:32:05.132617 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-47r5p" Dec 11 02:32:05 crc kubenswrapper[4824]: I1211 02:32:05.216410 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6c6c5ca9-ed78-4f51-a88b-12ec3df9efa2-ssh-key\") pod \"6c6c5ca9-ed78-4f51-a88b-12ec3df9efa2\" (UID: \"6c6c5ca9-ed78-4f51-a88b-12ec3df9efa2\") " Dec 11 02:32:05 crc kubenswrapper[4824]: I1211 02:32:05.216518 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5nkbz\" (UniqueName: \"kubernetes.io/projected/6c6c5ca9-ed78-4f51-a88b-12ec3df9efa2-kube-api-access-5nkbz\") pod \"6c6c5ca9-ed78-4f51-a88b-12ec3df9efa2\" (UID: \"6c6c5ca9-ed78-4f51-a88b-12ec3df9efa2\") " Dec 11 02:32:05 crc kubenswrapper[4824]: I1211 02:32:05.216677 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6c6c5ca9-ed78-4f51-a88b-12ec3df9efa2-inventory\") pod \"6c6c5ca9-ed78-4f51-a88b-12ec3df9efa2\" (UID: \"6c6c5ca9-ed78-4f51-a88b-12ec3df9efa2\") " Dec 11 02:32:05 crc kubenswrapper[4824]: I1211 02:32:05.224435 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c6c5ca9-ed78-4f51-a88b-12ec3df9efa2-kube-api-access-5nkbz" (OuterVolumeSpecName: "kube-api-access-5nkbz") pod "6c6c5ca9-ed78-4f51-a88b-12ec3df9efa2" (UID: "6c6c5ca9-ed78-4f51-a88b-12ec3df9efa2"). InnerVolumeSpecName "kube-api-access-5nkbz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:32:05 crc kubenswrapper[4824]: I1211 02:32:05.263443 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c6c5ca9-ed78-4f51-a88b-12ec3df9efa2-inventory" (OuterVolumeSpecName: "inventory") pod "6c6c5ca9-ed78-4f51-a88b-12ec3df9efa2" (UID: "6c6c5ca9-ed78-4f51-a88b-12ec3df9efa2"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:32:05 crc kubenswrapper[4824]: I1211 02:32:05.279899 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c6c5ca9-ed78-4f51-a88b-12ec3df9efa2-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "6c6c5ca9-ed78-4f51-a88b-12ec3df9efa2" (UID: "6c6c5ca9-ed78-4f51-a88b-12ec3df9efa2"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:32:05 crc kubenswrapper[4824]: I1211 02:32:05.318950 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6c6c5ca9-ed78-4f51-a88b-12ec3df9efa2-inventory\") on node \"crc\" DevicePath \"\"" Dec 11 02:32:05 crc kubenswrapper[4824]: I1211 02:32:05.318986 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6c6c5ca9-ed78-4f51-a88b-12ec3df9efa2-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 11 02:32:05 crc kubenswrapper[4824]: I1211 02:32:05.318997 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5nkbz\" (UniqueName: \"kubernetes.io/projected/6c6c5ca9-ed78-4f51-a88b-12ec3df9efa2-kube-api-access-5nkbz\") on node \"crc\" DevicePath \"\"" Dec 11 02:32:05 crc kubenswrapper[4824]: I1211 02:32:05.639466 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-47r5p" event={"ID":"6c6c5ca9-ed78-4f51-a88b-12ec3df9efa2","Type":"ContainerDied","Data":"4903a7a26efbabb12df19d0f5c74ebeac9787500cb4b681c105e2350a916d8a5"} Dec 11 02:32:05 crc kubenswrapper[4824]: I1211 02:32:05.639790 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4903a7a26efbabb12df19d0f5c74ebeac9787500cb4b681c105e2350a916d8a5" Dec 11 02:32:05 crc kubenswrapper[4824]: I1211 02:32:05.639611 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-47r5p" Dec 11 02:32:05 crc kubenswrapper[4824]: I1211 02:32:05.722797 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-v4fz9"] Dec 11 02:32:05 crc kubenswrapper[4824]: E1211 02:32:05.723415 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c6c5ca9-ed78-4f51-a88b-12ec3df9efa2" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 11 02:32:05 crc kubenswrapper[4824]: I1211 02:32:05.723437 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c6c5ca9-ed78-4f51-a88b-12ec3df9efa2" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 11 02:32:05 crc kubenswrapper[4824]: I1211 02:32:05.723795 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c6c5ca9-ed78-4f51-a88b-12ec3df9efa2" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 11 02:32:05 crc kubenswrapper[4824]: I1211 02:32:05.724856 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-v4fz9" Dec 11 02:32:05 crc kubenswrapper[4824]: I1211 02:32:05.728485 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 11 02:32:05 crc kubenswrapper[4824]: I1211 02:32:05.728875 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 11 02:32:05 crc kubenswrapper[4824]: I1211 02:32:05.729137 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 11 02:32:05 crc kubenswrapper[4824]: I1211 02:32:05.730254 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-hs2fv" Dec 11 02:32:05 crc kubenswrapper[4824]: I1211 02:32:05.731059 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-v4fz9"] Dec 11 02:32:05 crc kubenswrapper[4824]: I1211 02:32:05.829604 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9n4zc\" (UniqueName: \"kubernetes.io/projected/aeee6ae2-66ac-45e6-9a98-753649777abe-kube-api-access-9n4zc\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-v4fz9\" (UID: \"aeee6ae2-66ac-45e6-9a98-753649777abe\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-v4fz9" Dec 11 02:32:05 crc kubenswrapper[4824]: I1211 02:32:05.829715 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aeee6ae2-66ac-45e6-9a98-753649777abe-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-v4fz9\" (UID: \"aeee6ae2-66ac-45e6-9a98-753649777abe\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-v4fz9" Dec 11 02:32:05 crc kubenswrapper[4824]: I1211 02:32:05.829750 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aeee6ae2-66ac-45e6-9a98-753649777abe-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-v4fz9\" (UID: \"aeee6ae2-66ac-45e6-9a98-753649777abe\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-v4fz9" Dec 11 02:32:05 crc kubenswrapper[4824]: I1211 02:32:05.931792 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9n4zc\" (UniqueName: \"kubernetes.io/projected/aeee6ae2-66ac-45e6-9a98-753649777abe-kube-api-access-9n4zc\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-v4fz9\" (UID: \"aeee6ae2-66ac-45e6-9a98-753649777abe\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-v4fz9" Dec 11 02:32:05 crc kubenswrapper[4824]: I1211 02:32:05.931899 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aeee6ae2-66ac-45e6-9a98-753649777abe-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-v4fz9\" (UID: \"aeee6ae2-66ac-45e6-9a98-753649777abe\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-v4fz9" Dec 11 02:32:05 crc kubenswrapper[4824]: I1211 02:32:05.931940 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aeee6ae2-66ac-45e6-9a98-753649777abe-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-v4fz9\" (UID: \"aeee6ae2-66ac-45e6-9a98-753649777abe\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-v4fz9" Dec 11 02:32:05 crc kubenswrapper[4824]: I1211 02:32:05.938045 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aeee6ae2-66ac-45e6-9a98-753649777abe-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-v4fz9\" (UID: \"aeee6ae2-66ac-45e6-9a98-753649777abe\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-v4fz9" Dec 11 02:32:05 crc kubenswrapper[4824]: I1211 02:32:05.938103 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aeee6ae2-66ac-45e6-9a98-753649777abe-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-v4fz9\" (UID: \"aeee6ae2-66ac-45e6-9a98-753649777abe\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-v4fz9" Dec 11 02:32:05 crc kubenswrapper[4824]: I1211 02:32:05.953367 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9n4zc\" (UniqueName: \"kubernetes.io/projected/aeee6ae2-66ac-45e6-9a98-753649777abe-kube-api-access-9n4zc\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-v4fz9\" (UID: \"aeee6ae2-66ac-45e6-9a98-753649777abe\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-v4fz9" Dec 11 02:32:06 crc kubenswrapper[4824]: I1211 02:32:06.049699 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-v4fz9" Dec 11 02:32:06 crc kubenswrapper[4824]: I1211 02:32:06.642267 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68cc858a-9296-4a34-8322-b2dabc8ac1a2" path="/var/lib/kubelet/pods/68cc858a-9296-4a34-8322-b2dabc8ac1a2/volumes" Dec 11 02:32:06 crc kubenswrapper[4824]: I1211 02:32:06.708938 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-v4fz9"] Dec 11 02:32:07 crc kubenswrapper[4824]: I1211 02:32:07.662318 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-v4fz9" event={"ID":"aeee6ae2-66ac-45e6-9a98-753649777abe","Type":"ContainerStarted","Data":"6bbb949cbebf5f7fd32cfbf0c69db7a8cb8cdf89cbc276e369bb566f332a1f51"} Dec 11 02:32:07 crc kubenswrapper[4824]: I1211 02:32:07.662646 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-v4fz9" event={"ID":"aeee6ae2-66ac-45e6-9a98-753649777abe","Type":"ContainerStarted","Data":"aafa8c929bb224e937de9402b323881f9cd81ac8fcb0b944be38bde45b3a148f"} Dec 11 02:32:08 crc kubenswrapper[4824]: I1211 02:32:08.695028 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-v4fz9" podStartSLOduration=3.554346584 podStartE2EDuration="3.695001622s" podCreationTimestamp="2025-12-11 02:32:05 +0000 UTC" firstStartedPulling="2025-12-11 02:32:06.699132623 +0000 UTC m=+1868.388170002" lastFinishedPulling="2025-12-11 02:32:06.839787671 +0000 UTC m=+1868.528825040" observedRunningTime="2025-12-11 02:32:08.684706803 +0000 UTC m=+1870.373744212" watchObservedRunningTime="2025-12-11 02:32:08.695001622 +0000 UTC m=+1870.384039011" Dec 11 02:32:11 crc kubenswrapper[4824]: I1211 02:32:11.045701 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-752rf"] Dec 11 02:32:11 crc kubenswrapper[4824]: I1211 02:32:11.059482 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-752rf"] Dec 11 02:32:12 crc kubenswrapper[4824]: I1211 02:32:12.649604 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68990bd4-eecf-4181-a76c-bf0d24154774" path="/var/lib/kubelet/pods/68990bd4-eecf-4181-a76c-bf0d24154774/volumes" Dec 11 02:32:20 crc kubenswrapper[4824]: I1211 02:32:20.042298 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-pbsrf"] Dec 11 02:32:20 crc kubenswrapper[4824]: I1211 02:32:20.050193 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-pbsrf"] Dec 11 02:32:20 crc kubenswrapper[4824]: I1211 02:32:20.648187 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d536a1b1-baa2-497a-ba0c-ad03599168cc" path="/var/lib/kubelet/pods/d536a1b1-baa2-497a-ba0c-ad03599168cc/volumes" Dec 11 02:32:23 crc kubenswrapper[4824]: I1211 02:32:23.051704 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-tn658"] Dec 11 02:32:23 crc kubenswrapper[4824]: I1211 02:32:23.064032 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-tn658"] Dec 11 02:32:24 crc kubenswrapper[4824]: I1211 02:32:24.653786 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7" path="/var/lib/kubelet/pods/8116a08e-ffca-48a9-87dd-1a0b2a8a1cd7/volumes" Dec 11 02:33:03 crc kubenswrapper[4824]: I1211 02:33:03.073877 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-kchr8"] Dec 11 02:33:03 crc kubenswrapper[4824]: I1211 02:33:03.092072 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-kchr8"] Dec 11 02:33:03 crc kubenswrapper[4824]: I1211 02:33:03.766033 4824 scope.go:117] "RemoveContainer" containerID="6b97e27f717fafb1660abc5454cbf78410b26ec165a238a422af3cc98a9cf41b" Dec 11 02:33:03 crc kubenswrapper[4824]: I1211 02:33:03.798255 4824 scope.go:117] "RemoveContainer" containerID="a3b44f56c6628b1b240b0bad78da5c6cc9a39004a9cb7deab6fe0bd7586b95a4" Dec 11 02:33:03 crc kubenswrapper[4824]: I1211 02:33:03.836845 4824 scope.go:117] "RemoveContainer" containerID="61c05165b5969b15e231ae24f60dcbd4ca978cb8d81f0055fe9a2ab1ba1519d5" Dec 11 02:33:03 crc kubenswrapper[4824]: I1211 02:33:03.890239 4824 scope.go:117] "RemoveContainer" containerID="a01fcd21e235ca3330d07c56038ef155168ad9722f0844751b85828e6c035962" Dec 11 02:33:03 crc kubenswrapper[4824]: I1211 02:33:03.921716 4824 scope.go:117] "RemoveContainer" containerID="d38997cd7f66152cf930c9deb794b8e30589e769a2b2d85afbc34bbe959d1c6b" Dec 11 02:33:04 crc kubenswrapper[4824]: I1211 02:33:04.029892 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-kpv29"] Dec 11 02:33:04 crc kubenswrapper[4824]: I1211 02:33:04.037056 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-tqtd7"] Dec 11 02:33:04 crc kubenswrapper[4824]: I1211 02:33:04.044524 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-tqtd7"] Dec 11 02:33:04 crc kubenswrapper[4824]: I1211 02:33:04.052744 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-2225-account-create-update-f9zht"] Dec 11 02:33:04 crc kubenswrapper[4824]: I1211 02:33:04.060573 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-kpv29"] Dec 11 02:33:04 crc kubenswrapper[4824]: I1211 02:33:04.068144 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-32c8-account-create-update-stjs4"] Dec 11 02:33:04 crc kubenswrapper[4824]: I1211 02:33:04.075759 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-2225-account-create-update-f9zht"] Dec 11 02:33:04 crc kubenswrapper[4824]: I1211 02:33:04.082820 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-32c8-account-create-update-stjs4"] Dec 11 02:33:04 crc kubenswrapper[4824]: I1211 02:33:04.659476 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="14d1e80f-7c53-4c15-a002-eacde62186bc" path="/var/lib/kubelet/pods/14d1e80f-7c53-4c15-a002-eacde62186bc/volumes" Dec 11 02:33:04 crc kubenswrapper[4824]: I1211 02:33:04.660973 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="295fe4d0-5910-49d6-b8ac-56ff445666f9" path="/var/lib/kubelet/pods/295fe4d0-5910-49d6-b8ac-56ff445666f9/volumes" Dec 11 02:33:04 crc kubenswrapper[4824]: I1211 02:33:04.662401 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="33034069-6622-45c9-9045-44a5e01682bd" path="/var/lib/kubelet/pods/33034069-6622-45c9-9045-44a5e01682bd/volumes" Dec 11 02:33:04 crc kubenswrapper[4824]: I1211 02:33:04.663868 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb82f620-fb10-4f76-b2eb-adde800579bf" path="/var/lib/kubelet/pods/eb82f620-fb10-4f76-b2eb-adde800579bf/volumes" Dec 11 02:33:04 crc kubenswrapper[4824]: I1211 02:33:04.666309 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f0fefd9f-c55d-4bb0-8d9c-b3f7c42ac4d4" path="/var/lib/kubelet/pods/f0fefd9f-c55d-4bb0-8d9c-b3f7c42ac4d4/volumes" Dec 11 02:33:05 crc kubenswrapper[4824]: I1211 02:33:05.030872 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-7a4e-account-create-update-c7h9j"] Dec 11 02:33:05 crc kubenswrapper[4824]: I1211 02:33:05.040725 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-7a4e-account-create-update-c7h9j"] Dec 11 02:33:06 crc kubenswrapper[4824]: I1211 02:33:06.653333 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="78553f90-b619-4ff6-84b4-05835483a29e" path="/var/lib/kubelet/pods/78553f90-b619-4ff6-84b4-05835483a29e/volumes" Dec 11 02:33:25 crc kubenswrapper[4824]: I1211 02:33:25.473272 4824 generic.go:334] "Generic (PLEG): container finished" podID="aeee6ae2-66ac-45e6-9a98-753649777abe" containerID="6bbb949cbebf5f7fd32cfbf0c69db7a8cb8cdf89cbc276e369bb566f332a1f51" exitCode=0 Dec 11 02:33:25 crc kubenswrapper[4824]: I1211 02:33:25.473895 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-v4fz9" event={"ID":"aeee6ae2-66ac-45e6-9a98-753649777abe","Type":"ContainerDied","Data":"6bbb949cbebf5f7fd32cfbf0c69db7a8cb8cdf89cbc276e369bb566f332a1f51"} Dec 11 02:33:26 crc kubenswrapper[4824]: I1211 02:33:26.899757 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-v4fz9" Dec 11 02:33:26 crc kubenswrapper[4824]: I1211 02:33:26.932985 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aeee6ae2-66ac-45e6-9a98-753649777abe-ssh-key\") pod \"aeee6ae2-66ac-45e6-9a98-753649777abe\" (UID: \"aeee6ae2-66ac-45e6-9a98-753649777abe\") " Dec 11 02:33:26 crc kubenswrapper[4824]: I1211 02:33:26.933082 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9n4zc\" (UniqueName: \"kubernetes.io/projected/aeee6ae2-66ac-45e6-9a98-753649777abe-kube-api-access-9n4zc\") pod \"aeee6ae2-66ac-45e6-9a98-753649777abe\" (UID: \"aeee6ae2-66ac-45e6-9a98-753649777abe\") " Dec 11 02:33:26 crc kubenswrapper[4824]: I1211 02:33:26.933158 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aeee6ae2-66ac-45e6-9a98-753649777abe-inventory\") pod \"aeee6ae2-66ac-45e6-9a98-753649777abe\" (UID: \"aeee6ae2-66ac-45e6-9a98-753649777abe\") " Dec 11 02:33:26 crc kubenswrapper[4824]: I1211 02:33:26.987154 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aeee6ae2-66ac-45e6-9a98-753649777abe-inventory" (OuterVolumeSpecName: "inventory") pod "aeee6ae2-66ac-45e6-9a98-753649777abe" (UID: "aeee6ae2-66ac-45e6-9a98-753649777abe"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:33:26 crc kubenswrapper[4824]: I1211 02:33:26.987197 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aeee6ae2-66ac-45e6-9a98-753649777abe-kube-api-access-9n4zc" (OuterVolumeSpecName: "kube-api-access-9n4zc") pod "aeee6ae2-66ac-45e6-9a98-753649777abe" (UID: "aeee6ae2-66ac-45e6-9a98-753649777abe"). InnerVolumeSpecName "kube-api-access-9n4zc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:33:27 crc kubenswrapper[4824]: I1211 02:33:27.009377 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aeee6ae2-66ac-45e6-9a98-753649777abe-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "aeee6ae2-66ac-45e6-9a98-753649777abe" (UID: "aeee6ae2-66ac-45e6-9a98-753649777abe"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:33:27 crc kubenswrapper[4824]: I1211 02:33:27.036025 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9n4zc\" (UniqueName: \"kubernetes.io/projected/aeee6ae2-66ac-45e6-9a98-753649777abe-kube-api-access-9n4zc\") on node \"crc\" DevicePath \"\"" Dec 11 02:33:27 crc kubenswrapper[4824]: I1211 02:33:27.036060 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aeee6ae2-66ac-45e6-9a98-753649777abe-inventory\") on node \"crc\" DevicePath \"\"" Dec 11 02:33:27 crc kubenswrapper[4824]: I1211 02:33:27.036071 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aeee6ae2-66ac-45e6-9a98-753649777abe-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 11 02:33:27 crc kubenswrapper[4824]: I1211 02:33:27.492752 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-v4fz9" event={"ID":"aeee6ae2-66ac-45e6-9a98-753649777abe","Type":"ContainerDied","Data":"aafa8c929bb224e937de9402b323881f9cd81ac8fcb0b944be38bde45b3a148f"} Dec 11 02:33:27 crc kubenswrapper[4824]: I1211 02:33:27.492796 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aafa8c929bb224e937de9402b323881f9cd81ac8fcb0b944be38bde45b3a148f" Dec 11 02:33:27 crc kubenswrapper[4824]: I1211 02:33:27.492851 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-v4fz9" Dec 11 02:33:27 crc kubenswrapper[4824]: I1211 02:33:27.574697 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-t4pfx"] Dec 11 02:33:27 crc kubenswrapper[4824]: E1211 02:33:27.575371 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aeee6ae2-66ac-45e6-9a98-753649777abe" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 11 02:33:27 crc kubenswrapper[4824]: I1211 02:33:27.575389 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="aeee6ae2-66ac-45e6-9a98-753649777abe" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 11 02:33:27 crc kubenswrapper[4824]: I1211 02:33:27.575594 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="aeee6ae2-66ac-45e6-9a98-753649777abe" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 11 02:33:27 crc kubenswrapper[4824]: I1211 02:33:27.576564 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-t4pfx" Dec 11 02:33:27 crc kubenswrapper[4824]: I1211 02:33:27.577834 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 11 02:33:27 crc kubenswrapper[4824]: I1211 02:33:27.577924 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-hs2fv" Dec 11 02:33:27 crc kubenswrapper[4824]: I1211 02:33:27.578290 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 11 02:33:27 crc kubenswrapper[4824]: I1211 02:33:27.578508 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 11 02:33:27 crc kubenswrapper[4824]: I1211 02:33:27.590778 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-t4pfx"] Dec 11 02:33:27 crc kubenswrapper[4824]: I1211 02:33:27.747615 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f694423f-b8e7-49bd-9032-4b29a64fda4e-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-t4pfx\" (UID: \"f694423f-b8e7-49bd-9032-4b29a64fda4e\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-t4pfx" Dec 11 02:33:27 crc kubenswrapper[4824]: I1211 02:33:27.747779 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f694423f-b8e7-49bd-9032-4b29a64fda4e-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-t4pfx\" (UID: \"f694423f-b8e7-49bd-9032-4b29a64fda4e\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-t4pfx" Dec 11 02:33:27 crc kubenswrapper[4824]: I1211 02:33:27.747941 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6446\" (UniqueName: \"kubernetes.io/projected/f694423f-b8e7-49bd-9032-4b29a64fda4e-kube-api-access-m6446\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-t4pfx\" (UID: \"f694423f-b8e7-49bd-9032-4b29a64fda4e\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-t4pfx" Dec 11 02:33:27 crc kubenswrapper[4824]: I1211 02:33:27.849598 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m6446\" (UniqueName: \"kubernetes.io/projected/f694423f-b8e7-49bd-9032-4b29a64fda4e-kube-api-access-m6446\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-t4pfx\" (UID: \"f694423f-b8e7-49bd-9032-4b29a64fda4e\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-t4pfx" Dec 11 02:33:27 crc kubenswrapper[4824]: I1211 02:33:27.849691 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f694423f-b8e7-49bd-9032-4b29a64fda4e-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-t4pfx\" (UID: \"f694423f-b8e7-49bd-9032-4b29a64fda4e\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-t4pfx" Dec 11 02:33:27 crc kubenswrapper[4824]: I1211 02:33:27.849767 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f694423f-b8e7-49bd-9032-4b29a64fda4e-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-t4pfx\" (UID: \"f694423f-b8e7-49bd-9032-4b29a64fda4e\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-t4pfx" Dec 11 02:33:27 crc kubenswrapper[4824]: I1211 02:33:27.856913 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f694423f-b8e7-49bd-9032-4b29a64fda4e-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-t4pfx\" (UID: \"f694423f-b8e7-49bd-9032-4b29a64fda4e\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-t4pfx" Dec 11 02:33:27 crc kubenswrapper[4824]: I1211 02:33:27.858023 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f694423f-b8e7-49bd-9032-4b29a64fda4e-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-t4pfx\" (UID: \"f694423f-b8e7-49bd-9032-4b29a64fda4e\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-t4pfx" Dec 11 02:33:27 crc kubenswrapper[4824]: I1211 02:33:27.883496 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m6446\" (UniqueName: \"kubernetes.io/projected/f694423f-b8e7-49bd-9032-4b29a64fda4e-kube-api-access-m6446\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-t4pfx\" (UID: \"f694423f-b8e7-49bd-9032-4b29a64fda4e\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-t4pfx" Dec 11 02:33:27 crc kubenswrapper[4824]: I1211 02:33:27.899628 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-t4pfx" Dec 11 02:33:28 crc kubenswrapper[4824]: I1211 02:33:28.089191 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-lx54d"] Dec 11 02:33:28 crc kubenswrapper[4824]: I1211 02:33:28.100442 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-lx54d"] Dec 11 02:33:28 crc kubenswrapper[4824]: I1211 02:33:28.525299 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-t4pfx"] Dec 11 02:33:28 crc kubenswrapper[4824]: I1211 02:33:28.646266 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="10c51806-9b95-4b93-825f-45cd566834c2" path="/var/lib/kubelet/pods/10c51806-9b95-4b93-825f-45cd566834c2/volumes" Dec 11 02:33:29 crc kubenswrapper[4824]: I1211 02:33:29.515912 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-t4pfx" event={"ID":"f694423f-b8e7-49bd-9032-4b29a64fda4e","Type":"ContainerStarted","Data":"46216157a2d33f71a778c6bde8e60dbe1f04866627cdc3c76e3fb4a5140d3dae"} Dec 11 02:33:29 crc kubenswrapper[4824]: I1211 02:33:29.517085 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-t4pfx" event={"ID":"f694423f-b8e7-49bd-9032-4b29a64fda4e","Type":"ContainerStarted","Data":"6462242f6fc310880b313665ccd1289d8e2550c0747e3f0789d1ed95ffc4b7ff"} Dec 11 02:33:29 crc kubenswrapper[4824]: I1211 02:33:29.539482 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-t4pfx" podStartSLOduration=2.266407538 podStartE2EDuration="2.539462047s" podCreationTimestamp="2025-12-11 02:33:27 +0000 UTC" firstStartedPulling="2025-12-11 02:33:28.545558287 +0000 UTC m=+1950.234595706" lastFinishedPulling="2025-12-11 02:33:28.818612806 +0000 UTC m=+1950.507650215" observedRunningTime="2025-12-11 02:33:29.536215618 +0000 UTC m=+1951.225253017" watchObservedRunningTime="2025-12-11 02:33:29.539462047 +0000 UTC m=+1951.228499426" Dec 11 02:33:34 crc kubenswrapper[4824]: I1211 02:33:34.566028 4824 generic.go:334] "Generic (PLEG): container finished" podID="f694423f-b8e7-49bd-9032-4b29a64fda4e" containerID="46216157a2d33f71a778c6bde8e60dbe1f04866627cdc3c76e3fb4a5140d3dae" exitCode=0 Dec 11 02:33:34 crc kubenswrapper[4824]: I1211 02:33:34.566139 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-t4pfx" event={"ID":"f694423f-b8e7-49bd-9032-4b29a64fda4e","Type":"ContainerDied","Data":"46216157a2d33f71a778c6bde8e60dbe1f04866627cdc3c76e3fb4a5140d3dae"} Dec 11 02:33:36 crc kubenswrapper[4824]: I1211 02:33:36.109489 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-t4pfx" Dec 11 02:33:36 crc kubenswrapper[4824]: I1211 02:33:36.251763 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f694423f-b8e7-49bd-9032-4b29a64fda4e-ssh-key\") pod \"f694423f-b8e7-49bd-9032-4b29a64fda4e\" (UID: \"f694423f-b8e7-49bd-9032-4b29a64fda4e\") " Dec 11 02:33:36 crc kubenswrapper[4824]: I1211 02:33:36.251843 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f694423f-b8e7-49bd-9032-4b29a64fda4e-inventory\") pod \"f694423f-b8e7-49bd-9032-4b29a64fda4e\" (UID: \"f694423f-b8e7-49bd-9032-4b29a64fda4e\") " Dec 11 02:33:36 crc kubenswrapper[4824]: I1211 02:33:36.251975 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m6446\" (UniqueName: \"kubernetes.io/projected/f694423f-b8e7-49bd-9032-4b29a64fda4e-kube-api-access-m6446\") pod \"f694423f-b8e7-49bd-9032-4b29a64fda4e\" (UID: \"f694423f-b8e7-49bd-9032-4b29a64fda4e\") " Dec 11 02:33:36 crc kubenswrapper[4824]: I1211 02:33:36.263167 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f694423f-b8e7-49bd-9032-4b29a64fda4e-kube-api-access-m6446" (OuterVolumeSpecName: "kube-api-access-m6446") pod "f694423f-b8e7-49bd-9032-4b29a64fda4e" (UID: "f694423f-b8e7-49bd-9032-4b29a64fda4e"). InnerVolumeSpecName "kube-api-access-m6446". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:33:36 crc kubenswrapper[4824]: I1211 02:33:36.299745 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f694423f-b8e7-49bd-9032-4b29a64fda4e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f694423f-b8e7-49bd-9032-4b29a64fda4e" (UID: "f694423f-b8e7-49bd-9032-4b29a64fda4e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:33:36 crc kubenswrapper[4824]: I1211 02:33:36.300914 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f694423f-b8e7-49bd-9032-4b29a64fda4e-inventory" (OuterVolumeSpecName: "inventory") pod "f694423f-b8e7-49bd-9032-4b29a64fda4e" (UID: "f694423f-b8e7-49bd-9032-4b29a64fda4e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:33:36 crc kubenswrapper[4824]: I1211 02:33:36.354418 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m6446\" (UniqueName: \"kubernetes.io/projected/f694423f-b8e7-49bd-9032-4b29a64fda4e-kube-api-access-m6446\") on node \"crc\" DevicePath \"\"" Dec 11 02:33:36 crc kubenswrapper[4824]: I1211 02:33:36.354849 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f694423f-b8e7-49bd-9032-4b29a64fda4e-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 11 02:33:36 crc kubenswrapper[4824]: I1211 02:33:36.354869 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f694423f-b8e7-49bd-9032-4b29a64fda4e-inventory\") on node \"crc\" DevicePath \"\"" Dec 11 02:33:36 crc kubenswrapper[4824]: I1211 02:33:36.591306 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-t4pfx" event={"ID":"f694423f-b8e7-49bd-9032-4b29a64fda4e","Type":"ContainerDied","Data":"6462242f6fc310880b313665ccd1289d8e2550c0747e3f0789d1ed95ffc4b7ff"} Dec 11 02:33:36 crc kubenswrapper[4824]: I1211 02:33:36.591358 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6462242f6fc310880b313665ccd1289d8e2550c0747e3f0789d1ed95ffc4b7ff" Dec 11 02:33:36 crc kubenswrapper[4824]: I1211 02:33:36.591449 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-t4pfx" Dec 11 02:33:36 crc kubenswrapper[4824]: I1211 02:33:36.674480 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-9k74z"] Dec 11 02:33:36 crc kubenswrapper[4824]: E1211 02:33:36.676173 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f694423f-b8e7-49bd-9032-4b29a64fda4e" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 11 02:33:36 crc kubenswrapper[4824]: I1211 02:33:36.676221 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f694423f-b8e7-49bd-9032-4b29a64fda4e" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 11 02:33:36 crc kubenswrapper[4824]: I1211 02:33:36.676474 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f694423f-b8e7-49bd-9032-4b29a64fda4e" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 11 02:33:36 crc kubenswrapper[4824]: I1211 02:33:36.678270 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-9k74z" Dec 11 02:33:36 crc kubenswrapper[4824]: I1211 02:33:36.682829 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 11 02:33:36 crc kubenswrapper[4824]: I1211 02:33:36.683224 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 11 02:33:36 crc kubenswrapper[4824]: I1211 02:33:36.683563 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 11 02:33:36 crc kubenswrapper[4824]: I1211 02:33:36.684091 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-9k74z"] Dec 11 02:33:36 crc kubenswrapper[4824]: I1211 02:33:36.686637 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-hs2fv" Dec 11 02:33:36 crc kubenswrapper[4824]: I1211 02:33:36.762920 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a786c754-b2c4-4f19-b082-e8d64d47ac44-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-9k74z\" (UID: \"a786c754-b2c4-4f19-b082-e8d64d47ac44\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-9k74z" Dec 11 02:33:36 crc kubenswrapper[4824]: I1211 02:33:36.762990 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a786c754-b2c4-4f19-b082-e8d64d47ac44-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-9k74z\" (UID: \"a786c754-b2c4-4f19-b082-e8d64d47ac44\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-9k74z" Dec 11 02:33:36 crc kubenswrapper[4824]: I1211 02:33:36.763101 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9qfj\" (UniqueName: \"kubernetes.io/projected/a786c754-b2c4-4f19-b082-e8d64d47ac44-kube-api-access-h9qfj\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-9k74z\" (UID: \"a786c754-b2c4-4f19-b082-e8d64d47ac44\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-9k74z" Dec 11 02:33:36 crc kubenswrapper[4824]: I1211 02:33:36.865033 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a786c754-b2c4-4f19-b082-e8d64d47ac44-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-9k74z\" (UID: \"a786c754-b2c4-4f19-b082-e8d64d47ac44\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-9k74z" Dec 11 02:33:36 crc kubenswrapper[4824]: I1211 02:33:36.865187 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9qfj\" (UniqueName: \"kubernetes.io/projected/a786c754-b2c4-4f19-b082-e8d64d47ac44-kube-api-access-h9qfj\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-9k74z\" (UID: \"a786c754-b2c4-4f19-b082-e8d64d47ac44\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-9k74z" Dec 11 02:33:36 crc kubenswrapper[4824]: I1211 02:33:36.865291 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a786c754-b2c4-4f19-b082-e8d64d47ac44-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-9k74z\" (UID: \"a786c754-b2c4-4f19-b082-e8d64d47ac44\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-9k74z" Dec 11 02:33:36 crc kubenswrapper[4824]: I1211 02:33:36.869815 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a786c754-b2c4-4f19-b082-e8d64d47ac44-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-9k74z\" (UID: \"a786c754-b2c4-4f19-b082-e8d64d47ac44\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-9k74z" Dec 11 02:33:36 crc kubenswrapper[4824]: I1211 02:33:36.870612 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a786c754-b2c4-4f19-b082-e8d64d47ac44-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-9k74z\" (UID: \"a786c754-b2c4-4f19-b082-e8d64d47ac44\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-9k74z" Dec 11 02:33:36 crc kubenswrapper[4824]: I1211 02:33:36.883875 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9qfj\" (UniqueName: \"kubernetes.io/projected/a786c754-b2c4-4f19-b082-e8d64d47ac44-kube-api-access-h9qfj\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-9k74z\" (UID: \"a786c754-b2c4-4f19-b082-e8d64d47ac44\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-9k74z" Dec 11 02:33:36 crc kubenswrapper[4824]: I1211 02:33:36.997709 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-9k74z" Dec 11 02:33:37 crc kubenswrapper[4824]: I1211 02:33:37.250635 4824 patch_prober.go:28] interesting pod/machine-config-daemon-gx6xt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 02:33:37 crc kubenswrapper[4824]: I1211 02:33:37.250999 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 02:33:37 crc kubenswrapper[4824]: I1211 02:33:37.526769 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-9k74z"] Dec 11 02:33:37 crc kubenswrapper[4824]: W1211 02:33:37.531034 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda786c754_b2c4_4f19_b082_e8d64d47ac44.slice/crio-8ca9e3b655648d7c1fb8767acea84621627b63bd214a10cec1cec99ae49b98b5 WatchSource:0}: Error finding container 8ca9e3b655648d7c1fb8767acea84621627b63bd214a10cec1cec99ae49b98b5: Status 404 returned error can't find the container with id 8ca9e3b655648d7c1fb8767acea84621627b63bd214a10cec1cec99ae49b98b5 Dec 11 02:33:37 crc kubenswrapper[4824]: I1211 02:33:37.601536 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-9k74z" event={"ID":"a786c754-b2c4-4f19-b082-e8d64d47ac44","Type":"ContainerStarted","Data":"8ca9e3b655648d7c1fb8767acea84621627b63bd214a10cec1cec99ae49b98b5"} Dec 11 02:33:38 crc kubenswrapper[4824]: I1211 02:33:38.614927 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-9k74z" event={"ID":"a786c754-b2c4-4f19-b082-e8d64d47ac44","Type":"ContainerStarted","Data":"c08ee99e6eb258f123d12246e7f841a8efdc2cf76b436a75b440caee04c12d1d"} Dec 11 02:33:38 crc kubenswrapper[4824]: I1211 02:33:38.643627 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-9k74z" podStartSLOduration=2.473830686 podStartE2EDuration="2.643608222s" podCreationTimestamp="2025-12-11 02:33:36 +0000 UTC" firstStartedPulling="2025-12-11 02:33:37.53399136 +0000 UTC m=+1959.223028739" lastFinishedPulling="2025-12-11 02:33:37.703768896 +0000 UTC m=+1959.392806275" observedRunningTime="2025-12-11 02:33:38.636052627 +0000 UTC m=+1960.325090006" watchObservedRunningTime="2025-12-11 02:33:38.643608222 +0000 UTC m=+1960.332645601" Dec 11 02:33:50 crc kubenswrapper[4824]: I1211 02:33:50.062942 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-lzc98"] Dec 11 02:33:50 crc kubenswrapper[4824]: I1211 02:33:50.076357 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-lzc98"] Dec 11 02:33:50 crc kubenswrapper[4824]: I1211 02:33:50.652819 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f973a712-53f1-4d52-8767-c65f665dcb41" path="/var/lib/kubelet/pods/f973a712-53f1-4d52-8767-c65f665dcb41/volumes" Dec 11 02:33:51 crc kubenswrapper[4824]: I1211 02:33:51.049688 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-2k725"] Dec 11 02:33:51 crc kubenswrapper[4824]: I1211 02:33:51.057167 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-2k725"] Dec 11 02:33:52 crc kubenswrapper[4824]: I1211 02:33:52.650883 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a30aa0d2-55dc-4375-b72d-953a3c57043f" path="/var/lib/kubelet/pods/a30aa0d2-55dc-4375-b72d-953a3c57043f/volumes" Dec 11 02:34:04 crc kubenswrapper[4824]: I1211 02:34:04.077574 4824 scope.go:117] "RemoveContainer" containerID="6d0e5e492932c0331515a185fb06051ade0a660f2674731bbdf380d6d80fa182" Dec 11 02:34:04 crc kubenswrapper[4824]: I1211 02:34:04.112139 4824 scope.go:117] "RemoveContainer" containerID="b8e1b97c75d160789ea16a7148ba70c4ad22979eb01a893cc8e81de1b00e9d1d" Dec 11 02:34:04 crc kubenswrapper[4824]: I1211 02:34:04.157494 4824 scope.go:117] "RemoveContainer" containerID="46278df8294fe1a2c45a468015a14a6cd20406e47e9a25c003822d43a2f84332" Dec 11 02:34:04 crc kubenswrapper[4824]: I1211 02:34:04.210763 4824 scope.go:117] "RemoveContainer" containerID="9b989c8ec4608d3afac27cb73f4c53f059b0fa135849cc91dbe60b75eaa17e85" Dec 11 02:34:04 crc kubenswrapper[4824]: I1211 02:34:04.242954 4824 scope.go:117] "RemoveContainer" containerID="84eddbf85534e881ea9bf082a39086269e957efa5e6265a2e9b4b74a8ae4e37b" Dec 11 02:34:04 crc kubenswrapper[4824]: I1211 02:34:04.291813 4824 scope.go:117] "RemoveContainer" containerID="eec6b03b5b66e40b5d93beb65d557f9e3afa4db04c99efeb00feb123c145b50b" Dec 11 02:34:04 crc kubenswrapper[4824]: I1211 02:34:04.357080 4824 scope.go:117] "RemoveContainer" containerID="adc63ff1b02af6a4a9cc9409bb1319f46d832f8180fd67da9d5a27531bffdcea" Dec 11 02:34:04 crc kubenswrapper[4824]: I1211 02:34:04.393522 4824 scope.go:117] "RemoveContainer" containerID="6f30f175ac348afce59696aa8ba26d931409e4ccddc9f33846b288313d383a58" Dec 11 02:34:04 crc kubenswrapper[4824]: I1211 02:34:04.413266 4824 scope.go:117] "RemoveContainer" containerID="12e254d930693f63ccdfa5e852f7d30d78edf40699e481052f063ab4dc6cba73" Dec 11 02:34:07 crc kubenswrapper[4824]: I1211 02:34:07.252008 4824 patch_prober.go:28] interesting pod/machine-config-daemon-gx6xt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 02:34:07 crc kubenswrapper[4824]: I1211 02:34:07.252459 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 02:34:19 crc kubenswrapper[4824]: I1211 02:34:19.089662 4824 generic.go:334] "Generic (PLEG): container finished" podID="a786c754-b2c4-4f19-b082-e8d64d47ac44" containerID="c08ee99e6eb258f123d12246e7f841a8efdc2cf76b436a75b440caee04c12d1d" exitCode=0 Dec 11 02:34:19 crc kubenswrapper[4824]: I1211 02:34:19.089767 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-9k74z" event={"ID":"a786c754-b2c4-4f19-b082-e8d64d47ac44","Type":"ContainerDied","Data":"c08ee99e6eb258f123d12246e7f841a8efdc2cf76b436a75b440caee04c12d1d"} Dec 11 02:34:20 crc kubenswrapper[4824]: I1211 02:34:20.622368 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-9k74z" Dec 11 02:34:20 crc kubenswrapper[4824]: I1211 02:34:20.696324 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a786c754-b2c4-4f19-b082-e8d64d47ac44-inventory\") pod \"a786c754-b2c4-4f19-b082-e8d64d47ac44\" (UID: \"a786c754-b2c4-4f19-b082-e8d64d47ac44\") " Dec 11 02:34:20 crc kubenswrapper[4824]: I1211 02:34:20.696480 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h9qfj\" (UniqueName: \"kubernetes.io/projected/a786c754-b2c4-4f19-b082-e8d64d47ac44-kube-api-access-h9qfj\") pod \"a786c754-b2c4-4f19-b082-e8d64d47ac44\" (UID: \"a786c754-b2c4-4f19-b082-e8d64d47ac44\") " Dec 11 02:34:20 crc kubenswrapper[4824]: I1211 02:34:20.696532 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a786c754-b2c4-4f19-b082-e8d64d47ac44-ssh-key\") pod \"a786c754-b2c4-4f19-b082-e8d64d47ac44\" (UID: \"a786c754-b2c4-4f19-b082-e8d64d47ac44\") " Dec 11 02:34:20 crc kubenswrapper[4824]: I1211 02:34:20.703340 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a786c754-b2c4-4f19-b082-e8d64d47ac44-kube-api-access-h9qfj" (OuterVolumeSpecName: "kube-api-access-h9qfj") pod "a786c754-b2c4-4f19-b082-e8d64d47ac44" (UID: "a786c754-b2c4-4f19-b082-e8d64d47ac44"). InnerVolumeSpecName "kube-api-access-h9qfj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:34:20 crc kubenswrapper[4824]: E1211 02:34:20.727629 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a786c754-b2c4-4f19-b082-e8d64d47ac44-inventory podName:a786c754-b2c4-4f19-b082-e8d64d47ac44 nodeName:}" failed. No retries permitted until 2025-12-11 02:34:21.227594748 +0000 UTC m=+2002.916632167 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "inventory" (UniqueName: "kubernetes.io/secret/a786c754-b2c4-4f19-b082-e8d64d47ac44-inventory") pod "a786c754-b2c4-4f19-b082-e8d64d47ac44" (UID: "a786c754-b2c4-4f19-b082-e8d64d47ac44") : error deleting /var/lib/kubelet/pods/a786c754-b2c4-4f19-b082-e8d64d47ac44/volume-subpaths: remove /var/lib/kubelet/pods/a786c754-b2c4-4f19-b082-e8d64d47ac44/volume-subpaths: no such file or directory Dec 11 02:34:20 crc kubenswrapper[4824]: I1211 02:34:20.734377 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a786c754-b2c4-4f19-b082-e8d64d47ac44-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a786c754-b2c4-4f19-b082-e8d64d47ac44" (UID: "a786c754-b2c4-4f19-b082-e8d64d47ac44"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:34:20 crc kubenswrapper[4824]: I1211 02:34:20.798582 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h9qfj\" (UniqueName: \"kubernetes.io/projected/a786c754-b2c4-4f19-b082-e8d64d47ac44-kube-api-access-h9qfj\") on node \"crc\" DevicePath \"\"" Dec 11 02:34:20 crc kubenswrapper[4824]: I1211 02:34:20.798613 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a786c754-b2c4-4f19-b082-e8d64d47ac44-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 11 02:34:21 crc kubenswrapper[4824]: I1211 02:34:21.119776 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-9k74z" event={"ID":"a786c754-b2c4-4f19-b082-e8d64d47ac44","Type":"ContainerDied","Data":"8ca9e3b655648d7c1fb8767acea84621627b63bd214a10cec1cec99ae49b98b5"} Dec 11 02:34:21 crc kubenswrapper[4824]: I1211 02:34:21.120410 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8ca9e3b655648d7c1fb8767acea84621627b63bd214a10cec1cec99ae49b98b5" Dec 11 02:34:21 crc kubenswrapper[4824]: I1211 02:34:21.119835 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-9k74z" Dec 11 02:34:21 crc kubenswrapper[4824]: I1211 02:34:21.244894 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4qkdg"] Dec 11 02:34:21 crc kubenswrapper[4824]: E1211 02:34:21.245471 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a786c754-b2c4-4f19-b082-e8d64d47ac44" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 11 02:34:21 crc kubenswrapper[4824]: I1211 02:34:21.245539 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="a786c754-b2c4-4f19-b082-e8d64d47ac44" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 11 02:34:21 crc kubenswrapper[4824]: I1211 02:34:21.245814 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="a786c754-b2c4-4f19-b082-e8d64d47ac44" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 11 02:34:21 crc kubenswrapper[4824]: I1211 02:34:21.246761 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4qkdg" Dec 11 02:34:21 crc kubenswrapper[4824]: I1211 02:34:21.257553 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4qkdg"] Dec 11 02:34:21 crc kubenswrapper[4824]: I1211 02:34:21.306727 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a786c754-b2c4-4f19-b082-e8d64d47ac44-inventory\") pod \"a786c754-b2c4-4f19-b082-e8d64d47ac44\" (UID: \"a786c754-b2c4-4f19-b082-e8d64d47ac44\") " Dec 11 02:34:21 crc kubenswrapper[4824]: I1211 02:34:21.307939 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6b2de157-739b-48d1-8bcb-585e4e6f475a-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-4qkdg\" (UID: \"6b2de157-739b-48d1-8bcb-585e4e6f475a\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4qkdg" Dec 11 02:34:21 crc kubenswrapper[4824]: I1211 02:34:21.308340 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6b2de157-739b-48d1-8bcb-585e4e6f475a-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-4qkdg\" (UID: \"6b2de157-739b-48d1-8bcb-585e4e6f475a\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4qkdg" Dec 11 02:34:21 crc kubenswrapper[4824]: I1211 02:34:21.308578 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h8vll\" (UniqueName: \"kubernetes.io/projected/6b2de157-739b-48d1-8bcb-585e4e6f475a-kube-api-access-h8vll\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-4qkdg\" (UID: \"6b2de157-739b-48d1-8bcb-585e4e6f475a\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4qkdg" Dec 11 02:34:21 crc kubenswrapper[4824]: I1211 02:34:21.312779 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a786c754-b2c4-4f19-b082-e8d64d47ac44-inventory" (OuterVolumeSpecName: "inventory") pod "a786c754-b2c4-4f19-b082-e8d64d47ac44" (UID: "a786c754-b2c4-4f19-b082-e8d64d47ac44"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:34:21 crc kubenswrapper[4824]: I1211 02:34:21.410066 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h8vll\" (UniqueName: \"kubernetes.io/projected/6b2de157-739b-48d1-8bcb-585e4e6f475a-kube-api-access-h8vll\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-4qkdg\" (UID: \"6b2de157-739b-48d1-8bcb-585e4e6f475a\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4qkdg" Dec 11 02:34:21 crc kubenswrapper[4824]: I1211 02:34:21.410166 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6b2de157-739b-48d1-8bcb-585e4e6f475a-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-4qkdg\" (UID: \"6b2de157-739b-48d1-8bcb-585e4e6f475a\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4qkdg" Dec 11 02:34:21 crc kubenswrapper[4824]: I1211 02:34:21.410272 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6b2de157-739b-48d1-8bcb-585e4e6f475a-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-4qkdg\" (UID: \"6b2de157-739b-48d1-8bcb-585e4e6f475a\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4qkdg" Dec 11 02:34:21 crc kubenswrapper[4824]: I1211 02:34:21.410349 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a786c754-b2c4-4f19-b082-e8d64d47ac44-inventory\") on node \"crc\" DevicePath \"\"" Dec 11 02:34:21 crc kubenswrapper[4824]: I1211 02:34:21.413909 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6b2de157-739b-48d1-8bcb-585e4e6f475a-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-4qkdg\" (UID: \"6b2de157-739b-48d1-8bcb-585e4e6f475a\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4qkdg" Dec 11 02:34:21 crc kubenswrapper[4824]: I1211 02:34:21.414724 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6b2de157-739b-48d1-8bcb-585e4e6f475a-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-4qkdg\" (UID: \"6b2de157-739b-48d1-8bcb-585e4e6f475a\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4qkdg" Dec 11 02:34:21 crc kubenswrapper[4824]: I1211 02:34:21.428662 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h8vll\" (UniqueName: \"kubernetes.io/projected/6b2de157-739b-48d1-8bcb-585e4e6f475a-kube-api-access-h8vll\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-4qkdg\" (UID: \"6b2de157-739b-48d1-8bcb-585e4e6f475a\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4qkdg" Dec 11 02:34:21 crc kubenswrapper[4824]: I1211 02:34:21.566642 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4qkdg" Dec 11 02:34:22 crc kubenswrapper[4824]: I1211 02:34:22.121366 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4qkdg"] Dec 11 02:34:22 crc kubenswrapper[4824]: W1211 02:34:22.141486 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6b2de157_739b_48d1_8bcb_585e4e6f475a.slice/crio-ae3f1d3c2c6062c99a806b44ccce0a362f47a60da22fef4aaac8558c01b26b96 WatchSource:0}: Error finding container ae3f1d3c2c6062c99a806b44ccce0a362f47a60da22fef4aaac8558c01b26b96: Status 404 returned error can't find the container with id ae3f1d3c2c6062c99a806b44ccce0a362f47a60da22fef4aaac8558c01b26b96 Dec 11 02:34:23 crc kubenswrapper[4824]: I1211 02:34:23.141644 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4qkdg" event={"ID":"6b2de157-739b-48d1-8bcb-585e4e6f475a","Type":"ContainerStarted","Data":"b465c271eaeb7e696a011d4e003ca38b4ec65ea4816638ebdbe32b47187571a5"} Dec 11 02:34:23 crc kubenswrapper[4824]: I1211 02:34:23.141710 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4qkdg" event={"ID":"6b2de157-739b-48d1-8bcb-585e4e6f475a","Type":"ContainerStarted","Data":"ae3f1d3c2c6062c99a806b44ccce0a362f47a60da22fef4aaac8558c01b26b96"} Dec 11 02:34:23 crc kubenswrapper[4824]: I1211 02:34:23.169152 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4qkdg" podStartSLOduration=1.93905869 podStartE2EDuration="2.169133699s" podCreationTimestamp="2025-12-11 02:34:21 +0000 UTC" firstStartedPulling="2025-12-11 02:34:22.143670769 +0000 UTC m=+2003.832708148" lastFinishedPulling="2025-12-11 02:34:22.373745768 +0000 UTC m=+2004.062783157" observedRunningTime="2025-12-11 02:34:23.164063211 +0000 UTC m=+2004.853100640" watchObservedRunningTime="2025-12-11 02:34:23.169133699 +0000 UTC m=+2004.858171088" Dec 11 02:34:35 crc kubenswrapper[4824]: I1211 02:34:35.070473 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-trgbn"] Dec 11 02:34:35 crc kubenswrapper[4824]: I1211 02:34:35.086546 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-trgbn"] Dec 11 02:34:36 crc kubenswrapper[4824]: I1211 02:34:36.653105 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="767d0d22-091d-4768-8fda-f51e6bac2504" path="/var/lib/kubelet/pods/767d0d22-091d-4768-8fda-f51e6bac2504/volumes" Dec 11 02:34:37 crc kubenswrapper[4824]: I1211 02:34:37.251445 4824 patch_prober.go:28] interesting pod/machine-config-daemon-gx6xt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 02:34:37 crc kubenswrapper[4824]: I1211 02:34:37.251532 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 02:34:37 crc kubenswrapper[4824]: I1211 02:34:37.251597 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" Dec 11 02:34:37 crc kubenswrapper[4824]: I1211 02:34:37.252444 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"80aa24371216c8cd3cae76744f65819f6cda361ceaf0439f8ef7d77b23dd30ff"} pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 11 02:34:37 crc kubenswrapper[4824]: I1211 02:34:37.252580 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" containerName="machine-config-daemon" containerID="cri-o://80aa24371216c8cd3cae76744f65819f6cda361ceaf0439f8ef7d77b23dd30ff" gracePeriod=600 Dec 11 02:34:38 crc kubenswrapper[4824]: I1211 02:34:38.339585 4824 generic.go:334] "Generic (PLEG): container finished" podID="44b156e4-64a4-4d45-aa5e-9b10a862faed" containerID="80aa24371216c8cd3cae76744f65819f6cda361ceaf0439f8ef7d77b23dd30ff" exitCode=0 Dec 11 02:34:38 crc kubenswrapper[4824]: I1211 02:34:38.339715 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" event={"ID":"44b156e4-64a4-4d45-aa5e-9b10a862faed","Type":"ContainerDied","Data":"80aa24371216c8cd3cae76744f65819f6cda361ceaf0439f8ef7d77b23dd30ff"} Dec 11 02:34:38 crc kubenswrapper[4824]: I1211 02:34:38.340311 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" event={"ID":"44b156e4-64a4-4d45-aa5e-9b10a862faed","Type":"ContainerStarted","Data":"592f00198a08d7f30f2bd4761177c27cba23107c7971ce2c5b2441d9ff2d1144"} Dec 11 02:34:38 crc kubenswrapper[4824]: I1211 02:34:38.340349 4824 scope.go:117] "RemoveContainer" containerID="1d7148bf050b9a450d49eb09a5933cb6d8dd6490c731a4e84f5c37876b3c58a6" Dec 11 02:35:04 crc kubenswrapper[4824]: I1211 02:35:04.629652 4824 scope.go:117] "RemoveContainer" containerID="ecb08a99e948f9fdfcffdae254cf531db2878ad91679263a52a6ed6e24f48c97" Dec 11 02:35:06 crc kubenswrapper[4824]: I1211 02:35:06.842094 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-jpknk"] Dec 11 02:35:06 crc kubenswrapper[4824]: I1211 02:35:06.844768 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jpknk" Dec 11 02:35:06 crc kubenswrapper[4824]: I1211 02:35:06.861311 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jpknk"] Dec 11 02:35:06 crc kubenswrapper[4824]: I1211 02:35:06.889745 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6c6c488f-e898-4d34-a737-dabd055c45ad-utilities\") pod \"redhat-marketplace-jpknk\" (UID: \"6c6c488f-e898-4d34-a737-dabd055c45ad\") " pod="openshift-marketplace/redhat-marketplace-jpknk" Dec 11 02:35:06 crc kubenswrapper[4824]: I1211 02:35:06.889894 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6c6c488f-e898-4d34-a737-dabd055c45ad-catalog-content\") pod \"redhat-marketplace-jpknk\" (UID: \"6c6c488f-e898-4d34-a737-dabd055c45ad\") " pod="openshift-marketplace/redhat-marketplace-jpknk" Dec 11 02:35:06 crc kubenswrapper[4824]: I1211 02:35:06.889927 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jzdst\" (UniqueName: \"kubernetes.io/projected/6c6c488f-e898-4d34-a737-dabd055c45ad-kube-api-access-jzdst\") pod \"redhat-marketplace-jpknk\" (UID: \"6c6c488f-e898-4d34-a737-dabd055c45ad\") " pod="openshift-marketplace/redhat-marketplace-jpknk" Dec 11 02:35:06 crc kubenswrapper[4824]: I1211 02:35:06.992208 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6c6c488f-e898-4d34-a737-dabd055c45ad-utilities\") pod \"redhat-marketplace-jpknk\" (UID: \"6c6c488f-e898-4d34-a737-dabd055c45ad\") " pod="openshift-marketplace/redhat-marketplace-jpknk" Dec 11 02:35:06 crc kubenswrapper[4824]: I1211 02:35:06.992294 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6c6c488f-e898-4d34-a737-dabd055c45ad-catalog-content\") pod \"redhat-marketplace-jpknk\" (UID: \"6c6c488f-e898-4d34-a737-dabd055c45ad\") " pod="openshift-marketplace/redhat-marketplace-jpknk" Dec 11 02:35:06 crc kubenswrapper[4824]: I1211 02:35:06.992319 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jzdst\" (UniqueName: \"kubernetes.io/projected/6c6c488f-e898-4d34-a737-dabd055c45ad-kube-api-access-jzdst\") pod \"redhat-marketplace-jpknk\" (UID: \"6c6c488f-e898-4d34-a737-dabd055c45ad\") " pod="openshift-marketplace/redhat-marketplace-jpknk" Dec 11 02:35:06 crc kubenswrapper[4824]: I1211 02:35:06.992793 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6c6c488f-e898-4d34-a737-dabd055c45ad-utilities\") pod \"redhat-marketplace-jpknk\" (UID: \"6c6c488f-e898-4d34-a737-dabd055c45ad\") " pod="openshift-marketplace/redhat-marketplace-jpknk" Dec 11 02:35:06 crc kubenswrapper[4824]: I1211 02:35:06.992826 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6c6c488f-e898-4d34-a737-dabd055c45ad-catalog-content\") pod \"redhat-marketplace-jpknk\" (UID: \"6c6c488f-e898-4d34-a737-dabd055c45ad\") " pod="openshift-marketplace/redhat-marketplace-jpknk" Dec 11 02:35:07 crc kubenswrapper[4824]: I1211 02:35:07.025306 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jzdst\" (UniqueName: \"kubernetes.io/projected/6c6c488f-e898-4d34-a737-dabd055c45ad-kube-api-access-jzdst\") pod \"redhat-marketplace-jpknk\" (UID: \"6c6c488f-e898-4d34-a737-dabd055c45ad\") " pod="openshift-marketplace/redhat-marketplace-jpknk" Dec 11 02:35:07 crc kubenswrapper[4824]: I1211 02:35:07.170739 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jpknk" Dec 11 02:35:07 crc kubenswrapper[4824]: I1211 02:35:07.626332 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jpknk"] Dec 11 02:35:07 crc kubenswrapper[4824]: I1211 02:35:07.748182 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jpknk" event={"ID":"6c6c488f-e898-4d34-a737-dabd055c45ad","Type":"ContainerStarted","Data":"008dac7de783e7090b560f445f2afb93e2b8e9eef2d2b42678543bee18661872"} Dec 11 02:35:08 crc kubenswrapper[4824]: I1211 02:35:08.759017 4824 generic.go:334] "Generic (PLEG): container finished" podID="6c6c488f-e898-4d34-a737-dabd055c45ad" containerID="1af651361d0b3f88b99c9f75bb30db4d1296e71c76f2fc8b1a738eb48efc28a5" exitCode=0 Dec 11 02:35:08 crc kubenswrapper[4824]: I1211 02:35:08.759068 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jpknk" event={"ID":"6c6c488f-e898-4d34-a737-dabd055c45ad","Type":"ContainerDied","Data":"1af651361d0b3f88b99c9f75bb30db4d1296e71c76f2fc8b1a738eb48efc28a5"} Dec 11 02:35:08 crc kubenswrapper[4824]: I1211 02:35:08.761484 4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 11 02:35:10 crc kubenswrapper[4824]: I1211 02:35:10.776944 4824 generic.go:334] "Generic (PLEG): container finished" podID="6c6c488f-e898-4d34-a737-dabd055c45ad" containerID="544bae55ceae11b35d4707e8ad94907cb80236b7318c68ba14be503f13ae78a9" exitCode=0 Dec 11 02:35:10 crc kubenswrapper[4824]: I1211 02:35:10.777021 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jpknk" event={"ID":"6c6c488f-e898-4d34-a737-dabd055c45ad","Type":"ContainerDied","Data":"544bae55ceae11b35d4707e8ad94907cb80236b7318c68ba14be503f13ae78a9"} Dec 11 02:35:12 crc kubenswrapper[4824]: I1211 02:35:12.803392 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jpknk" event={"ID":"6c6c488f-e898-4d34-a737-dabd055c45ad","Type":"ContainerStarted","Data":"e3ff9d27bd16372cdebb79fdd78da29bb65805e458c8bfd791b47854d4fdec13"} Dec 11 02:35:12 crc kubenswrapper[4824]: I1211 02:35:12.827831 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-jpknk" podStartSLOduration=3.067614991 podStartE2EDuration="6.827809649s" podCreationTimestamp="2025-12-11 02:35:06 +0000 UTC" firstStartedPulling="2025-12-11 02:35:08.761252294 +0000 UTC m=+2050.450289673" lastFinishedPulling="2025-12-11 02:35:12.521446912 +0000 UTC m=+2054.210484331" observedRunningTime="2025-12-11 02:35:12.821752994 +0000 UTC m=+2054.510790373" watchObservedRunningTime="2025-12-11 02:35:12.827809649 +0000 UTC m=+2054.516847028" Dec 11 02:35:17 crc kubenswrapper[4824]: I1211 02:35:17.172963 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-jpknk" Dec 11 02:35:17 crc kubenswrapper[4824]: I1211 02:35:17.173674 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-jpknk" Dec 11 02:35:17 crc kubenswrapper[4824]: I1211 02:35:17.246891 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-jpknk" Dec 11 02:35:17 crc kubenswrapper[4824]: I1211 02:35:17.941685 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-jpknk" Dec 11 02:35:18 crc kubenswrapper[4824]: I1211 02:35:18.002631 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jpknk"] Dec 11 02:35:18 crc kubenswrapper[4824]: I1211 02:35:18.886516 4824 generic.go:334] "Generic (PLEG): container finished" podID="6b2de157-739b-48d1-8bcb-585e4e6f475a" containerID="b465c271eaeb7e696a011d4e003ca38b4ec65ea4816638ebdbe32b47187571a5" exitCode=0 Dec 11 02:35:18 crc kubenswrapper[4824]: I1211 02:35:18.886582 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4qkdg" event={"ID":"6b2de157-739b-48d1-8bcb-585e4e6f475a","Type":"ContainerDied","Data":"b465c271eaeb7e696a011d4e003ca38b4ec65ea4816638ebdbe32b47187571a5"} Dec 11 02:35:19 crc kubenswrapper[4824]: I1211 02:35:19.893688 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-jpknk" podUID="6c6c488f-e898-4d34-a737-dabd055c45ad" containerName="registry-server" containerID="cri-o://e3ff9d27bd16372cdebb79fdd78da29bb65805e458c8bfd791b47854d4fdec13" gracePeriod=2 Dec 11 02:35:20 crc kubenswrapper[4824]: I1211 02:35:20.411102 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jpknk" Dec 11 02:35:20 crc kubenswrapper[4824]: I1211 02:35:20.417945 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4qkdg" Dec 11 02:35:20 crc kubenswrapper[4824]: I1211 02:35:20.480912 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6c6c488f-e898-4d34-a737-dabd055c45ad-utilities\") pod \"6c6c488f-e898-4d34-a737-dabd055c45ad\" (UID: \"6c6c488f-e898-4d34-a737-dabd055c45ad\") " Dec 11 02:35:20 crc kubenswrapper[4824]: I1211 02:35:20.481055 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jzdst\" (UniqueName: \"kubernetes.io/projected/6c6c488f-e898-4d34-a737-dabd055c45ad-kube-api-access-jzdst\") pod \"6c6c488f-e898-4d34-a737-dabd055c45ad\" (UID: \"6c6c488f-e898-4d34-a737-dabd055c45ad\") " Dec 11 02:35:20 crc kubenswrapper[4824]: I1211 02:35:20.481157 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6c6c488f-e898-4d34-a737-dabd055c45ad-catalog-content\") pod \"6c6c488f-e898-4d34-a737-dabd055c45ad\" (UID: \"6c6c488f-e898-4d34-a737-dabd055c45ad\") " Dec 11 02:35:20 crc kubenswrapper[4824]: I1211 02:35:20.481213 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6b2de157-739b-48d1-8bcb-585e4e6f475a-ssh-key\") pod \"6b2de157-739b-48d1-8bcb-585e4e6f475a\" (UID: \"6b2de157-739b-48d1-8bcb-585e4e6f475a\") " Dec 11 02:35:20 crc kubenswrapper[4824]: I1211 02:35:20.481258 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h8vll\" (UniqueName: \"kubernetes.io/projected/6b2de157-739b-48d1-8bcb-585e4e6f475a-kube-api-access-h8vll\") pod \"6b2de157-739b-48d1-8bcb-585e4e6f475a\" (UID: \"6b2de157-739b-48d1-8bcb-585e4e6f475a\") " Dec 11 02:35:20 crc kubenswrapper[4824]: I1211 02:35:20.481410 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6b2de157-739b-48d1-8bcb-585e4e6f475a-inventory\") pod \"6b2de157-739b-48d1-8bcb-585e4e6f475a\" (UID: \"6b2de157-739b-48d1-8bcb-585e4e6f475a\") " Dec 11 02:35:20 crc kubenswrapper[4824]: I1211 02:35:20.487202 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6c6c488f-e898-4d34-a737-dabd055c45ad-utilities" (OuterVolumeSpecName: "utilities") pod "6c6c488f-e898-4d34-a737-dabd055c45ad" (UID: "6c6c488f-e898-4d34-a737-dabd055c45ad"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:35:20 crc kubenswrapper[4824]: I1211 02:35:20.489681 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c6c488f-e898-4d34-a737-dabd055c45ad-kube-api-access-jzdst" (OuterVolumeSpecName: "kube-api-access-jzdst") pod "6c6c488f-e898-4d34-a737-dabd055c45ad" (UID: "6c6c488f-e898-4d34-a737-dabd055c45ad"). InnerVolumeSpecName "kube-api-access-jzdst". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:35:20 crc kubenswrapper[4824]: I1211 02:35:20.491861 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b2de157-739b-48d1-8bcb-585e4e6f475a-kube-api-access-h8vll" (OuterVolumeSpecName: "kube-api-access-h8vll") pod "6b2de157-739b-48d1-8bcb-585e4e6f475a" (UID: "6b2de157-739b-48d1-8bcb-585e4e6f475a"). InnerVolumeSpecName "kube-api-access-h8vll". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:35:20 crc kubenswrapper[4824]: I1211 02:35:20.509622 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6c6c488f-e898-4d34-a737-dabd055c45ad-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6c6c488f-e898-4d34-a737-dabd055c45ad" (UID: "6c6c488f-e898-4d34-a737-dabd055c45ad"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:35:20 crc kubenswrapper[4824]: I1211 02:35:20.514537 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b2de157-739b-48d1-8bcb-585e4e6f475a-inventory" (OuterVolumeSpecName: "inventory") pod "6b2de157-739b-48d1-8bcb-585e4e6f475a" (UID: "6b2de157-739b-48d1-8bcb-585e4e6f475a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:35:20 crc kubenswrapper[4824]: I1211 02:35:20.519784 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b2de157-739b-48d1-8bcb-585e4e6f475a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "6b2de157-739b-48d1-8bcb-585e4e6f475a" (UID: "6b2de157-739b-48d1-8bcb-585e4e6f475a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:35:20 crc kubenswrapper[4824]: I1211 02:35:20.584326 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6c6c488f-e898-4d34-a737-dabd055c45ad-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 02:35:20 crc kubenswrapper[4824]: I1211 02:35:20.584368 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6b2de157-739b-48d1-8bcb-585e4e6f475a-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 11 02:35:20 crc kubenswrapper[4824]: I1211 02:35:20.584382 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h8vll\" (UniqueName: \"kubernetes.io/projected/6b2de157-739b-48d1-8bcb-585e4e6f475a-kube-api-access-h8vll\") on node \"crc\" DevicePath \"\"" Dec 11 02:35:20 crc kubenswrapper[4824]: I1211 02:35:20.584396 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6b2de157-739b-48d1-8bcb-585e4e6f475a-inventory\") on node \"crc\" DevicePath \"\"" Dec 11 02:35:20 crc kubenswrapper[4824]: I1211 02:35:20.584406 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6c6c488f-e898-4d34-a737-dabd055c45ad-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 02:35:20 crc kubenswrapper[4824]: I1211 02:35:20.584415 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jzdst\" (UniqueName: \"kubernetes.io/projected/6c6c488f-e898-4d34-a737-dabd055c45ad-kube-api-access-jzdst\") on node \"crc\" DevicePath \"\"" Dec 11 02:35:20 crc kubenswrapper[4824]: I1211 02:35:20.909449 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4qkdg" event={"ID":"6b2de157-739b-48d1-8bcb-585e4e6f475a","Type":"ContainerDied","Data":"ae3f1d3c2c6062c99a806b44ccce0a362f47a60da22fef4aaac8558c01b26b96"} Dec 11 02:35:20 crc kubenswrapper[4824]: I1211 02:35:20.909514 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4qkdg" Dec 11 02:35:20 crc kubenswrapper[4824]: I1211 02:35:20.909548 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ae3f1d3c2c6062c99a806b44ccce0a362f47a60da22fef4aaac8558c01b26b96" Dec 11 02:35:20 crc kubenswrapper[4824]: I1211 02:35:20.914473 4824 generic.go:334] "Generic (PLEG): container finished" podID="6c6c488f-e898-4d34-a737-dabd055c45ad" containerID="e3ff9d27bd16372cdebb79fdd78da29bb65805e458c8bfd791b47854d4fdec13" exitCode=0 Dec 11 02:35:20 crc kubenswrapper[4824]: I1211 02:35:20.914550 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jpknk" event={"ID":"6c6c488f-e898-4d34-a737-dabd055c45ad","Type":"ContainerDied","Data":"e3ff9d27bd16372cdebb79fdd78da29bb65805e458c8bfd791b47854d4fdec13"} Dec 11 02:35:20 crc kubenswrapper[4824]: I1211 02:35:20.914606 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jpknk" event={"ID":"6c6c488f-e898-4d34-a737-dabd055c45ad","Type":"ContainerDied","Data":"008dac7de783e7090b560f445f2afb93e2b8e9eef2d2b42678543bee18661872"} Dec 11 02:35:20 crc kubenswrapper[4824]: I1211 02:35:20.914637 4824 scope.go:117] "RemoveContainer" containerID="e3ff9d27bd16372cdebb79fdd78da29bb65805e458c8bfd791b47854d4fdec13" Dec 11 02:35:20 crc kubenswrapper[4824]: I1211 02:35:20.914702 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jpknk" Dec 11 02:35:20 crc kubenswrapper[4824]: I1211 02:35:20.985585 4824 scope.go:117] "RemoveContainer" containerID="544bae55ceae11b35d4707e8ad94907cb80236b7318c68ba14be503f13ae78a9" Dec 11 02:35:20 crc kubenswrapper[4824]: I1211 02:35:20.995747 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jpknk"] Dec 11 02:35:21 crc kubenswrapper[4824]: I1211 02:35:21.015317 4824 scope.go:117] "RemoveContainer" containerID="1af651361d0b3f88b99c9f75bb30db4d1296e71c76f2fc8b1a738eb48efc28a5" Dec 11 02:35:21 crc kubenswrapper[4824]: I1211 02:35:21.038967 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-jpknk"] Dec 11 02:35:21 crc kubenswrapper[4824]: I1211 02:35:21.047967 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-pkkbb"] Dec 11 02:35:21 crc kubenswrapper[4824]: E1211 02:35:21.048823 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c6c488f-e898-4d34-a737-dabd055c45ad" containerName="extract-utilities" Dec 11 02:35:21 crc kubenswrapper[4824]: I1211 02:35:21.048844 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c6c488f-e898-4d34-a737-dabd055c45ad" containerName="extract-utilities" Dec 11 02:35:21 crc kubenswrapper[4824]: E1211 02:35:21.048856 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c6c488f-e898-4d34-a737-dabd055c45ad" containerName="extract-content" Dec 11 02:35:21 crc kubenswrapper[4824]: I1211 02:35:21.048863 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c6c488f-e898-4d34-a737-dabd055c45ad" containerName="extract-content" Dec 11 02:35:21 crc kubenswrapper[4824]: E1211 02:35:21.048875 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b2de157-739b-48d1-8bcb-585e4e6f475a" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 11 02:35:21 crc kubenswrapper[4824]: I1211 02:35:21.048885 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b2de157-739b-48d1-8bcb-585e4e6f475a" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 11 02:35:21 crc kubenswrapper[4824]: E1211 02:35:21.048905 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c6c488f-e898-4d34-a737-dabd055c45ad" containerName="registry-server" Dec 11 02:35:21 crc kubenswrapper[4824]: I1211 02:35:21.048910 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c6c488f-e898-4d34-a737-dabd055c45ad" containerName="registry-server" Dec 11 02:35:21 crc kubenswrapper[4824]: I1211 02:35:21.049082 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b2de157-739b-48d1-8bcb-585e4e6f475a" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 11 02:35:21 crc kubenswrapper[4824]: I1211 02:35:21.049094 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c6c488f-e898-4d34-a737-dabd055c45ad" containerName="registry-server" Dec 11 02:35:21 crc kubenswrapper[4824]: I1211 02:35:21.053506 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-pkkbb" Dec 11 02:35:21 crc kubenswrapper[4824]: I1211 02:35:21.056148 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 11 02:35:21 crc kubenswrapper[4824]: I1211 02:35:21.056587 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 11 02:35:21 crc kubenswrapper[4824]: I1211 02:35:21.056721 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 11 02:35:21 crc kubenswrapper[4824]: I1211 02:35:21.056665 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-hs2fv" Dec 11 02:35:21 crc kubenswrapper[4824]: I1211 02:35:21.080744 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-pkkbb"] Dec 11 02:35:21 crc kubenswrapper[4824]: I1211 02:35:21.097161 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/bf0de509-4023-4c82-b39a-f42d78d535a7-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-pkkbb\" (UID: \"bf0de509-4023-4c82-b39a-f42d78d535a7\") " pod="openstack/ssh-known-hosts-edpm-deployment-pkkbb" Dec 11 02:35:21 crc kubenswrapper[4824]: I1211 02:35:21.097240 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/bf0de509-4023-4c82-b39a-f42d78d535a7-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-pkkbb\" (UID: \"bf0de509-4023-4c82-b39a-f42d78d535a7\") " pod="openstack/ssh-known-hosts-edpm-deployment-pkkbb" Dec 11 02:35:21 crc kubenswrapper[4824]: I1211 02:35:21.097301 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gkl4v\" (UniqueName: \"kubernetes.io/projected/bf0de509-4023-4c82-b39a-f42d78d535a7-kube-api-access-gkl4v\") pod \"ssh-known-hosts-edpm-deployment-pkkbb\" (UID: \"bf0de509-4023-4c82-b39a-f42d78d535a7\") " pod="openstack/ssh-known-hosts-edpm-deployment-pkkbb" Dec 11 02:35:21 crc kubenswrapper[4824]: I1211 02:35:21.105336 4824 scope.go:117] "RemoveContainer" containerID="e3ff9d27bd16372cdebb79fdd78da29bb65805e458c8bfd791b47854d4fdec13" Dec 11 02:35:21 crc kubenswrapper[4824]: E1211 02:35:21.112657 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e3ff9d27bd16372cdebb79fdd78da29bb65805e458c8bfd791b47854d4fdec13\": container with ID starting with e3ff9d27bd16372cdebb79fdd78da29bb65805e458c8bfd791b47854d4fdec13 not found: ID does not exist" containerID="e3ff9d27bd16372cdebb79fdd78da29bb65805e458c8bfd791b47854d4fdec13" Dec 11 02:35:21 crc kubenswrapper[4824]: I1211 02:35:21.112734 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3ff9d27bd16372cdebb79fdd78da29bb65805e458c8bfd791b47854d4fdec13"} err="failed to get container status \"e3ff9d27bd16372cdebb79fdd78da29bb65805e458c8bfd791b47854d4fdec13\": rpc error: code = NotFound desc = could not find container \"e3ff9d27bd16372cdebb79fdd78da29bb65805e458c8bfd791b47854d4fdec13\": container with ID starting with e3ff9d27bd16372cdebb79fdd78da29bb65805e458c8bfd791b47854d4fdec13 not found: ID does not exist" Dec 11 02:35:21 crc kubenswrapper[4824]: I1211 02:35:21.112775 4824 scope.go:117] "RemoveContainer" containerID="544bae55ceae11b35d4707e8ad94907cb80236b7318c68ba14be503f13ae78a9" Dec 11 02:35:21 crc kubenswrapper[4824]: E1211 02:35:21.114935 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"544bae55ceae11b35d4707e8ad94907cb80236b7318c68ba14be503f13ae78a9\": container with ID starting with 544bae55ceae11b35d4707e8ad94907cb80236b7318c68ba14be503f13ae78a9 not found: ID does not exist" containerID="544bae55ceae11b35d4707e8ad94907cb80236b7318c68ba14be503f13ae78a9" Dec 11 02:35:21 crc kubenswrapper[4824]: I1211 02:35:21.115046 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"544bae55ceae11b35d4707e8ad94907cb80236b7318c68ba14be503f13ae78a9"} err="failed to get container status \"544bae55ceae11b35d4707e8ad94907cb80236b7318c68ba14be503f13ae78a9\": rpc error: code = NotFound desc = could not find container \"544bae55ceae11b35d4707e8ad94907cb80236b7318c68ba14be503f13ae78a9\": container with ID starting with 544bae55ceae11b35d4707e8ad94907cb80236b7318c68ba14be503f13ae78a9 not found: ID does not exist" Dec 11 02:35:21 crc kubenswrapper[4824]: I1211 02:35:21.115146 4824 scope.go:117] "RemoveContainer" containerID="1af651361d0b3f88b99c9f75bb30db4d1296e71c76f2fc8b1a738eb48efc28a5" Dec 11 02:35:21 crc kubenswrapper[4824]: E1211 02:35:21.116922 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1af651361d0b3f88b99c9f75bb30db4d1296e71c76f2fc8b1a738eb48efc28a5\": container with ID starting with 1af651361d0b3f88b99c9f75bb30db4d1296e71c76f2fc8b1a738eb48efc28a5 not found: ID does not exist" containerID="1af651361d0b3f88b99c9f75bb30db4d1296e71c76f2fc8b1a738eb48efc28a5" Dec 11 02:35:21 crc kubenswrapper[4824]: I1211 02:35:21.116984 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1af651361d0b3f88b99c9f75bb30db4d1296e71c76f2fc8b1a738eb48efc28a5"} err="failed to get container status \"1af651361d0b3f88b99c9f75bb30db4d1296e71c76f2fc8b1a738eb48efc28a5\": rpc error: code = NotFound desc = could not find container \"1af651361d0b3f88b99c9f75bb30db4d1296e71c76f2fc8b1a738eb48efc28a5\": container with ID starting with 1af651361d0b3f88b99c9f75bb30db4d1296e71c76f2fc8b1a738eb48efc28a5 not found: ID does not exist" Dec 11 02:35:21 crc kubenswrapper[4824]: I1211 02:35:21.198350 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/bf0de509-4023-4c82-b39a-f42d78d535a7-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-pkkbb\" (UID: \"bf0de509-4023-4c82-b39a-f42d78d535a7\") " pod="openstack/ssh-known-hosts-edpm-deployment-pkkbb" Dec 11 02:35:21 crc kubenswrapper[4824]: I1211 02:35:21.198403 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/bf0de509-4023-4c82-b39a-f42d78d535a7-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-pkkbb\" (UID: \"bf0de509-4023-4c82-b39a-f42d78d535a7\") " pod="openstack/ssh-known-hosts-edpm-deployment-pkkbb" Dec 11 02:35:21 crc kubenswrapper[4824]: I1211 02:35:21.198446 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gkl4v\" (UniqueName: \"kubernetes.io/projected/bf0de509-4023-4c82-b39a-f42d78d535a7-kube-api-access-gkl4v\") pod \"ssh-known-hosts-edpm-deployment-pkkbb\" (UID: \"bf0de509-4023-4c82-b39a-f42d78d535a7\") " pod="openstack/ssh-known-hosts-edpm-deployment-pkkbb" Dec 11 02:35:21 crc kubenswrapper[4824]: I1211 02:35:21.203801 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/bf0de509-4023-4c82-b39a-f42d78d535a7-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-pkkbb\" (UID: \"bf0de509-4023-4c82-b39a-f42d78d535a7\") " pod="openstack/ssh-known-hosts-edpm-deployment-pkkbb" Dec 11 02:35:21 crc kubenswrapper[4824]: I1211 02:35:21.204946 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/bf0de509-4023-4c82-b39a-f42d78d535a7-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-pkkbb\" (UID: \"bf0de509-4023-4c82-b39a-f42d78d535a7\") " pod="openstack/ssh-known-hosts-edpm-deployment-pkkbb" Dec 11 02:35:21 crc kubenswrapper[4824]: I1211 02:35:21.225285 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gkl4v\" (UniqueName: \"kubernetes.io/projected/bf0de509-4023-4c82-b39a-f42d78d535a7-kube-api-access-gkl4v\") pod \"ssh-known-hosts-edpm-deployment-pkkbb\" (UID: \"bf0de509-4023-4c82-b39a-f42d78d535a7\") " pod="openstack/ssh-known-hosts-edpm-deployment-pkkbb" Dec 11 02:35:21 crc kubenswrapper[4824]: I1211 02:35:21.438505 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-pkkbb" Dec 11 02:35:22 crc kubenswrapper[4824]: I1211 02:35:22.008783 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-pkkbb"] Dec 11 02:35:22 crc kubenswrapper[4824]: W1211 02:35:22.023530 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbf0de509_4023_4c82_b39a_f42d78d535a7.slice/crio-6f2b1ec3d140138f9eae5545b802d1b6efa76ee754eb7b4fde8d025f441da1a9 WatchSource:0}: Error finding container 6f2b1ec3d140138f9eae5545b802d1b6efa76ee754eb7b4fde8d025f441da1a9: Status 404 returned error can't find the container with id 6f2b1ec3d140138f9eae5545b802d1b6efa76ee754eb7b4fde8d025f441da1a9 Dec 11 02:35:22 crc kubenswrapper[4824]: I1211 02:35:22.650187 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6c6c488f-e898-4d34-a737-dabd055c45ad" path="/var/lib/kubelet/pods/6c6c488f-e898-4d34-a737-dabd055c45ad/volumes" Dec 11 02:35:22 crc kubenswrapper[4824]: I1211 02:35:22.937374 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-pkkbb" event={"ID":"bf0de509-4023-4c82-b39a-f42d78d535a7","Type":"ContainerStarted","Data":"3a4c06621652045693d4305ed62a441e581f462296b132a849be141b221f9381"} Dec 11 02:35:22 crc kubenswrapper[4824]: I1211 02:35:22.937435 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-pkkbb" event={"ID":"bf0de509-4023-4c82-b39a-f42d78d535a7","Type":"ContainerStarted","Data":"6f2b1ec3d140138f9eae5545b802d1b6efa76ee754eb7b4fde8d025f441da1a9"} Dec 11 02:35:22 crc kubenswrapper[4824]: I1211 02:35:22.963983 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-pkkbb" podStartSLOduration=1.808541398 podStartE2EDuration="1.963957672s" podCreationTimestamp="2025-12-11 02:35:21 +0000 UTC" firstStartedPulling="2025-12-11 02:35:22.02646004 +0000 UTC m=+2063.715497439" lastFinishedPulling="2025-12-11 02:35:22.181876334 +0000 UTC m=+2063.870913713" observedRunningTime="2025-12-11 02:35:22.963739856 +0000 UTC m=+2064.652777245" watchObservedRunningTime="2025-12-11 02:35:22.963957672 +0000 UTC m=+2064.652995061" Dec 11 02:35:30 crc kubenswrapper[4824]: I1211 02:35:30.009884 4824 generic.go:334] "Generic (PLEG): container finished" podID="bf0de509-4023-4c82-b39a-f42d78d535a7" containerID="3a4c06621652045693d4305ed62a441e581f462296b132a849be141b221f9381" exitCode=0 Dec 11 02:35:30 crc kubenswrapper[4824]: I1211 02:35:30.009995 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-pkkbb" event={"ID":"bf0de509-4023-4c82-b39a-f42d78d535a7","Type":"ContainerDied","Data":"3a4c06621652045693d4305ed62a441e581f462296b132a849be141b221f9381"} Dec 11 02:35:31 crc kubenswrapper[4824]: I1211 02:35:31.522814 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-pkkbb" Dec 11 02:35:31 crc kubenswrapper[4824]: I1211 02:35:31.627826 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/bf0de509-4023-4c82-b39a-f42d78d535a7-ssh-key-openstack-edpm-ipam\") pod \"bf0de509-4023-4c82-b39a-f42d78d535a7\" (UID: \"bf0de509-4023-4c82-b39a-f42d78d535a7\") " Dec 11 02:35:31 crc kubenswrapper[4824]: I1211 02:35:31.628309 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gkl4v\" (UniqueName: \"kubernetes.io/projected/bf0de509-4023-4c82-b39a-f42d78d535a7-kube-api-access-gkl4v\") pod \"bf0de509-4023-4c82-b39a-f42d78d535a7\" (UID: \"bf0de509-4023-4c82-b39a-f42d78d535a7\") " Dec 11 02:35:31 crc kubenswrapper[4824]: I1211 02:35:31.628353 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/bf0de509-4023-4c82-b39a-f42d78d535a7-inventory-0\") pod \"bf0de509-4023-4c82-b39a-f42d78d535a7\" (UID: \"bf0de509-4023-4c82-b39a-f42d78d535a7\") " Dec 11 02:35:31 crc kubenswrapper[4824]: I1211 02:35:31.636890 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf0de509-4023-4c82-b39a-f42d78d535a7-kube-api-access-gkl4v" (OuterVolumeSpecName: "kube-api-access-gkl4v") pod "bf0de509-4023-4c82-b39a-f42d78d535a7" (UID: "bf0de509-4023-4c82-b39a-f42d78d535a7"). InnerVolumeSpecName "kube-api-access-gkl4v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:35:31 crc kubenswrapper[4824]: I1211 02:35:31.666188 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf0de509-4023-4c82-b39a-f42d78d535a7-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "bf0de509-4023-4c82-b39a-f42d78d535a7" (UID: "bf0de509-4023-4c82-b39a-f42d78d535a7"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:35:31 crc kubenswrapper[4824]: I1211 02:35:31.678730 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf0de509-4023-4c82-b39a-f42d78d535a7-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "bf0de509-4023-4c82-b39a-f42d78d535a7" (UID: "bf0de509-4023-4c82-b39a-f42d78d535a7"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:35:31 crc kubenswrapper[4824]: I1211 02:35:31.731585 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/bf0de509-4023-4c82-b39a-f42d78d535a7-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 11 02:35:31 crc kubenswrapper[4824]: I1211 02:35:31.731626 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gkl4v\" (UniqueName: \"kubernetes.io/projected/bf0de509-4023-4c82-b39a-f42d78d535a7-kube-api-access-gkl4v\") on node \"crc\" DevicePath \"\"" Dec 11 02:35:31 crc kubenswrapper[4824]: I1211 02:35:31.731642 4824 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/bf0de509-4023-4c82-b39a-f42d78d535a7-inventory-0\") on node \"crc\" DevicePath \"\"" Dec 11 02:35:32 crc kubenswrapper[4824]: I1211 02:35:32.038972 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-pkkbb" event={"ID":"bf0de509-4023-4c82-b39a-f42d78d535a7","Type":"ContainerDied","Data":"6f2b1ec3d140138f9eae5545b802d1b6efa76ee754eb7b4fde8d025f441da1a9"} Dec 11 02:35:32 crc kubenswrapper[4824]: I1211 02:35:32.039045 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6f2b1ec3d140138f9eae5545b802d1b6efa76ee754eb7b4fde8d025f441da1a9" Dec 11 02:35:32 crc kubenswrapper[4824]: I1211 02:35:32.039054 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-pkkbb" Dec 11 02:35:32 crc kubenswrapper[4824]: I1211 02:35:32.140798 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-gk2qc"] Dec 11 02:35:32 crc kubenswrapper[4824]: E1211 02:35:32.141357 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf0de509-4023-4c82-b39a-f42d78d535a7" containerName="ssh-known-hosts-edpm-deployment" Dec 11 02:35:32 crc kubenswrapper[4824]: I1211 02:35:32.141378 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf0de509-4023-4c82-b39a-f42d78d535a7" containerName="ssh-known-hosts-edpm-deployment" Dec 11 02:35:32 crc kubenswrapper[4824]: I1211 02:35:32.141592 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf0de509-4023-4c82-b39a-f42d78d535a7" containerName="ssh-known-hosts-edpm-deployment" Dec 11 02:35:32 crc kubenswrapper[4824]: I1211 02:35:32.142359 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gk2qc" Dec 11 02:35:32 crc kubenswrapper[4824]: I1211 02:35:32.147806 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-hs2fv" Dec 11 02:35:32 crc kubenswrapper[4824]: I1211 02:35:32.148040 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 11 02:35:32 crc kubenswrapper[4824]: I1211 02:35:32.148261 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 11 02:35:32 crc kubenswrapper[4824]: I1211 02:35:32.148413 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 11 02:35:32 crc kubenswrapper[4824]: I1211 02:35:32.153890 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-gk2qc"] Dec 11 02:35:32 crc kubenswrapper[4824]: I1211 02:35:32.242659 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ab605768-f69e-4079-892a-a5b78e8dfc7d-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-gk2qc\" (UID: \"ab605768-f69e-4079-892a-a5b78e8dfc7d\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gk2qc" Dec 11 02:35:32 crc kubenswrapper[4824]: I1211 02:35:32.242727 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hq4dr\" (UniqueName: \"kubernetes.io/projected/ab605768-f69e-4079-892a-a5b78e8dfc7d-kube-api-access-hq4dr\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-gk2qc\" (UID: \"ab605768-f69e-4079-892a-a5b78e8dfc7d\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gk2qc" Dec 11 02:35:32 crc kubenswrapper[4824]: I1211 02:35:32.243089 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ab605768-f69e-4079-892a-a5b78e8dfc7d-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-gk2qc\" (UID: \"ab605768-f69e-4079-892a-a5b78e8dfc7d\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gk2qc" Dec 11 02:35:32 crc kubenswrapper[4824]: I1211 02:35:32.344933 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hq4dr\" (UniqueName: \"kubernetes.io/projected/ab605768-f69e-4079-892a-a5b78e8dfc7d-kube-api-access-hq4dr\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-gk2qc\" (UID: \"ab605768-f69e-4079-892a-a5b78e8dfc7d\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gk2qc" Dec 11 02:35:32 crc kubenswrapper[4824]: I1211 02:35:32.345072 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ab605768-f69e-4079-892a-a5b78e8dfc7d-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-gk2qc\" (UID: \"ab605768-f69e-4079-892a-a5b78e8dfc7d\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gk2qc" Dec 11 02:35:32 crc kubenswrapper[4824]: I1211 02:35:32.345154 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ab605768-f69e-4079-892a-a5b78e8dfc7d-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-gk2qc\" (UID: \"ab605768-f69e-4079-892a-a5b78e8dfc7d\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gk2qc" Dec 11 02:35:32 crc kubenswrapper[4824]: I1211 02:35:32.349656 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ab605768-f69e-4079-892a-a5b78e8dfc7d-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-gk2qc\" (UID: \"ab605768-f69e-4079-892a-a5b78e8dfc7d\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gk2qc" Dec 11 02:35:32 crc kubenswrapper[4824]: I1211 02:35:32.350732 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ab605768-f69e-4079-892a-a5b78e8dfc7d-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-gk2qc\" (UID: \"ab605768-f69e-4079-892a-a5b78e8dfc7d\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gk2qc" Dec 11 02:35:32 crc kubenswrapper[4824]: I1211 02:35:32.365975 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hq4dr\" (UniqueName: \"kubernetes.io/projected/ab605768-f69e-4079-892a-a5b78e8dfc7d-kube-api-access-hq4dr\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-gk2qc\" (UID: \"ab605768-f69e-4079-892a-a5b78e8dfc7d\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gk2qc" Dec 11 02:35:32 crc kubenswrapper[4824]: I1211 02:35:32.481965 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gk2qc" Dec 11 02:35:33 crc kubenswrapper[4824]: I1211 02:35:33.077476 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-gk2qc"] Dec 11 02:35:34 crc kubenswrapper[4824]: I1211 02:35:34.062895 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gk2qc" event={"ID":"ab605768-f69e-4079-892a-a5b78e8dfc7d","Type":"ContainerStarted","Data":"605125edcea6a9dce155503e4ae1229d011be5876e2646e3515594abfbbe8064"} Dec 11 02:35:34 crc kubenswrapper[4824]: I1211 02:35:34.064313 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gk2qc" event={"ID":"ab605768-f69e-4079-892a-a5b78e8dfc7d","Type":"ContainerStarted","Data":"2b74099c10eb36c0636e53cc51483bddfab1aa5fc87b931849f3208a74dfd0a6"} Dec 11 02:35:34 crc kubenswrapper[4824]: I1211 02:35:34.094913 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gk2qc" podStartSLOduration=1.866881416 podStartE2EDuration="2.094894139s" podCreationTimestamp="2025-12-11 02:35:32 +0000 UTC" firstStartedPulling="2025-12-11 02:35:33.091218113 +0000 UTC m=+2074.780255492" lastFinishedPulling="2025-12-11 02:35:33.319230806 +0000 UTC m=+2075.008268215" observedRunningTime="2025-12-11 02:35:34.081056792 +0000 UTC m=+2075.770094201" watchObservedRunningTime="2025-12-11 02:35:34.094894139 +0000 UTC m=+2075.783931518" Dec 11 02:35:35 crc kubenswrapper[4824]: I1211 02:35:35.532290 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-6gr8d"] Dec 11 02:35:35 crc kubenswrapper[4824]: I1211 02:35:35.536275 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6gr8d" Dec 11 02:35:35 crc kubenswrapper[4824]: I1211 02:35:35.542516 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6gr8d"] Dec 11 02:35:35 crc kubenswrapper[4824]: I1211 02:35:35.641389 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tg2sj\" (UniqueName: \"kubernetes.io/projected/5eb46a8d-73b9-4cb2-ad01-ba94e8633979-kube-api-access-tg2sj\") pod \"certified-operators-6gr8d\" (UID: \"5eb46a8d-73b9-4cb2-ad01-ba94e8633979\") " pod="openshift-marketplace/certified-operators-6gr8d" Dec 11 02:35:35 crc kubenswrapper[4824]: I1211 02:35:35.641447 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5eb46a8d-73b9-4cb2-ad01-ba94e8633979-catalog-content\") pod \"certified-operators-6gr8d\" (UID: \"5eb46a8d-73b9-4cb2-ad01-ba94e8633979\") " pod="openshift-marketplace/certified-operators-6gr8d" Dec 11 02:35:35 crc kubenswrapper[4824]: I1211 02:35:35.641564 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5eb46a8d-73b9-4cb2-ad01-ba94e8633979-utilities\") pod \"certified-operators-6gr8d\" (UID: \"5eb46a8d-73b9-4cb2-ad01-ba94e8633979\") " pod="openshift-marketplace/certified-operators-6gr8d" Dec 11 02:35:35 crc kubenswrapper[4824]: I1211 02:35:35.743457 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tg2sj\" (UniqueName: \"kubernetes.io/projected/5eb46a8d-73b9-4cb2-ad01-ba94e8633979-kube-api-access-tg2sj\") pod \"certified-operators-6gr8d\" (UID: \"5eb46a8d-73b9-4cb2-ad01-ba94e8633979\") " pod="openshift-marketplace/certified-operators-6gr8d" Dec 11 02:35:35 crc kubenswrapper[4824]: I1211 02:35:35.743515 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5eb46a8d-73b9-4cb2-ad01-ba94e8633979-catalog-content\") pod \"certified-operators-6gr8d\" (UID: \"5eb46a8d-73b9-4cb2-ad01-ba94e8633979\") " pod="openshift-marketplace/certified-operators-6gr8d" Dec 11 02:35:35 crc kubenswrapper[4824]: I1211 02:35:35.743642 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5eb46a8d-73b9-4cb2-ad01-ba94e8633979-utilities\") pod \"certified-operators-6gr8d\" (UID: \"5eb46a8d-73b9-4cb2-ad01-ba94e8633979\") " pod="openshift-marketplace/certified-operators-6gr8d" Dec 11 02:35:35 crc kubenswrapper[4824]: I1211 02:35:35.744955 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5eb46a8d-73b9-4cb2-ad01-ba94e8633979-catalog-content\") pod \"certified-operators-6gr8d\" (UID: \"5eb46a8d-73b9-4cb2-ad01-ba94e8633979\") " pod="openshift-marketplace/certified-operators-6gr8d" Dec 11 02:35:35 crc kubenswrapper[4824]: I1211 02:35:35.744973 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5eb46a8d-73b9-4cb2-ad01-ba94e8633979-utilities\") pod \"certified-operators-6gr8d\" (UID: \"5eb46a8d-73b9-4cb2-ad01-ba94e8633979\") " pod="openshift-marketplace/certified-operators-6gr8d" Dec 11 02:35:35 crc kubenswrapper[4824]: I1211 02:35:35.765516 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tg2sj\" (UniqueName: \"kubernetes.io/projected/5eb46a8d-73b9-4cb2-ad01-ba94e8633979-kube-api-access-tg2sj\") pod \"certified-operators-6gr8d\" (UID: \"5eb46a8d-73b9-4cb2-ad01-ba94e8633979\") " pod="openshift-marketplace/certified-operators-6gr8d" Dec 11 02:35:35 crc kubenswrapper[4824]: I1211 02:35:35.878552 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6gr8d" Dec 11 02:35:36 crc kubenswrapper[4824]: I1211 02:35:36.383297 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6gr8d"] Dec 11 02:35:36 crc kubenswrapper[4824]: W1211 02:35:36.396423 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5eb46a8d_73b9_4cb2_ad01_ba94e8633979.slice/crio-2140b7559383951b7d50c51918e33fbc54c8fc7852abe89fd83e58a44acbb803 WatchSource:0}: Error finding container 2140b7559383951b7d50c51918e33fbc54c8fc7852abe89fd83e58a44acbb803: Status 404 returned error can't find the container with id 2140b7559383951b7d50c51918e33fbc54c8fc7852abe89fd83e58a44acbb803 Dec 11 02:35:37 crc kubenswrapper[4824]: I1211 02:35:37.112928 4824 generic.go:334] "Generic (PLEG): container finished" podID="5eb46a8d-73b9-4cb2-ad01-ba94e8633979" containerID="1255a7f9ad143a632e177460de9e1fefdb79706ee2f6bda18d1b486cc488f642" exitCode=0 Dec 11 02:35:37 crc kubenswrapper[4824]: I1211 02:35:37.113189 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6gr8d" event={"ID":"5eb46a8d-73b9-4cb2-ad01-ba94e8633979","Type":"ContainerDied","Data":"1255a7f9ad143a632e177460de9e1fefdb79706ee2f6bda18d1b486cc488f642"} Dec 11 02:35:37 crc kubenswrapper[4824]: I1211 02:35:37.113691 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6gr8d" event={"ID":"5eb46a8d-73b9-4cb2-ad01-ba94e8633979","Type":"ContainerStarted","Data":"2140b7559383951b7d50c51918e33fbc54c8fc7852abe89fd83e58a44acbb803"} Dec 11 02:35:38 crc kubenswrapper[4824]: I1211 02:35:38.123548 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6gr8d" event={"ID":"5eb46a8d-73b9-4cb2-ad01-ba94e8633979","Type":"ContainerStarted","Data":"0d529b189e1894e4d961c44090954debf0cfbeff0b272cbade404ceba61e97d0"} Dec 11 02:35:40 crc kubenswrapper[4824]: I1211 02:35:40.146521 4824 generic.go:334] "Generic (PLEG): container finished" podID="5eb46a8d-73b9-4cb2-ad01-ba94e8633979" containerID="0d529b189e1894e4d961c44090954debf0cfbeff0b272cbade404ceba61e97d0" exitCode=0 Dec 11 02:35:40 crc kubenswrapper[4824]: I1211 02:35:40.146606 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6gr8d" event={"ID":"5eb46a8d-73b9-4cb2-ad01-ba94e8633979","Type":"ContainerDied","Data":"0d529b189e1894e4d961c44090954debf0cfbeff0b272cbade404ceba61e97d0"} Dec 11 02:35:41 crc kubenswrapper[4824]: I1211 02:35:41.161415 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6gr8d" event={"ID":"5eb46a8d-73b9-4cb2-ad01-ba94e8633979","Type":"ContainerStarted","Data":"0bd01a2695cfa7f54f7efcc03334586b7dcbc57dda29ad5c77a859865a420ad9"} Dec 11 02:35:41 crc kubenswrapper[4824]: I1211 02:35:41.192315 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-6gr8d" podStartSLOduration=2.495625072 podStartE2EDuration="6.19229466s" podCreationTimestamp="2025-12-11 02:35:35 +0000 UTC" firstStartedPulling="2025-12-11 02:35:37.11637716 +0000 UTC m=+2078.805414569" lastFinishedPulling="2025-12-11 02:35:40.813046758 +0000 UTC m=+2082.502084157" observedRunningTime="2025-12-11 02:35:41.18677918 +0000 UTC m=+2082.875816629" watchObservedRunningTime="2025-12-11 02:35:41.19229466 +0000 UTC m=+2082.881332059" Dec 11 02:35:42 crc kubenswrapper[4824]: I1211 02:35:42.170856 4824 generic.go:334] "Generic (PLEG): container finished" podID="ab605768-f69e-4079-892a-a5b78e8dfc7d" containerID="605125edcea6a9dce155503e4ae1229d011be5876e2646e3515594abfbbe8064" exitCode=0 Dec 11 02:35:42 crc kubenswrapper[4824]: I1211 02:35:42.170913 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gk2qc" event={"ID":"ab605768-f69e-4079-892a-a5b78e8dfc7d","Type":"ContainerDied","Data":"605125edcea6a9dce155503e4ae1229d011be5876e2646e3515594abfbbe8064"} Dec 11 02:35:43 crc kubenswrapper[4824]: I1211 02:35:43.712461 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gk2qc" Dec 11 02:35:43 crc kubenswrapper[4824]: I1211 02:35:43.851779 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hq4dr\" (UniqueName: \"kubernetes.io/projected/ab605768-f69e-4079-892a-a5b78e8dfc7d-kube-api-access-hq4dr\") pod \"ab605768-f69e-4079-892a-a5b78e8dfc7d\" (UID: \"ab605768-f69e-4079-892a-a5b78e8dfc7d\") " Dec 11 02:35:43 crc kubenswrapper[4824]: I1211 02:35:43.852052 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ab605768-f69e-4079-892a-a5b78e8dfc7d-ssh-key\") pod \"ab605768-f69e-4079-892a-a5b78e8dfc7d\" (UID: \"ab605768-f69e-4079-892a-a5b78e8dfc7d\") " Dec 11 02:35:43 crc kubenswrapper[4824]: I1211 02:35:43.852140 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ab605768-f69e-4079-892a-a5b78e8dfc7d-inventory\") pod \"ab605768-f69e-4079-892a-a5b78e8dfc7d\" (UID: \"ab605768-f69e-4079-892a-a5b78e8dfc7d\") " Dec 11 02:35:43 crc kubenswrapper[4824]: I1211 02:35:43.857558 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab605768-f69e-4079-892a-a5b78e8dfc7d-kube-api-access-hq4dr" (OuterVolumeSpecName: "kube-api-access-hq4dr") pod "ab605768-f69e-4079-892a-a5b78e8dfc7d" (UID: "ab605768-f69e-4079-892a-a5b78e8dfc7d"). InnerVolumeSpecName "kube-api-access-hq4dr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:35:43 crc kubenswrapper[4824]: I1211 02:35:43.889861 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab605768-f69e-4079-892a-a5b78e8dfc7d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ab605768-f69e-4079-892a-a5b78e8dfc7d" (UID: "ab605768-f69e-4079-892a-a5b78e8dfc7d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:35:43 crc kubenswrapper[4824]: I1211 02:35:43.900849 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab605768-f69e-4079-892a-a5b78e8dfc7d-inventory" (OuterVolumeSpecName: "inventory") pod "ab605768-f69e-4079-892a-a5b78e8dfc7d" (UID: "ab605768-f69e-4079-892a-a5b78e8dfc7d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:35:43 crc kubenswrapper[4824]: I1211 02:35:43.954672 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hq4dr\" (UniqueName: \"kubernetes.io/projected/ab605768-f69e-4079-892a-a5b78e8dfc7d-kube-api-access-hq4dr\") on node \"crc\" DevicePath \"\"" Dec 11 02:35:43 crc kubenswrapper[4824]: I1211 02:35:43.954710 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ab605768-f69e-4079-892a-a5b78e8dfc7d-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 11 02:35:43 crc kubenswrapper[4824]: I1211 02:35:43.954720 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ab605768-f69e-4079-892a-a5b78e8dfc7d-inventory\") on node \"crc\" DevicePath \"\"" Dec 11 02:35:44 crc kubenswrapper[4824]: I1211 02:35:44.198944 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gk2qc" event={"ID":"ab605768-f69e-4079-892a-a5b78e8dfc7d","Type":"ContainerDied","Data":"2b74099c10eb36c0636e53cc51483bddfab1aa5fc87b931849f3208a74dfd0a6"} Dec 11 02:35:44 crc kubenswrapper[4824]: I1211 02:35:44.199025 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2b74099c10eb36c0636e53cc51483bddfab1aa5fc87b931849f3208a74dfd0a6" Dec 11 02:35:44 crc kubenswrapper[4824]: I1211 02:35:44.199096 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gk2qc" Dec 11 02:35:44 crc kubenswrapper[4824]: I1211 02:35:44.318646 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-wx57q"] Dec 11 02:35:44 crc kubenswrapper[4824]: E1211 02:35:44.319475 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab605768-f69e-4079-892a-a5b78e8dfc7d" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 11 02:35:44 crc kubenswrapper[4824]: I1211 02:35:44.319526 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab605768-f69e-4079-892a-a5b78e8dfc7d" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 11 02:35:44 crc kubenswrapper[4824]: I1211 02:35:44.319993 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab605768-f69e-4079-892a-a5b78e8dfc7d" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 11 02:35:44 crc kubenswrapper[4824]: I1211 02:35:44.321212 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-wx57q" Dec 11 02:35:44 crc kubenswrapper[4824]: I1211 02:35:44.324642 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-hs2fv" Dec 11 02:35:44 crc kubenswrapper[4824]: I1211 02:35:44.324696 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 11 02:35:44 crc kubenswrapper[4824]: I1211 02:35:44.324717 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 11 02:35:44 crc kubenswrapper[4824]: I1211 02:35:44.324892 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 11 02:35:44 crc kubenswrapper[4824]: I1211 02:35:44.328676 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-wx57q"] Dec 11 02:35:44 crc kubenswrapper[4824]: I1211 02:35:44.463882 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6xhjh\" (UniqueName: \"kubernetes.io/projected/a24e9b38-d375-4b6b-a641-12aa6bf0fe18-kube-api-access-6xhjh\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-wx57q\" (UID: \"a24e9b38-d375-4b6b-a641-12aa6bf0fe18\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-wx57q" Dec 11 02:35:44 crc kubenswrapper[4824]: I1211 02:35:44.464305 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a24e9b38-d375-4b6b-a641-12aa6bf0fe18-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-wx57q\" (UID: \"a24e9b38-d375-4b6b-a641-12aa6bf0fe18\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-wx57q" Dec 11 02:35:44 crc kubenswrapper[4824]: I1211 02:35:44.464339 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a24e9b38-d375-4b6b-a641-12aa6bf0fe18-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-wx57q\" (UID: \"a24e9b38-d375-4b6b-a641-12aa6bf0fe18\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-wx57q" Dec 11 02:35:44 crc kubenswrapper[4824]: I1211 02:35:44.565967 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a24e9b38-d375-4b6b-a641-12aa6bf0fe18-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-wx57q\" (UID: \"a24e9b38-d375-4b6b-a641-12aa6bf0fe18\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-wx57q" Dec 11 02:35:44 crc kubenswrapper[4824]: I1211 02:35:44.566025 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a24e9b38-d375-4b6b-a641-12aa6bf0fe18-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-wx57q\" (UID: \"a24e9b38-d375-4b6b-a641-12aa6bf0fe18\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-wx57q" Dec 11 02:35:44 crc kubenswrapper[4824]: I1211 02:35:44.566103 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6xhjh\" (UniqueName: \"kubernetes.io/projected/a24e9b38-d375-4b6b-a641-12aa6bf0fe18-kube-api-access-6xhjh\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-wx57q\" (UID: \"a24e9b38-d375-4b6b-a641-12aa6bf0fe18\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-wx57q" Dec 11 02:35:44 crc kubenswrapper[4824]: I1211 02:35:44.572687 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a24e9b38-d375-4b6b-a641-12aa6bf0fe18-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-wx57q\" (UID: \"a24e9b38-d375-4b6b-a641-12aa6bf0fe18\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-wx57q" Dec 11 02:35:44 crc kubenswrapper[4824]: I1211 02:35:44.574075 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a24e9b38-d375-4b6b-a641-12aa6bf0fe18-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-wx57q\" (UID: \"a24e9b38-d375-4b6b-a641-12aa6bf0fe18\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-wx57q" Dec 11 02:35:44 crc kubenswrapper[4824]: I1211 02:35:44.591654 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6xhjh\" (UniqueName: \"kubernetes.io/projected/a24e9b38-d375-4b6b-a641-12aa6bf0fe18-kube-api-access-6xhjh\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-wx57q\" (UID: \"a24e9b38-d375-4b6b-a641-12aa6bf0fe18\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-wx57q" Dec 11 02:35:44 crc kubenswrapper[4824]: I1211 02:35:44.646733 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-wx57q" Dec 11 02:35:45 crc kubenswrapper[4824]: I1211 02:35:45.245792 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-wx57q"] Dec 11 02:35:45 crc kubenswrapper[4824]: W1211 02:35:45.254337 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda24e9b38_d375_4b6b_a641_12aa6bf0fe18.slice/crio-8984690b70f5a8ea5a607484f496aa6dd9afafdf14e4f37ab3a0d004d7b28e7a WatchSource:0}: Error finding container 8984690b70f5a8ea5a607484f496aa6dd9afafdf14e4f37ab3a0d004d7b28e7a: Status 404 returned error can't find the container with id 8984690b70f5a8ea5a607484f496aa6dd9afafdf14e4f37ab3a0d004d7b28e7a Dec 11 02:35:45 crc kubenswrapper[4824]: I1211 02:35:45.878712 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-6gr8d" Dec 11 02:35:45 crc kubenswrapper[4824]: I1211 02:35:45.880729 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-6gr8d" Dec 11 02:35:45 crc kubenswrapper[4824]: I1211 02:35:45.945192 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-6gr8d" Dec 11 02:35:46 crc kubenswrapper[4824]: I1211 02:35:46.220853 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-wx57q" event={"ID":"a24e9b38-d375-4b6b-a641-12aa6bf0fe18","Type":"ContainerStarted","Data":"46ffd5815482dd2a73f6115e0d3d8d6e70532a1217dc9afc4d8360b17f86f2dc"} Dec 11 02:35:46 crc kubenswrapper[4824]: I1211 02:35:46.220926 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-wx57q" event={"ID":"a24e9b38-d375-4b6b-a641-12aa6bf0fe18","Type":"ContainerStarted","Data":"8984690b70f5a8ea5a607484f496aa6dd9afafdf14e4f37ab3a0d004d7b28e7a"} Dec 11 02:35:46 crc kubenswrapper[4824]: I1211 02:35:46.247056 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-wx57q" podStartSLOduration=2.061668978 podStartE2EDuration="2.247025109s" podCreationTimestamp="2025-12-11 02:35:44 +0000 UTC" firstStartedPulling="2025-12-11 02:35:45.258017812 +0000 UTC m=+2086.947055191" lastFinishedPulling="2025-12-11 02:35:45.443373943 +0000 UTC m=+2087.132411322" observedRunningTime="2025-12-11 02:35:46.244543761 +0000 UTC m=+2087.933581170" watchObservedRunningTime="2025-12-11 02:35:46.247025109 +0000 UTC m=+2087.936062548" Dec 11 02:35:46 crc kubenswrapper[4824]: I1211 02:35:46.304811 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-6gr8d" Dec 11 02:35:46 crc kubenswrapper[4824]: I1211 02:35:46.376020 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6gr8d"] Dec 11 02:35:48 crc kubenswrapper[4824]: I1211 02:35:48.246436 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-6gr8d" podUID="5eb46a8d-73b9-4cb2-ad01-ba94e8633979" containerName="registry-server" containerID="cri-o://0bd01a2695cfa7f54f7efcc03334586b7dcbc57dda29ad5c77a859865a420ad9" gracePeriod=2 Dec 11 02:35:48 crc kubenswrapper[4824]: I1211 02:35:48.788215 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6gr8d" Dec 11 02:35:48 crc kubenswrapper[4824]: I1211 02:35:48.974915 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tg2sj\" (UniqueName: \"kubernetes.io/projected/5eb46a8d-73b9-4cb2-ad01-ba94e8633979-kube-api-access-tg2sj\") pod \"5eb46a8d-73b9-4cb2-ad01-ba94e8633979\" (UID: \"5eb46a8d-73b9-4cb2-ad01-ba94e8633979\") " Dec 11 02:35:48 crc kubenswrapper[4824]: I1211 02:35:48.975097 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5eb46a8d-73b9-4cb2-ad01-ba94e8633979-catalog-content\") pod \"5eb46a8d-73b9-4cb2-ad01-ba94e8633979\" (UID: \"5eb46a8d-73b9-4cb2-ad01-ba94e8633979\") " Dec 11 02:35:48 crc kubenswrapper[4824]: I1211 02:35:48.975146 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5eb46a8d-73b9-4cb2-ad01-ba94e8633979-utilities\") pod \"5eb46a8d-73b9-4cb2-ad01-ba94e8633979\" (UID: \"5eb46a8d-73b9-4cb2-ad01-ba94e8633979\") " Dec 11 02:35:48 crc kubenswrapper[4824]: I1211 02:35:48.976784 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5eb46a8d-73b9-4cb2-ad01-ba94e8633979-utilities" (OuterVolumeSpecName: "utilities") pod "5eb46a8d-73b9-4cb2-ad01-ba94e8633979" (UID: "5eb46a8d-73b9-4cb2-ad01-ba94e8633979"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:35:48 crc kubenswrapper[4824]: I1211 02:35:48.985781 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5eb46a8d-73b9-4cb2-ad01-ba94e8633979-kube-api-access-tg2sj" (OuterVolumeSpecName: "kube-api-access-tg2sj") pod "5eb46a8d-73b9-4cb2-ad01-ba94e8633979" (UID: "5eb46a8d-73b9-4cb2-ad01-ba94e8633979"). InnerVolumeSpecName "kube-api-access-tg2sj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:35:49 crc kubenswrapper[4824]: I1211 02:35:49.025162 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5eb46a8d-73b9-4cb2-ad01-ba94e8633979-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5eb46a8d-73b9-4cb2-ad01-ba94e8633979" (UID: "5eb46a8d-73b9-4cb2-ad01-ba94e8633979"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:35:49 crc kubenswrapper[4824]: I1211 02:35:49.076967 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tg2sj\" (UniqueName: \"kubernetes.io/projected/5eb46a8d-73b9-4cb2-ad01-ba94e8633979-kube-api-access-tg2sj\") on node \"crc\" DevicePath \"\"" Dec 11 02:35:49 crc kubenswrapper[4824]: I1211 02:35:49.077001 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5eb46a8d-73b9-4cb2-ad01-ba94e8633979-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 02:35:49 crc kubenswrapper[4824]: I1211 02:35:49.077009 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5eb46a8d-73b9-4cb2-ad01-ba94e8633979-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 02:35:49 crc kubenswrapper[4824]: I1211 02:35:49.282604 4824 generic.go:334] "Generic (PLEG): container finished" podID="5eb46a8d-73b9-4cb2-ad01-ba94e8633979" containerID="0bd01a2695cfa7f54f7efcc03334586b7dcbc57dda29ad5c77a859865a420ad9" exitCode=0 Dec 11 02:35:49 crc kubenswrapper[4824]: I1211 02:35:49.282697 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6gr8d" event={"ID":"5eb46a8d-73b9-4cb2-ad01-ba94e8633979","Type":"ContainerDied","Data":"0bd01a2695cfa7f54f7efcc03334586b7dcbc57dda29ad5c77a859865a420ad9"} Dec 11 02:35:49 crc kubenswrapper[4824]: I1211 02:35:49.282768 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6gr8d" event={"ID":"5eb46a8d-73b9-4cb2-ad01-ba94e8633979","Type":"ContainerDied","Data":"2140b7559383951b7d50c51918e33fbc54c8fc7852abe89fd83e58a44acbb803"} Dec 11 02:35:49 crc kubenswrapper[4824]: I1211 02:35:49.282800 4824 scope.go:117] "RemoveContainer" containerID="0bd01a2695cfa7f54f7efcc03334586b7dcbc57dda29ad5c77a859865a420ad9" Dec 11 02:35:49 crc kubenswrapper[4824]: I1211 02:35:49.282832 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6gr8d" Dec 11 02:35:49 crc kubenswrapper[4824]: I1211 02:35:49.328520 4824 scope.go:117] "RemoveContainer" containerID="0d529b189e1894e4d961c44090954debf0cfbeff0b272cbade404ceba61e97d0" Dec 11 02:35:49 crc kubenswrapper[4824]: I1211 02:35:49.341282 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6gr8d"] Dec 11 02:35:49 crc kubenswrapper[4824]: I1211 02:35:49.347676 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-6gr8d"] Dec 11 02:35:49 crc kubenswrapper[4824]: I1211 02:35:49.361368 4824 scope.go:117] "RemoveContainer" containerID="1255a7f9ad143a632e177460de9e1fefdb79706ee2f6bda18d1b486cc488f642" Dec 11 02:35:49 crc kubenswrapper[4824]: I1211 02:35:49.435257 4824 scope.go:117] "RemoveContainer" containerID="0bd01a2695cfa7f54f7efcc03334586b7dcbc57dda29ad5c77a859865a420ad9" Dec 11 02:35:49 crc kubenswrapper[4824]: E1211 02:35:49.435686 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0bd01a2695cfa7f54f7efcc03334586b7dcbc57dda29ad5c77a859865a420ad9\": container with ID starting with 0bd01a2695cfa7f54f7efcc03334586b7dcbc57dda29ad5c77a859865a420ad9 not found: ID does not exist" containerID="0bd01a2695cfa7f54f7efcc03334586b7dcbc57dda29ad5c77a859865a420ad9" Dec 11 02:35:49 crc kubenswrapper[4824]: I1211 02:35:49.435726 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0bd01a2695cfa7f54f7efcc03334586b7dcbc57dda29ad5c77a859865a420ad9"} err="failed to get container status \"0bd01a2695cfa7f54f7efcc03334586b7dcbc57dda29ad5c77a859865a420ad9\": rpc error: code = NotFound desc = could not find container \"0bd01a2695cfa7f54f7efcc03334586b7dcbc57dda29ad5c77a859865a420ad9\": container with ID starting with 0bd01a2695cfa7f54f7efcc03334586b7dcbc57dda29ad5c77a859865a420ad9 not found: ID does not exist" Dec 11 02:35:49 crc kubenswrapper[4824]: I1211 02:35:49.435748 4824 scope.go:117] "RemoveContainer" containerID="0d529b189e1894e4d961c44090954debf0cfbeff0b272cbade404ceba61e97d0" Dec 11 02:35:49 crc kubenswrapper[4824]: E1211 02:35:49.436003 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d529b189e1894e4d961c44090954debf0cfbeff0b272cbade404ceba61e97d0\": container with ID starting with 0d529b189e1894e4d961c44090954debf0cfbeff0b272cbade404ceba61e97d0 not found: ID does not exist" containerID="0d529b189e1894e4d961c44090954debf0cfbeff0b272cbade404ceba61e97d0" Dec 11 02:35:49 crc kubenswrapper[4824]: I1211 02:35:49.436037 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d529b189e1894e4d961c44090954debf0cfbeff0b272cbade404ceba61e97d0"} err="failed to get container status \"0d529b189e1894e4d961c44090954debf0cfbeff0b272cbade404ceba61e97d0\": rpc error: code = NotFound desc = could not find container \"0d529b189e1894e4d961c44090954debf0cfbeff0b272cbade404ceba61e97d0\": container with ID starting with 0d529b189e1894e4d961c44090954debf0cfbeff0b272cbade404ceba61e97d0 not found: ID does not exist" Dec 11 02:35:49 crc kubenswrapper[4824]: I1211 02:35:49.436052 4824 scope.go:117] "RemoveContainer" containerID="1255a7f9ad143a632e177460de9e1fefdb79706ee2f6bda18d1b486cc488f642" Dec 11 02:35:49 crc kubenswrapper[4824]: E1211 02:35:49.436289 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1255a7f9ad143a632e177460de9e1fefdb79706ee2f6bda18d1b486cc488f642\": container with ID starting with 1255a7f9ad143a632e177460de9e1fefdb79706ee2f6bda18d1b486cc488f642 not found: ID does not exist" containerID="1255a7f9ad143a632e177460de9e1fefdb79706ee2f6bda18d1b486cc488f642" Dec 11 02:35:49 crc kubenswrapper[4824]: I1211 02:35:49.436318 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1255a7f9ad143a632e177460de9e1fefdb79706ee2f6bda18d1b486cc488f642"} err="failed to get container status \"1255a7f9ad143a632e177460de9e1fefdb79706ee2f6bda18d1b486cc488f642\": rpc error: code = NotFound desc = could not find container \"1255a7f9ad143a632e177460de9e1fefdb79706ee2f6bda18d1b486cc488f642\": container with ID starting with 1255a7f9ad143a632e177460de9e1fefdb79706ee2f6bda18d1b486cc488f642 not found: ID does not exist" Dec 11 02:35:50 crc kubenswrapper[4824]: I1211 02:35:50.652206 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5eb46a8d-73b9-4cb2-ad01-ba94e8633979" path="/var/lib/kubelet/pods/5eb46a8d-73b9-4cb2-ad01-ba94e8633979/volumes" Dec 11 02:35:56 crc kubenswrapper[4824]: I1211 02:35:56.364094 4824 generic.go:334] "Generic (PLEG): container finished" podID="a24e9b38-d375-4b6b-a641-12aa6bf0fe18" containerID="46ffd5815482dd2a73f6115e0d3d8d6e70532a1217dc9afc4d8360b17f86f2dc" exitCode=0 Dec 11 02:35:56 crc kubenswrapper[4824]: I1211 02:35:56.364165 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-wx57q" event={"ID":"a24e9b38-d375-4b6b-a641-12aa6bf0fe18","Type":"ContainerDied","Data":"46ffd5815482dd2a73f6115e0d3d8d6e70532a1217dc9afc4d8360b17f86f2dc"} Dec 11 02:35:57 crc kubenswrapper[4824]: I1211 02:35:57.901519 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-wx57q" Dec 11 02:35:58 crc kubenswrapper[4824]: I1211 02:35:58.013481 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a24e9b38-d375-4b6b-a641-12aa6bf0fe18-ssh-key\") pod \"a24e9b38-d375-4b6b-a641-12aa6bf0fe18\" (UID: \"a24e9b38-d375-4b6b-a641-12aa6bf0fe18\") " Dec 11 02:35:58 crc kubenswrapper[4824]: I1211 02:35:58.013593 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6xhjh\" (UniqueName: \"kubernetes.io/projected/a24e9b38-d375-4b6b-a641-12aa6bf0fe18-kube-api-access-6xhjh\") pod \"a24e9b38-d375-4b6b-a641-12aa6bf0fe18\" (UID: \"a24e9b38-d375-4b6b-a641-12aa6bf0fe18\") " Dec 11 02:35:58 crc kubenswrapper[4824]: I1211 02:35:58.013635 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a24e9b38-d375-4b6b-a641-12aa6bf0fe18-inventory\") pod \"a24e9b38-d375-4b6b-a641-12aa6bf0fe18\" (UID: \"a24e9b38-d375-4b6b-a641-12aa6bf0fe18\") " Dec 11 02:35:58 crc kubenswrapper[4824]: I1211 02:35:58.027413 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a24e9b38-d375-4b6b-a641-12aa6bf0fe18-kube-api-access-6xhjh" (OuterVolumeSpecName: "kube-api-access-6xhjh") pod "a24e9b38-d375-4b6b-a641-12aa6bf0fe18" (UID: "a24e9b38-d375-4b6b-a641-12aa6bf0fe18"). InnerVolumeSpecName "kube-api-access-6xhjh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:35:58 crc kubenswrapper[4824]: I1211 02:35:58.059217 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a24e9b38-d375-4b6b-a641-12aa6bf0fe18-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a24e9b38-d375-4b6b-a641-12aa6bf0fe18" (UID: "a24e9b38-d375-4b6b-a641-12aa6bf0fe18"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:35:58 crc kubenswrapper[4824]: I1211 02:35:58.061855 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a24e9b38-d375-4b6b-a641-12aa6bf0fe18-inventory" (OuterVolumeSpecName: "inventory") pod "a24e9b38-d375-4b6b-a641-12aa6bf0fe18" (UID: "a24e9b38-d375-4b6b-a641-12aa6bf0fe18"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:35:58 crc kubenswrapper[4824]: I1211 02:35:58.115961 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a24e9b38-d375-4b6b-a641-12aa6bf0fe18-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 11 02:35:58 crc kubenswrapper[4824]: I1211 02:35:58.115991 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6xhjh\" (UniqueName: \"kubernetes.io/projected/a24e9b38-d375-4b6b-a641-12aa6bf0fe18-kube-api-access-6xhjh\") on node \"crc\" DevicePath \"\"" Dec 11 02:35:58 crc kubenswrapper[4824]: I1211 02:35:58.116001 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a24e9b38-d375-4b6b-a641-12aa6bf0fe18-inventory\") on node \"crc\" DevicePath \"\"" Dec 11 02:35:58 crc kubenswrapper[4824]: I1211 02:35:58.433876 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-wx57q" event={"ID":"a24e9b38-d375-4b6b-a641-12aa6bf0fe18","Type":"ContainerDied","Data":"8984690b70f5a8ea5a607484f496aa6dd9afafdf14e4f37ab3a0d004d7b28e7a"} Dec 11 02:35:58 crc kubenswrapper[4824]: I1211 02:35:58.433939 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8984690b70f5a8ea5a607484f496aa6dd9afafdf14e4f37ab3a0d004d7b28e7a" Dec 11 02:35:58 crc kubenswrapper[4824]: I1211 02:35:58.434083 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-wx57q" Dec 11 02:35:58 crc kubenswrapper[4824]: I1211 02:35:58.529864 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lw68c"] Dec 11 02:35:58 crc kubenswrapper[4824]: E1211 02:35:58.530261 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a24e9b38-d375-4b6b-a641-12aa6bf0fe18" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 11 02:35:58 crc kubenswrapper[4824]: I1211 02:35:58.530279 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="a24e9b38-d375-4b6b-a641-12aa6bf0fe18" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 11 02:35:58 crc kubenswrapper[4824]: E1211 02:35:58.530306 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5eb46a8d-73b9-4cb2-ad01-ba94e8633979" containerName="extract-content" Dec 11 02:35:58 crc kubenswrapper[4824]: I1211 02:35:58.530313 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="5eb46a8d-73b9-4cb2-ad01-ba94e8633979" containerName="extract-content" Dec 11 02:35:58 crc kubenswrapper[4824]: E1211 02:35:58.530321 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5eb46a8d-73b9-4cb2-ad01-ba94e8633979" containerName="extract-utilities" Dec 11 02:35:58 crc kubenswrapper[4824]: I1211 02:35:58.530327 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="5eb46a8d-73b9-4cb2-ad01-ba94e8633979" containerName="extract-utilities" Dec 11 02:35:58 crc kubenswrapper[4824]: E1211 02:35:58.530344 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5eb46a8d-73b9-4cb2-ad01-ba94e8633979" containerName="registry-server" Dec 11 02:35:58 crc kubenswrapper[4824]: I1211 02:35:58.530350 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="5eb46a8d-73b9-4cb2-ad01-ba94e8633979" containerName="registry-server" Dec 11 02:35:58 crc kubenswrapper[4824]: I1211 02:35:58.530520 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="a24e9b38-d375-4b6b-a641-12aa6bf0fe18" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 11 02:35:58 crc kubenswrapper[4824]: I1211 02:35:58.530532 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="5eb46a8d-73b9-4cb2-ad01-ba94e8633979" containerName="registry-server" Dec 11 02:35:58 crc kubenswrapper[4824]: I1211 02:35:58.531176 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lw68c" Dec 11 02:35:58 crc kubenswrapper[4824]: I1211 02:35:58.539703 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-hs2fv" Dec 11 02:35:58 crc kubenswrapper[4824]: I1211 02:35:58.539960 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Dec 11 02:35:58 crc kubenswrapper[4824]: I1211 02:35:58.540072 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 11 02:35:58 crc kubenswrapper[4824]: I1211 02:35:58.540294 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 11 02:35:58 crc kubenswrapper[4824]: I1211 02:35:58.540436 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 11 02:35:58 crc kubenswrapper[4824]: I1211 02:35:58.542129 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Dec 11 02:35:58 crc kubenswrapper[4824]: I1211 02:35:58.543558 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Dec 11 02:35:58 crc kubenswrapper[4824]: I1211 02:35:58.543735 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Dec 11 02:35:58 crc kubenswrapper[4824]: I1211 02:35:58.562042 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lw68c"] Dec 11 02:35:58 crc kubenswrapper[4824]: I1211 02:35:58.628352 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20ff7897-84b7-4c8c-b9f9-fc7721488ea7-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lw68c\" (UID: \"20ff7897-84b7-4c8c-b9f9-fc7721488ea7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lw68c" Dec 11 02:35:58 crc kubenswrapper[4824]: I1211 02:35:58.628421 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/20ff7897-84b7-4c8c-b9f9-fc7721488ea7-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lw68c\" (UID: \"20ff7897-84b7-4c8c-b9f9-fc7721488ea7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lw68c" Dec 11 02:35:58 crc kubenswrapper[4824]: I1211 02:35:58.628591 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/20ff7897-84b7-4c8c-b9f9-fc7721488ea7-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lw68c\" (UID: \"20ff7897-84b7-4c8c-b9f9-fc7721488ea7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lw68c" Dec 11 02:35:58 crc kubenswrapper[4824]: I1211 02:35:58.628648 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20ff7897-84b7-4c8c-b9f9-fc7721488ea7-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lw68c\" (UID: \"20ff7897-84b7-4c8c-b9f9-fc7721488ea7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lw68c" Dec 11 02:35:58 crc kubenswrapper[4824]: I1211 02:35:58.628709 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20ff7897-84b7-4c8c-b9f9-fc7721488ea7-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lw68c\" (UID: \"20ff7897-84b7-4c8c-b9f9-fc7721488ea7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lw68c" Dec 11 02:35:58 crc kubenswrapper[4824]: I1211 02:35:58.628762 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/20ff7897-84b7-4c8c-b9f9-fc7721488ea7-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lw68c\" (UID: \"20ff7897-84b7-4c8c-b9f9-fc7721488ea7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lw68c" Dec 11 02:35:58 crc kubenswrapper[4824]: I1211 02:35:58.628829 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20ff7897-84b7-4c8c-b9f9-fc7721488ea7-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lw68c\" (UID: \"20ff7897-84b7-4c8c-b9f9-fc7721488ea7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lw68c" Dec 11 02:35:58 crc kubenswrapper[4824]: I1211 02:35:58.628876 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/20ff7897-84b7-4c8c-b9f9-fc7721488ea7-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lw68c\" (UID: \"20ff7897-84b7-4c8c-b9f9-fc7721488ea7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lw68c" Dec 11 02:35:58 crc kubenswrapper[4824]: I1211 02:35:58.628963 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20ff7897-84b7-4c8c-b9f9-fc7721488ea7-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lw68c\" (UID: \"20ff7897-84b7-4c8c-b9f9-fc7721488ea7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lw68c" Dec 11 02:35:58 crc kubenswrapper[4824]: I1211 02:35:58.629013 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pscks\" (UniqueName: \"kubernetes.io/projected/20ff7897-84b7-4c8c-b9f9-fc7721488ea7-kube-api-access-pscks\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lw68c\" (UID: \"20ff7897-84b7-4c8c-b9f9-fc7721488ea7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lw68c" Dec 11 02:35:58 crc kubenswrapper[4824]: I1211 02:35:58.629243 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/20ff7897-84b7-4c8c-b9f9-fc7721488ea7-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lw68c\" (UID: \"20ff7897-84b7-4c8c-b9f9-fc7721488ea7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lw68c" Dec 11 02:35:58 crc kubenswrapper[4824]: I1211 02:35:58.629301 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20ff7897-84b7-4c8c-b9f9-fc7721488ea7-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lw68c\" (UID: \"20ff7897-84b7-4c8c-b9f9-fc7721488ea7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lw68c" Dec 11 02:35:58 crc kubenswrapper[4824]: I1211 02:35:58.629334 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/20ff7897-84b7-4c8c-b9f9-fc7721488ea7-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lw68c\" (UID: \"20ff7897-84b7-4c8c-b9f9-fc7721488ea7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lw68c" Dec 11 02:35:58 crc kubenswrapper[4824]: I1211 02:35:58.629375 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20ff7897-84b7-4c8c-b9f9-fc7721488ea7-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lw68c\" (UID: \"20ff7897-84b7-4c8c-b9f9-fc7721488ea7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lw68c" Dec 11 02:35:58 crc kubenswrapper[4824]: I1211 02:35:58.731064 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/20ff7897-84b7-4c8c-b9f9-fc7721488ea7-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lw68c\" (UID: \"20ff7897-84b7-4c8c-b9f9-fc7721488ea7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lw68c" Dec 11 02:35:58 crc kubenswrapper[4824]: I1211 02:35:58.731133 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20ff7897-84b7-4c8c-b9f9-fc7721488ea7-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lw68c\" (UID: \"20ff7897-84b7-4c8c-b9f9-fc7721488ea7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lw68c" Dec 11 02:35:58 crc kubenswrapper[4824]: I1211 02:35:58.731217 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/20ff7897-84b7-4c8c-b9f9-fc7721488ea7-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lw68c\" (UID: \"20ff7897-84b7-4c8c-b9f9-fc7721488ea7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lw68c" Dec 11 02:35:58 crc kubenswrapper[4824]: I1211 02:35:58.731241 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20ff7897-84b7-4c8c-b9f9-fc7721488ea7-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lw68c\" (UID: \"20ff7897-84b7-4c8c-b9f9-fc7721488ea7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lw68c" Dec 11 02:35:58 crc kubenswrapper[4824]: I1211 02:35:58.731263 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pscks\" (UniqueName: \"kubernetes.io/projected/20ff7897-84b7-4c8c-b9f9-fc7721488ea7-kube-api-access-pscks\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lw68c\" (UID: \"20ff7897-84b7-4c8c-b9f9-fc7721488ea7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lw68c" Dec 11 02:35:58 crc kubenswrapper[4824]: I1211 02:35:58.731343 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/20ff7897-84b7-4c8c-b9f9-fc7721488ea7-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lw68c\" (UID: \"20ff7897-84b7-4c8c-b9f9-fc7721488ea7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lw68c" Dec 11 02:35:58 crc kubenswrapper[4824]: I1211 02:35:58.731364 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20ff7897-84b7-4c8c-b9f9-fc7721488ea7-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lw68c\" (UID: \"20ff7897-84b7-4c8c-b9f9-fc7721488ea7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lw68c" Dec 11 02:35:58 crc kubenswrapper[4824]: I1211 02:35:58.732132 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/20ff7897-84b7-4c8c-b9f9-fc7721488ea7-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lw68c\" (UID: \"20ff7897-84b7-4c8c-b9f9-fc7721488ea7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lw68c" Dec 11 02:35:58 crc kubenswrapper[4824]: I1211 02:35:58.732182 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20ff7897-84b7-4c8c-b9f9-fc7721488ea7-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lw68c\" (UID: \"20ff7897-84b7-4c8c-b9f9-fc7721488ea7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lw68c" Dec 11 02:35:58 crc kubenswrapper[4824]: I1211 02:35:58.732209 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20ff7897-84b7-4c8c-b9f9-fc7721488ea7-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lw68c\" (UID: \"20ff7897-84b7-4c8c-b9f9-fc7721488ea7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lw68c" Dec 11 02:35:58 crc kubenswrapper[4824]: I1211 02:35:58.732238 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/20ff7897-84b7-4c8c-b9f9-fc7721488ea7-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lw68c\" (UID: \"20ff7897-84b7-4c8c-b9f9-fc7721488ea7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lw68c" Dec 11 02:35:58 crc kubenswrapper[4824]: I1211 02:35:58.732279 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/20ff7897-84b7-4c8c-b9f9-fc7721488ea7-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lw68c\" (UID: \"20ff7897-84b7-4c8c-b9f9-fc7721488ea7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lw68c" Dec 11 02:35:58 crc kubenswrapper[4824]: I1211 02:35:58.732296 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20ff7897-84b7-4c8c-b9f9-fc7721488ea7-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lw68c\" (UID: \"20ff7897-84b7-4c8c-b9f9-fc7721488ea7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lw68c" Dec 11 02:35:58 crc kubenswrapper[4824]: I1211 02:35:58.732332 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20ff7897-84b7-4c8c-b9f9-fc7721488ea7-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lw68c\" (UID: \"20ff7897-84b7-4c8c-b9f9-fc7721488ea7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lw68c" Dec 11 02:35:58 crc kubenswrapper[4824]: I1211 02:35:58.735145 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 11 02:35:58 crc kubenswrapper[4824]: I1211 02:35:58.735615 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Dec 11 02:35:58 crc kubenswrapper[4824]: I1211 02:35:58.735642 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Dec 11 02:35:58 crc kubenswrapper[4824]: I1211 02:35:58.735716 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Dec 11 02:35:58 crc kubenswrapper[4824]: I1211 02:35:58.735765 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 11 02:35:58 crc kubenswrapper[4824]: I1211 02:35:58.735770 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Dec 11 02:35:58 crc kubenswrapper[4824]: I1211 02:35:58.738787 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20ff7897-84b7-4c8c-b9f9-fc7721488ea7-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lw68c\" (UID: \"20ff7897-84b7-4c8c-b9f9-fc7721488ea7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lw68c" Dec 11 02:35:58 crc kubenswrapper[4824]: I1211 02:35:58.739273 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20ff7897-84b7-4c8c-b9f9-fc7721488ea7-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lw68c\" (UID: \"20ff7897-84b7-4c8c-b9f9-fc7721488ea7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lw68c" Dec 11 02:35:58 crc kubenswrapper[4824]: I1211 02:35:58.742141 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20ff7897-84b7-4c8c-b9f9-fc7721488ea7-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lw68c\" (UID: \"20ff7897-84b7-4c8c-b9f9-fc7721488ea7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lw68c" Dec 11 02:35:58 crc kubenswrapper[4824]: I1211 02:35:58.744230 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20ff7897-84b7-4c8c-b9f9-fc7721488ea7-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lw68c\" (UID: \"20ff7897-84b7-4c8c-b9f9-fc7721488ea7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lw68c" Dec 11 02:35:58 crc kubenswrapper[4824]: I1211 02:35:58.745125 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20ff7897-84b7-4c8c-b9f9-fc7721488ea7-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lw68c\" (UID: \"20ff7897-84b7-4c8c-b9f9-fc7721488ea7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lw68c" Dec 11 02:35:58 crc kubenswrapper[4824]: I1211 02:35:58.745301 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/20ff7897-84b7-4c8c-b9f9-fc7721488ea7-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lw68c\" (UID: \"20ff7897-84b7-4c8c-b9f9-fc7721488ea7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lw68c" Dec 11 02:35:58 crc kubenswrapper[4824]: I1211 02:35:58.745739 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/20ff7897-84b7-4c8c-b9f9-fc7721488ea7-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lw68c\" (UID: \"20ff7897-84b7-4c8c-b9f9-fc7721488ea7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lw68c" Dec 11 02:35:58 crc kubenswrapper[4824]: I1211 02:35:58.746562 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20ff7897-84b7-4c8c-b9f9-fc7721488ea7-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lw68c\" (UID: \"20ff7897-84b7-4c8c-b9f9-fc7721488ea7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lw68c" Dec 11 02:35:58 crc kubenswrapper[4824]: I1211 02:35:58.746769 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/20ff7897-84b7-4c8c-b9f9-fc7721488ea7-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lw68c\" (UID: \"20ff7897-84b7-4c8c-b9f9-fc7721488ea7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lw68c" Dec 11 02:35:58 crc kubenswrapper[4824]: I1211 02:35:58.747166 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20ff7897-84b7-4c8c-b9f9-fc7721488ea7-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lw68c\" (UID: \"20ff7897-84b7-4c8c-b9f9-fc7721488ea7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lw68c" Dec 11 02:35:58 crc kubenswrapper[4824]: I1211 02:35:58.747737 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/20ff7897-84b7-4c8c-b9f9-fc7721488ea7-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lw68c\" (UID: \"20ff7897-84b7-4c8c-b9f9-fc7721488ea7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lw68c" Dec 11 02:35:58 crc kubenswrapper[4824]: I1211 02:35:58.748726 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/20ff7897-84b7-4c8c-b9f9-fc7721488ea7-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lw68c\" (UID: \"20ff7897-84b7-4c8c-b9f9-fc7721488ea7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lw68c" Dec 11 02:35:58 crc kubenswrapper[4824]: I1211 02:35:58.752489 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pscks\" (UniqueName: \"kubernetes.io/projected/20ff7897-84b7-4c8c-b9f9-fc7721488ea7-kube-api-access-pscks\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lw68c\" (UID: \"20ff7897-84b7-4c8c-b9f9-fc7721488ea7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lw68c" Dec 11 02:35:58 crc kubenswrapper[4824]: I1211 02:35:58.754590 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/20ff7897-84b7-4c8c-b9f9-fc7721488ea7-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lw68c\" (UID: \"20ff7897-84b7-4c8c-b9f9-fc7721488ea7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lw68c" Dec 11 02:35:58 crc kubenswrapper[4824]: I1211 02:35:58.850433 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-hs2fv" Dec 11 02:35:58 crc kubenswrapper[4824]: I1211 02:35:58.858448 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lw68c" Dec 11 02:36:00 crc kubenswrapper[4824]: I1211 02:36:00.266215 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lw68c"] Dec 11 02:36:00 crc kubenswrapper[4824]: I1211 02:36:00.461686 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lw68c" event={"ID":"20ff7897-84b7-4c8c-b9f9-fc7721488ea7","Type":"ContainerStarted","Data":"753c5be8dd4a6bda7cbd89829c4f89628ff8b1bd8e1e25433cb304c5874d10e7"} Dec 11 02:36:00 crc kubenswrapper[4824]: I1211 02:36:00.507078 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 11 02:36:01 crc kubenswrapper[4824]: I1211 02:36:01.499227 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lw68c" event={"ID":"20ff7897-84b7-4c8c-b9f9-fc7721488ea7","Type":"ContainerStarted","Data":"50f6837ee8d2238d132444bc0844ba3ec2793f4a881dca1328f6ea3e15f6452c"} Dec 11 02:36:01 crc kubenswrapper[4824]: I1211 02:36:01.542335 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lw68c" podStartSLOduration=3.308923173 podStartE2EDuration="3.542313813s" podCreationTimestamp="2025-12-11 02:35:58 +0000 UTC" firstStartedPulling="2025-12-11 02:36:00.270341467 +0000 UTC m=+2101.959378846" lastFinishedPulling="2025-12-11 02:36:00.503732097 +0000 UTC m=+2102.192769486" observedRunningTime="2025-12-11 02:36:01.528490616 +0000 UTC m=+2103.217527995" watchObservedRunningTime="2025-12-11 02:36:01.542313813 +0000 UTC m=+2103.231351202" Dec 11 02:36:31 crc kubenswrapper[4824]: I1211 02:36:31.463797 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-grt6q"] Dec 11 02:36:31 crc kubenswrapper[4824]: I1211 02:36:31.466765 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-grt6q" Dec 11 02:36:31 crc kubenswrapper[4824]: I1211 02:36:31.478191 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-grt6q"] Dec 11 02:36:31 crc kubenswrapper[4824]: I1211 02:36:31.552773 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49660600-bcea-4c04-ad34-30a06d1b795c-utilities\") pod \"community-operators-grt6q\" (UID: \"49660600-bcea-4c04-ad34-30a06d1b795c\") " pod="openshift-marketplace/community-operators-grt6q" Dec 11 02:36:31 crc kubenswrapper[4824]: I1211 02:36:31.552887 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49660600-bcea-4c04-ad34-30a06d1b795c-catalog-content\") pod \"community-operators-grt6q\" (UID: \"49660600-bcea-4c04-ad34-30a06d1b795c\") " pod="openshift-marketplace/community-operators-grt6q" Dec 11 02:36:31 crc kubenswrapper[4824]: I1211 02:36:31.552991 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5q8vs\" (UniqueName: \"kubernetes.io/projected/49660600-bcea-4c04-ad34-30a06d1b795c-kube-api-access-5q8vs\") pod \"community-operators-grt6q\" (UID: \"49660600-bcea-4c04-ad34-30a06d1b795c\") " pod="openshift-marketplace/community-operators-grt6q" Dec 11 02:36:31 crc kubenswrapper[4824]: I1211 02:36:31.655313 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5q8vs\" (UniqueName: \"kubernetes.io/projected/49660600-bcea-4c04-ad34-30a06d1b795c-kube-api-access-5q8vs\") pod \"community-operators-grt6q\" (UID: \"49660600-bcea-4c04-ad34-30a06d1b795c\") " pod="openshift-marketplace/community-operators-grt6q" Dec 11 02:36:31 crc kubenswrapper[4824]: I1211 02:36:31.655543 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49660600-bcea-4c04-ad34-30a06d1b795c-utilities\") pod \"community-operators-grt6q\" (UID: \"49660600-bcea-4c04-ad34-30a06d1b795c\") " pod="openshift-marketplace/community-operators-grt6q" Dec 11 02:36:31 crc kubenswrapper[4824]: I1211 02:36:31.655966 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49660600-bcea-4c04-ad34-30a06d1b795c-utilities\") pod \"community-operators-grt6q\" (UID: \"49660600-bcea-4c04-ad34-30a06d1b795c\") " pod="openshift-marketplace/community-operators-grt6q" Dec 11 02:36:31 crc kubenswrapper[4824]: I1211 02:36:31.656077 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49660600-bcea-4c04-ad34-30a06d1b795c-catalog-content\") pod \"community-operators-grt6q\" (UID: \"49660600-bcea-4c04-ad34-30a06d1b795c\") " pod="openshift-marketplace/community-operators-grt6q" Dec 11 02:36:31 crc kubenswrapper[4824]: I1211 02:36:31.656317 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49660600-bcea-4c04-ad34-30a06d1b795c-catalog-content\") pod \"community-operators-grt6q\" (UID: \"49660600-bcea-4c04-ad34-30a06d1b795c\") " pod="openshift-marketplace/community-operators-grt6q" Dec 11 02:36:31 crc kubenswrapper[4824]: I1211 02:36:31.678104 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5q8vs\" (UniqueName: \"kubernetes.io/projected/49660600-bcea-4c04-ad34-30a06d1b795c-kube-api-access-5q8vs\") pod \"community-operators-grt6q\" (UID: \"49660600-bcea-4c04-ad34-30a06d1b795c\") " pod="openshift-marketplace/community-operators-grt6q" Dec 11 02:36:31 crc kubenswrapper[4824]: I1211 02:36:31.788739 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-grt6q" Dec 11 02:36:32 crc kubenswrapper[4824]: I1211 02:36:32.378263 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-grt6q"] Dec 11 02:36:32 crc kubenswrapper[4824]: W1211 02:36:32.380979 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod49660600_bcea_4c04_ad34_30a06d1b795c.slice/crio-cecaab8cd6ebec22e159b8ebfc8a1f87a1c135d17b1c26449b69b7b8886adcdc WatchSource:0}: Error finding container cecaab8cd6ebec22e159b8ebfc8a1f87a1c135d17b1c26449b69b7b8886adcdc: Status 404 returned error can't find the container with id cecaab8cd6ebec22e159b8ebfc8a1f87a1c135d17b1c26449b69b7b8886adcdc Dec 11 02:36:32 crc kubenswrapper[4824]: I1211 02:36:32.824230 4824 generic.go:334] "Generic (PLEG): container finished" podID="49660600-bcea-4c04-ad34-30a06d1b795c" containerID="d34d02720b57556792ddbe505c365ac4d0624cdbac38b025ea82412f079e3847" exitCode=0 Dec 11 02:36:32 crc kubenswrapper[4824]: I1211 02:36:32.824432 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-grt6q" event={"ID":"49660600-bcea-4c04-ad34-30a06d1b795c","Type":"ContainerDied","Data":"d34d02720b57556792ddbe505c365ac4d0624cdbac38b025ea82412f079e3847"} Dec 11 02:36:32 crc kubenswrapper[4824]: I1211 02:36:32.824525 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-grt6q" event={"ID":"49660600-bcea-4c04-ad34-30a06d1b795c","Type":"ContainerStarted","Data":"cecaab8cd6ebec22e159b8ebfc8a1f87a1c135d17b1c26449b69b7b8886adcdc"} Dec 11 02:36:34 crc kubenswrapper[4824]: I1211 02:36:34.847097 4824 generic.go:334] "Generic (PLEG): container finished" podID="49660600-bcea-4c04-ad34-30a06d1b795c" containerID="4371f3196f7ff6a2f27f3526afc59f8a83860c553475504b34b2d53db2c7871f" exitCode=0 Dec 11 02:36:34 crc kubenswrapper[4824]: I1211 02:36:34.847167 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-grt6q" event={"ID":"49660600-bcea-4c04-ad34-30a06d1b795c","Type":"ContainerDied","Data":"4371f3196f7ff6a2f27f3526afc59f8a83860c553475504b34b2d53db2c7871f"} Dec 11 02:36:35 crc kubenswrapper[4824]: I1211 02:36:35.862240 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-grt6q" event={"ID":"49660600-bcea-4c04-ad34-30a06d1b795c","Type":"ContainerStarted","Data":"7e6a74555dc9842c4265be4cb91de55597dff8feb027dda80ca0c88b10a6b0f6"} Dec 11 02:36:35 crc kubenswrapper[4824]: I1211 02:36:35.900432 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-grt6q" podStartSLOduration=2.33289971 podStartE2EDuration="4.900407172s" podCreationTimestamp="2025-12-11 02:36:31 +0000 UTC" firstStartedPulling="2025-12-11 02:36:32.826709188 +0000 UTC m=+2134.515746567" lastFinishedPulling="2025-12-11 02:36:35.39421664 +0000 UTC m=+2137.083254029" observedRunningTime="2025-12-11 02:36:35.887618464 +0000 UTC m=+2137.576655863" watchObservedRunningTime="2025-12-11 02:36:35.900407172 +0000 UTC m=+2137.589444591" Dec 11 02:36:37 crc kubenswrapper[4824]: I1211 02:36:37.251712 4824 patch_prober.go:28] interesting pod/machine-config-daemon-gx6xt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 02:36:37 crc kubenswrapper[4824]: I1211 02:36:37.251768 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 02:36:41 crc kubenswrapper[4824]: I1211 02:36:41.789995 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-grt6q" Dec 11 02:36:41 crc kubenswrapper[4824]: I1211 02:36:41.790795 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-grt6q" Dec 11 02:36:41 crc kubenswrapper[4824]: I1211 02:36:41.862904 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-grt6q" Dec 11 02:36:41 crc kubenswrapper[4824]: I1211 02:36:41.981169 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-grt6q" Dec 11 02:36:42 crc kubenswrapper[4824]: I1211 02:36:42.111309 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-grt6q"] Dec 11 02:36:42 crc kubenswrapper[4824]: E1211 02:36:42.604641 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod20ff7897_84b7_4c8c_b9f9_fc7721488ea7.slice/crio-50f6837ee8d2238d132444bc0844ba3ec2793f4a881dca1328f6ea3e15f6452c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod20ff7897_84b7_4c8c_b9f9_fc7721488ea7.slice/crio-conmon-50f6837ee8d2238d132444bc0844ba3ec2793f4a881dca1328f6ea3e15f6452c.scope\": RecentStats: unable to find data in memory cache]" Dec 11 02:36:42 crc kubenswrapper[4824]: I1211 02:36:42.940431 4824 generic.go:334] "Generic (PLEG): container finished" podID="20ff7897-84b7-4c8c-b9f9-fc7721488ea7" containerID="50f6837ee8d2238d132444bc0844ba3ec2793f4a881dca1328f6ea3e15f6452c" exitCode=0 Dec 11 02:36:42 crc kubenswrapper[4824]: I1211 02:36:42.940553 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lw68c" event={"ID":"20ff7897-84b7-4c8c-b9f9-fc7721488ea7","Type":"ContainerDied","Data":"50f6837ee8d2238d132444bc0844ba3ec2793f4a881dca1328f6ea3e15f6452c"} Dec 11 02:36:43 crc kubenswrapper[4824]: I1211 02:36:43.950270 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-grt6q" podUID="49660600-bcea-4c04-ad34-30a06d1b795c" containerName="registry-server" containerID="cri-o://7e6a74555dc9842c4265be4cb91de55597dff8feb027dda80ca0c88b10a6b0f6" gracePeriod=2 Dec 11 02:36:44 crc kubenswrapper[4824]: I1211 02:36:44.583464 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lw68c" Dec 11 02:36:44 crc kubenswrapper[4824]: I1211 02:36:44.610167 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-grt6q" Dec 11 02:36:44 crc kubenswrapper[4824]: I1211 02:36:44.663725 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/20ff7897-84b7-4c8c-b9f9-fc7721488ea7-inventory\") pod \"20ff7897-84b7-4c8c-b9f9-fc7721488ea7\" (UID: \"20ff7897-84b7-4c8c-b9f9-fc7721488ea7\") " Dec 11 02:36:44 crc kubenswrapper[4824]: I1211 02:36:44.663773 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20ff7897-84b7-4c8c-b9f9-fc7721488ea7-neutron-metadata-combined-ca-bundle\") pod \"20ff7897-84b7-4c8c-b9f9-fc7721488ea7\" (UID: \"20ff7897-84b7-4c8c-b9f9-fc7721488ea7\") " Dec 11 02:36:44 crc kubenswrapper[4824]: I1211 02:36:44.663811 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/20ff7897-84b7-4c8c-b9f9-fc7721488ea7-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"20ff7897-84b7-4c8c-b9f9-fc7721488ea7\" (UID: \"20ff7897-84b7-4c8c-b9f9-fc7721488ea7\") " Dec 11 02:36:44 crc kubenswrapper[4824]: I1211 02:36:44.663852 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/20ff7897-84b7-4c8c-b9f9-fc7721488ea7-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"20ff7897-84b7-4c8c-b9f9-fc7721488ea7\" (UID: \"20ff7897-84b7-4c8c-b9f9-fc7721488ea7\") " Dec 11 02:36:44 crc kubenswrapper[4824]: I1211 02:36:44.663875 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/20ff7897-84b7-4c8c-b9f9-fc7721488ea7-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"20ff7897-84b7-4c8c-b9f9-fc7721488ea7\" (UID: \"20ff7897-84b7-4c8c-b9f9-fc7721488ea7\") " Dec 11 02:36:44 crc kubenswrapper[4824]: I1211 02:36:44.663913 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pscks\" (UniqueName: \"kubernetes.io/projected/20ff7897-84b7-4c8c-b9f9-fc7721488ea7-kube-api-access-pscks\") pod \"20ff7897-84b7-4c8c-b9f9-fc7721488ea7\" (UID: \"20ff7897-84b7-4c8c-b9f9-fc7721488ea7\") " Dec 11 02:36:44 crc kubenswrapper[4824]: I1211 02:36:44.663965 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/20ff7897-84b7-4c8c-b9f9-fc7721488ea7-openstack-edpm-ipam-ovn-default-certs-0\") pod \"20ff7897-84b7-4c8c-b9f9-fc7721488ea7\" (UID: \"20ff7897-84b7-4c8c-b9f9-fc7721488ea7\") " Dec 11 02:36:44 crc kubenswrapper[4824]: I1211 02:36:44.663991 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/20ff7897-84b7-4c8c-b9f9-fc7721488ea7-ssh-key\") pod \"20ff7897-84b7-4c8c-b9f9-fc7721488ea7\" (UID: \"20ff7897-84b7-4c8c-b9f9-fc7721488ea7\") " Dec 11 02:36:44 crc kubenswrapper[4824]: I1211 02:36:44.664024 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20ff7897-84b7-4c8c-b9f9-fc7721488ea7-bootstrap-combined-ca-bundle\") pod \"20ff7897-84b7-4c8c-b9f9-fc7721488ea7\" (UID: \"20ff7897-84b7-4c8c-b9f9-fc7721488ea7\") " Dec 11 02:36:44 crc kubenswrapper[4824]: I1211 02:36:44.664102 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20ff7897-84b7-4c8c-b9f9-fc7721488ea7-libvirt-combined-ca-bundle\") pod \"20ff7897-84b7-4c8c-b9f9-fc7721488ea7\" (UID: \"20ff7897-84b7-4c8c-b9f9-fc7721488ea7\") " Dec 11 02:36:44 crc kubenswrapper[4824]: I1211 02:36:44.664159 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20ff7897-84b7-4c8c-b9f9-fc7721488ea7-nova-combined-ca-bundle\") pod \"20ff7897-84b7-4c8c-b9f9-fc7721488ea7\" (UID: \"20ff7897-84b7-4c8c-b9f9-fc7721488ea7\") " Dec 11 02:36:44 crc kubenswrapper[4824]: I1211 02:36:44.664179 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20ff7897-84b7-4c8c-b9f9-fc7721488ea7-ovn-combined-ca-bundle\") pod \"20ff7897-84b7-4c8c-b9f9-fc7721488ea7\" (UID: \"20ff7897-84b7-4c8c-b9f9-fc7721488ea7\") " Dec 11 02:36:44 crc kubenswrapper[4824]: I1211 02:36:44.664199 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20ff7897-84b7-4c8c-b9f9-fc7721488ea7-repo-setup-combined-ca-bundle\") pod \"20ff7897-84b7-4c8c-b9f9-fc7721488ea7\" (UID: \"20ff7897-84b7-4c8c-b9f9-fc7721488ea7\") " Dec 11 02:36:44 crc kubenswrapper[4824]: I1211 02:36:44.664262 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20ff7897-84b7-4c8c-b9f9-fc7721488ea7-telemetry-combined-ca-bundle\") pod \"20ff7897-84b7-4c8c-b9f9-fc7721488ea7\" (UID: \"20ff7897-84b7-4c8c-b9f9-fc7721488ea7\") " Dec 11 02:36:44 crc kubenswrapper[4824]: I1211 02:36:44.671009 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20ff7897-84b7-4c8c-b9f9-fc7721488ea7-kube-api-access-pscks" (OuterVolumeSpecName: "kube-api-access-pscks") pod "20ff7897-84b7-4c8c-b9f9-fc7721488ea7" (UID: "20ff7897-84b7-4c8c-b9f9-fc7721488ea7"). InnerVolumeSpecName "kube-api-access-pscks". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:36:44 crc kubenswrapper[4824]: I1211 02:36:44.671067 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20ff7897-84b7-4c8c-b9f9-fc7721488ea7-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "20ff7897-84b7-4c8c-b9f9-fc7721488ea7" (UID: "20ff7897-84b7-4c8c-b9f9-fc7721488ea7"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:36:44 crc kubenswrapper[4824]: I1211 02:36:44.671140 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20ff7897-84b7-4c8c-b9f9-fc7721488ea7-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "20ff7897-84b7-4c8c-b9f9-fc7721488ea7" (UID: "20ff7897-84b7-4c8c-b9f9-fc7721488ea7"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:36:44 crc kubenswrapper[4824]: I1211 02:36:44.671178 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20ff7897-84b7-4c8c-b9f9-fc7721488ea7-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "20ff7897-84b7-4c8c-b9f9-fc7721488ea7" (UID: "20ff7897-84b7-4c8c-b9f9-fc7721488ea7"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:36:44 crc kubenswrapper[4824]: I1211 02:36:44.674193 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20ff7897-84b7-4c8c-b9f9-fc7721488ea7-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "20ff7897-84b7-4c8c-b9f9-fc7721488ea7" (UID: "20ff7897-84b7-4c8c-b9f9-fc7721488ea7"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:36:44 crc kubenswrapper[4824]: I1211 02:36:44.674526 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20ff7897-84b7-4c8c-b9f9-fc7721488ea7-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "20ff7897-84b7-4c8c-b9f9-fc7721488ea7" (UID: "20ff7897-84b7-4c8c-b9f9-fc7721488ea7"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:36:44 crc kubenswrapper[4824]: I1211 02:36:44.674840 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20ff7897-84b7-4c8c-b9f9-fc7721488ea7-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "20ff7897-84b7-4c8c-b9f9-fc7721488ea7" (UID: "20ff7897-84b7-4c8c-b9f9-fc7721488ea7"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:36:44 crc kubenswrapper[4824]: I1211 02:36:44.674957 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20ff7897-84b7-4c8c-b9f9-fc7721488ea7-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "20ff7897-84b7-4c8c-b9f9-fc7721488ea7" (UID: "20ff7897-84b7-4c8c-b9f9-fc7721488ea7"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:36:44 crc kubenswrapper[4824]: I1211 02:36:44.675504 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20ff7897-84b7-4c8c-b9f9-fc7721488ea7-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "20ff7897-84b7-4c8c-b9f9-fc7721488ea7" (UID: "20ff7897-84b7-4c8c-b9f9-fc7721488ea7"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:36:44 crc kubenswrapper[4824]: I1211 02:36:44.675964 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20ff7897-84b7-4c8c-b9f9-fc7721488ea7-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "20ff7897-84b7-4c8c-b9f9-fc7721488ea7" (UID: "20ff7897-84b7-4c8c-b9f9-fc7721488ea7"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:36:44 crc kubenswrapper[4824]: I1211 02:36:44.676303 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20ff7897-84b7-4c8c-b9f9-fc7721488ea7-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "20ff7897-84b7-4c8c-b9f9-fc7721488ea7" (UID: "20ff7897-84b7-4c8c-b9f9-fc7721488ea7"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:36:44 crc kubenswrapper[4824]: I1211 02:36:44.698759 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20ff7897-84b7-4c8c-b9f9-fc7721488ea7-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "20ff7897-84b7-4c8c-b9f9-fc7721488ea7" (UID: "20ff7897-84b7-4c8c-b9f9-fc7721488ea7"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:36:44 crc kubenswrapper[4824]: I1211 02:36:44.707278 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20ff7897-84b7-4c8c-b9f9-fc7721488ea7-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "20ff7897-84b7-4c8c-b9f9-fc7721488ea7" (UID: "20ff7897-84b7-4c8c-b9f9-fc7721488ea7"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:36:44 crc kubenswrapper[4824]: I1211 02:36:44.707312 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20ff7897-84b7-4c8c-b9f9-fc7721488ea7-inventory" (OuterVolumeSpecName: "inventory") pod "20ff7897-84b7-4c8c-b9f9-fc7721488ea7" (UID: "20ff7897-84b7-4c8c-b9f9-fc7721488ea7"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:36:44 crc kubenswrapper[4824]: I1211 02:36:44.765878 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49660600-bcea-4c04-ad34-30a06d1b795c-utilities\") pod \"49660600-bcea-4c04-ad34-30a06d1b795c\" (UID: \"49660600-bcea-4c04-ad34-30a06d1b795c\") " Dec 11 02:36:44 crc kubenswrapper[4824]: I1211 02:36:44.765981 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49660600-bcea-4c04-ad34-30a06d1b795c-catalog-content\") pod \"49660600-bcea-4c04-ad34-30a06d1b795c\" (UID: \"49660600-bcea-4c04-ad34-30a06d1b795c\") " Dec 11 02:36:44 crc kubenswrapper[4824]: I1211 02:36:44.766253 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5q8vs\" (UniqueName: \"kubernetes.io/projected/49660600-bcea-4c04-ad34-30a06d1b795c-kube-api-access-5q8vs\") pod \"49660600-bcea-4c04-ad34-30a06d1b795c\" (UID: \"49660600-bcea-4c04-ad34-30a06d1b795c\") " Dec 11 02:36:44 crc kubenswrapper[4824]: I1211 02:36:44.766587 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/49660600-bcea-4c04-ad34-30a06d1b795c-utilities" (OuterVolumeSpecName: "utilities") pod "49660600-bcea-4c04-ad34-30a06d1b795c" (UID: "49660600-bcea-4c04-ad34-30a06d1b795c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:36:44 crc kubenswrapper[4824]: I1211 02:36:44.767267 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/20ff7897-84b7-4c8c-b9f9-fc7721488ea7-inventory\") on node \"crc\" DevicePath \"\"" Dec 11 02:36:44 crc kubenswrapper[4824]: I1211 02:36:44.767315 4824 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20ff7897-84b7-4c8c-b9f9-fc7721488ea7-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 02:36:44 crc kubenswrapper[4824]: I1211 02:36:44.767342 4824 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/20ff7897-84b7-4c8c-b9f9-fc7721488ea7-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 11 02:36:44 crc kubenswrapper[4824]: I1211 02:36:44.767370 4824 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/20ff7897-84b7-4c8c-b9f9-fc7721488ea7-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 11 02:36:44 crc kubenswrapper[4824]: I1211 02:36:44.767398 4824 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/20ff7897-84b7-4c8c-b9f9-fc7721488ea7-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 11 02:36:44 crc kubenswrapper[4824]: I1211 02:36:44.767426 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pscks\" (UniqueName: \"kubernetes.io/projected/20ff7897-84b7-4c8c-b9f9-fc7721488ea7-kube-api-access-pscks\") on node \"crc\" DevicePath \"\"" Dec 11 02:36:44 crc kubenswrapper[4824]: I1211 02:36:44.767457 4824 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/20ff7897-84b7-4c8c-b9f9-fc7721488ea7-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 11 02:36:44 crc kubenswrapper[4824]: I1211 02:36:44.767483 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/20ff7897-84b7-4c8c-b9f9-fc7721488ea7-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 11 02:36:44 crc kubenswrapper[4824]: I1211 02:36:44.767507 4824 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20ff7897-84b7-4c8c-b9f9-fc7721488ea7-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 02:36:44 crc kubenswrapper[4824]: I1211 02:36:44.767530 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49660600-bcea-4c04-ad34-30a06d1b795c-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 02:36:44 crc kubenswrapper[4824]: I1211 02:36:44.767556 4824 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20ff7897-84b7-4c8c-b9f9-fc7721488ea7-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 02:36:44 crc kubenswrapper[4824]: I1211 02:36:44.767578 4824 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20ff7897-84b7-4c8c-b9f9-fc7721488ea7-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 02:36:44 crc kubenswrapper[4824]: I1211 02:36:44.767602 4824 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20ff7897-84b7-4c8c-b9f9-fc7721488ea7-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 02:36:44 crc kubenswrapper[4824]: I1211 02:36:44.767627 4824 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20ff7897-84b7-4c8c-b9f9-fc7721488ea7-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 02:36:44 crc kubenswrapper[4824]: I1211 02:36:44.767650 4824 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20ff7897-84b7-4c8c-b9f9-fc7721488ea7-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 02:36:44 crc kubenswrapper[4824]: I1211 02:36:44.771308 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49660600-bcea-4c04-ad34-30a06d1b795c-kube-api-access-5q8vs" (OuterVolumeSpecName: "kube-api-access-5q8vs") pod "49660600-bcea-4c04-ad34-30a06d1b795c" (UID: "49660600-bcea-4c04-ad34-30a06d1b795c"). InnerVolumeSpecName "kube-api-access-5q8vs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:36:44 crc kubenswrapper[4824]: I1211 02:36:44.818798 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/49660600-bcea-4c04-ad34-30a06d1b795c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "49660600-bcea-4c04-ad34-30a06d1b795c" (UID: "49660600-bcea-4c04-ad34-30a06d1b795c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:36:44 crc kubenswrapper[4824]: I1211 02:36:44.870402 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5q8vs\" (UniqueName: \"kubernetes.io/projected/49660600-bcea-4c04-ad34-30a06d1b795c-kube-api-access-5q8vs\") on node \"crc\" DevicePath \"\"" Dec 11 02:36:44 crc kubenswrapper[4824]: I1211 02:36:44.870439 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49660600-bcea-4c04-ad34-30a06d1b795c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 02:36:44 crc kubenswrapper[4824]: I1211 02:36:44.964816 4824 generic.go:334] "Generic (PLEG): container finished" podID="49660600-bcea-4c04-ad34-30a06d1b795c" containerID="7e6a74555dc9842c4265be4cb91de55597dff8feb027dda80ca0c88b10a6b0f6" exitCode=0 Dec 11 02:36:44 crc kubenswrapper[4824]: I1211 02:36:44.964920 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-grt6q" Dec 11 02:36:44 crc kubenswrapper[4824]: I1211 02:36:44.964948 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-grt6q" event={"ID":"49660600-bcea-4c04-ad34-30a06d1b795c","Type":"ContainerDied","Data":"7e6a74555dc9842c4265be4cb91de55597dff8feb027dda80ca0c88b10a6b0f6"} Dec 11 02:36:44 crc kubenswrapper[4824]: I1211 02:36:44.965038 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-grt6q" event={"ID":"49660600-bcea-4c04-ad34-30a06d1b795c","Type":"ContainerDied","Data":"cecaab8cd6ebec22e159b8ebfc8a1f87a1c135d17b1c26449b69b7b8886adcdc"} Dec 11 02:36:44 crc kubenswrapper[4824]: I1211 02:36:44.965084 4824 scope.go:117] "RemoveContainer" containerID="7e6a74555dc9842c4265be4cb91de55597dff8feb027dda80ca0c88b10a6b0f6" Dec 11 02:36:44 crc kubenswrapper[4824]: I1211 02:36:44.967134 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lw68c" event={"ID":"20ff7897-84b7-4c8c-b9f9-fc7721488ea7","Type":"ContainerDied","Data":"753c5be8dd4a6bda7cbd89829c4f89628ff8b1bd8e1e25433cb304c5874d10e7"} Dec 11 02:36:44 crc kubenswrapper[4824]: I1211 02:36:44.967165 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="753c5be8dd4a6bda7cbd89829c4f89628ff8b1bd8e1e25433cb304c5874d10e7" Dec 11 02:36:44 crc kubenswrapper[4824]: I1211 02:36:44.967245 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lw68c" Dec 11 02:36:44 crc kubenswrapper[4824]: I1211 02:36:44.988383 4824 scope.go:117] "RemoveContainer" containerID="4371f3196f7ff6a2f27f3526afc59f8a83860c553475504b34b2d53db2c7871f" Dec 11 02:36:45 crc kubenswrapper[4824]: I1211 02:36:45.014993 4824 scope.go:117] "RemoveContainer" containerID="d34d02720b57556792ddbe505c365ac4d0624cdbac38b025ea82412f079e3847" Dec 11 02:36:45 crc kubenswrapper[4824]: I1211 02:36:45.057650 4824 scope.go:117] "RemoveContainer" containerID="7e6a74555dc9842c4265be4cb91de55597dff8feb027dda80ca0c88b10a6b0f6" Dec 11 02:36:45 crc kubenswrapper[4824]: E1211 02:36:45.058123 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7e6a74555dc9842c4265be4cb91de55597dff8feb027dda80ca0c88b10a6b0f6\": container with ID starting with 7e6a74555dc9842c4265be4cb91de55597dff8feb027dda80ca0c88b10a6b0f6 not found: ID does not exist" containerID="7e6a74555dc9842c4265be4cb91de55597dff8feb027dda80ca0c88b10a6b0f6" Dec 11 02:36:45 crc kubenswrapper[4824]: I1211 02:36:45.058165 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e6a74555dc9842c4265be4cb91de55597dff8feb027dda80ca0c88b10a6b0f6"} err="failed to get container status \"7e6a74555dc9842c4265be4cb91de55597dff8feb027dda80ca0c88b10a6b0f6\": rpc error: code = NotFound desc = could not find container \"7e6a74555dc9842c4265be4cb91de55597dff8feb027dda80ca0c88b10a6b0f6\": container with ID starting with 7e6a74555dc9842c4265be4cb91de55597dff8feb027dda80ca0c88b10a6b0f6 not found: ID does not exist" Dec 11 02:36:45 crc kubenswrapper[4824]: I1211 02:36:45.058185 4824 scope.go:117] "RemoveContainer" containerID="4371f3196f7ff6a2f27f3526afc59f8a83860c553475504b34b2d53db2c7871f" Dec 11 02:36:45 crc kubenswrapper[4824]: E1211 02:36:45.058474 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4371f3196f7ff6a2f27f3526afc59f8a83860c553475504b34b2d53db2c7871f\": container with ID starting with 4371f3196f7ff6a2f27f3526afc59f8a83860c553475504b34b2d53db2c7871f not found: ID does not exist" containerID="4371f3196f7ff6a2f27f3526afc59f8a83860c553475504b34b2d53db2c7871f" Dec 11 02:36:45 crc kubenswrapper[4824]: I1211 02:36:45.058542 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4371f3196f7ff6a2f27f3526afc59f8a83860c553475504b34b2d53db2c7871f"} err="failed to get container status \"4371f3196f7ff6a2f27f3526afc59f8a83860c553475504b34b2d53db2c7871f\": rpc error: code = NotFound desc = could not find container \"4371f3196f7ff6a2f27f3526afc59f8a83860c553475504b34b2d53db2c7871f\": container with ID starting with 4371f3196f7ff6a2f27f3526afc59f8a83860c553475504b34b2d53db2c7871f not found: ID does not exist" Dec 11 02:36:45 crc kubenswrapper[4824]: I1211 02:36:45.058589 4824 scope.go:117] "RemoveContainer" containerID="d34d02720b57556792ddbe505c365ac4d0624cdbac38b025ea82412f079e3847" Dec 11 02:36:45 crc kubenswrapper[4824]: E1211 02:36:45.059425 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d34d02720b57556792ddbe505c365ac4d0624cdbac38b025ea82412f079e3847\": container with ID starting with d34d02720b57556792ddbe505c365ac4d0624cdbac38b025ea82412f079e3847 not found: ID does not exist" containerID="d34d02720b57556792ddbe505c365ac4d0624cdbac38b025ea82412f079e3847" Dec 11 02:36:45 crc kubenswrapper[4824]: I1211 02:36:45.059462 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d34d02720b57556792ddbe505c365ac4d0624cdbac38b025ea82412f079e3847"} err="failed to get container status \"d34d02720b57556792ddbe505c365ac4d0624cdbac38b025ea82412f079e3847\": rpc error: code = NotFound desc = could not find container \"d34d02720b57556792ddbe505c365ac4d0624cdbac38b025ea82412f079e3847\": container with ID starting with d34d02720b57556792ddbe505c365ac4d0624cdbac38b025ea82412f079e3847 not found: ID does not exist" Dec 11 02:36:45 crc kubenswrapper[4824]: I1211 02:36:45.068883 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-grt6q"] Dec 11 02:36:45 crc kubenswrapper[4824]: I1211 02:36:45.106000 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-grt6q"] Dec 11 02:36:45 crc kubenswrapper[4824]: I1211 02:36:45.122090 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-grms5"] Dec 11 02:36:45 crc kubenswrapper[4824]: E1211 02:36:45.122577 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49660600-bcea-4c04-ad34-30a06d1b795c" containerName="extract-utilities" Dec 11 02:36:45 crc kubenswrapper[4824]: I1211 02:36:45.122598 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="49660600-bcea-4c04-ad34-30a06d1b795c" containerName="extract-utilities" Dec 11 02:36:45 crc kubenswrapper[4824]: E1211 02:36:45.122633 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49660600-bcea-4c04-ad34-30a06d1b795c" containerName="extract-content" Dec 11 02:36:45 crc kubenswrapper[4824]: I1211 02:36:45.122644 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="49660600-bcea-4c04-ad34-30a06d1b795c" containerName="extract-content" Dec 11 02:36:45 crc kubenswrapper[4824]: E1211 02:36:45.122660 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49660600-bcea-4c04-ad34-30a06d1b795c" containerName="registry-server" Dec 11 02:36:45 crc kubenswrapper[4824]: I1211 02:36:45.122668 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="49660600-bcea-4c04-ad34-30a06d1b795c" containerName="registry-server" Dec 11 02:36:45 crc kubenswrapper[4824]: E1211 02:36:45.122696 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20ff7897-84b7-4c8c-b9f9-fc7721488ea7" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 11 02:36:45 crc kubenswrapper[4824]: I1211 02:36:45.122706 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="20ff7897-84b7-4c8c-b9f9-fc7721488ea7" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 11 02:36:45 crc kubenswrapper[4824]: I1211 02:36:45.122933 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="20ff7897-84b7-4c8c-b9f9-fc7721488ea7" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 11 02:36:45 crc kubenswrapper[4824]: I1211 02:36:45.122946 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="49660600-bcea-4c04-ad34-30a06d1b795c" containerName="registry-server" Dec 11 02:36:45 crc kubenswrapper[4824]: I1211 02:36:45.123754 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-grms5" Dec 11 02:36:45 crc kubenswrapper[4824]: I1211 02:36:45.126228 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 11 02:36:45 crc kubenswrapper[4824]: I1211 02:36:45.126423 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 11 02:36:45 crc kubenswrapper[4824]: I1211 02:36:45.126534 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 11 02:36:45 crc kubenswrapper[4824]: I1211 02:36:45.126662 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-hs2fv" Dec 11 02:36:45 crc kubenswrapper[4824]: I1211 02:36:45.126804 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Dec 11 02:36:45 crc kubenswrapper[4824]: I1211 02:36:45.137131 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-grms5"] Dec 11 02:36:45 crc kubenswrapper[4824]: I1211 02:36:45.283026 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c8ca447b-581a-45ba-af85-a1a4ff83687f-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-grms5\" (UID: \"c8ca447b-581a-45ba-af85-a1a4ff83687f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-grms5" Dec 11 02:36:45 crc kubenswrapper[4824]: I1211 02:36:45.283202 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6clqk\" (UniqueName: \"kubernetes.io/projected/c8ca447b-581a-45ba-af85-a1a4ff83687f-kube-api-access-6clqk\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-grms5\" (UID: \"c8ca447b-581a-45ba-af85-a1a4ff83687f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-grms5" Dec 11 02:36:45 crc kubenswrapper[4824]: I1211 02:36:45.283548 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c8ca447b-581a-45ba-af85-a1a4ff83687f-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-grms5\" (UID: \"c8ca447b-581a-45ba-af85-a1a4ff83687f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-grms5" Dec 11 02:36:45 crc kubenswrapper[4824]: I1211 02:36:45.283725 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/c8ca447b-581a-45ba-af85-a1a4ff83687f-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-grms5\" (UID: \"c8ca447b-581a-45ba-af85-a1a4ff83687f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-grms5" Dec 11 02:36:45 crc kubenswrapper[4824]: I1211 02:36:45.283782 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8ca447b-581a-45ba-af85-a1a4ff83687f-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-grms5\" (UID: \"c8ca447b-581a-45ba-af85-a1a4ff83687f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-grms5" Dec 11 02:36:45 crc kubenswrapper[4824]: I1211 02:36:45.385889 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c8ca447b-581a-45ba-af85-a1a4ff83687f-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-grms5\" (UID: \"c8ca447b-581a-45ba-af85-a1a4ff83687f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-grms5" Dec 11 02:36:45 crc kubenswrapper[4824]: I1211 02:36:45.385984 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/c8ca447b-581a-45ba-af85-a1a4ff83687f-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-grms5\" (UID: \"c8ca447b-581a-45ba-af85-a1a4ff83687f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-grms5" Dec 11 02:36:45 crc kubenswrapper[4824]: I1211 02:36:45.386020 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8ca447b-581a-45ba-af85-a1a4ff83687f-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-grms5\" (UID: \"c8ca447b-581a-45ba-af85-a1a4ff83687f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-grms5" Dec 11 02:36:45 crc kubenswrapper[4824]: I1211 02:36:45.386093 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c8ca447b-581a-45ba-af85-a1a4ff83687f-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-grms5\" (UID: \"c8ca447b-581a-45ba-af85-a1a4ff83687f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-grms5" Dec 11 02:36:45 crc kubenswrapper[4824]: I1211 02:36:45.386261 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6clqk\" (UniqueName: \"kubernetes.io/projected/c8ca447b-581a-45ba-af85-a1a4ff83687f-kube-api-access-6clqk\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-grms5\" (UID: \"c8ca447b-581a-45ba-af85-a1a4ff83687f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-grms5" Dec 11 02:36:45 crc kubenswrapper[4824]: I1211 02:36:45.387587 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/c8ca447b-581a-45ba-af85-a1a4ff83687f-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-grms5\" (UID: \"c8ca447b-581a-45ba-af85-a1a4ff83687f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-grms5" Dec 11 02:36:45 crc kubenswrapper[4824]: I1211 02:36:45.391751 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8ca447b-581a-45ba-af85-a1a4ff83687f-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-grms5\" (UID: \"c8ca447b-581a-45ba-af85-a1a4ff83687f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-grms5" Dec 11 02:36:45 crc kubenswrapper[4824]: I1211 02:36:45.392428 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c8ca447b-581a-45ba-af85-a1a4ff83687f-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-grms5\" (UID: \"c8ca447b-581a-45ba-af85-a1a4ff83687f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-grms5" Dec 11 02:36:45 crc kubenswrapper[4824]: I1211 02:36:45.404366 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c8ca447b-581a-45ba-af85-a1a4ff83687f-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-grms5\" (UID: \"c8ca447b-581a-45ba-af85-a1a4ff83687f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-grms5" Dec 11 02:36:45 crc kubenswrapper[4824]: I1211 02:36:45.411763 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6clqk\" (UniqueName: \"kubernetes.io/projected/c8ca447b-581a-45ba-af85-a1a4ff83687f-kube-api-access-6clqk\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-grms5\" (UID: \"c8ca447b-581a-45ba-af85-a1a4ff83687f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-grms5" Dec 11 02:36:45 crc kubenswrapper[4824]: I1211 02:36:45.452406 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-grms5" Dec 11 02:36:46 crc kubenswrapper[4824]: I1211 02:36:46.156501 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-grms5"] Dec 11 02:36:46 crc kubenswrapper[4824]: I1211 02:36:46.644661 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49660600-bcea-4c04-ad34-30a06d1b795c" path="/var/lib/kubelet/pods/49660600-bcea-4c04-ad34-30a06d1b795c/volumes" Dec 11 02:36:46 crc kubenswrapper[4824]: I1211 02:36:46.989477 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-grms5" event={"ID":"c8ca447b-581a-45ba-af85-a1a4ff83687f","Type":"ContainerStarted","Data":"961c10e9952b0bbcbc72e2fab1193c745c96722120de058bfddd6e33d011e8fa"} Dec 11 02:36:46 crc kubenswrapper[4824]: I1211 02:36:46.989535 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-grms5" event={"ID":"c8ca447b-581a-45ba-af85-a1a4ff83687f","Type":"ContainerStarted","Data":"b18a833afad53ae37e49dc03000c4ea5e1dbfdae9f333a1a7daccbb44b28fc84"} Dec 11 02:36:47 crc kubenswrapper[4824]: I1211 02:36:47.008590 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-grms5" podStartSLOduration=1.810994615 podStartE2EDuration="2.008570598s" podCreationTimestamp="2025-12-11 02:36:45 +0000 UTC" firstStartedPulling="2025-12-11 02:36:46.165456877 +0000 UTC m=+2147.854494266" lastFinishedPulling="2025-12-11 02:36:46.36303285 +0000 UTC m=+2148.052070249" observedRunningTime="2025-12-11 02:36:47.003413108 +0000 UTC m=+2148.692450517" watchObservedRunningTime="2025-12-11 02:36:47.008570598 +0000 UTC m=+2148.697607977" Dec 11 02:36:54 crc kubenswrapper[4824]: I1211 02:36:54.793104 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-966n9"] Dec 11 02:36:54 crc kubenswrapper[4824]: I1211 02:36:54.801988 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-966n9" Dec 11 02:36:54 crc kubenswrapper[4824]: I1211 02:36:54.805415 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-966n9"] Dec 11 02:36:54 crc kubenswrapper[4824]: I1211 02:36:54.887748 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afe04fa0-ec8f-4e4a-a7ca-26ed402d3cc6-utilities\") pod \"redhat-operators-966n9\" (UID: \"afe04fa0-ec8f-4e4a-a7ca-26ed402d3cc6\") " pod="openshift-marketplace/redhat-operators-966n9" Dec 11 02:36:54 crc kubenswrapper[4824]: I1211 02:36:54.887795 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afe04fa0-ec8f-4e4a-a7ca-26ed402d3cc6-catalog-content\") pod \"redhat-operators-966n9\" (UID: \"afe04fa0-ec8f-4e4a-a7ca-26ed402d3cc6\") " pod="openshift-marketplace/redhat-operators-966n9" Dec 11 02:36:54 crc kubenswrapper[4824]: I1211 02:36:54.887896 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z9ccm\" (UniqueName: \"kubernetes.io/projected/afe04fa0-ec8f-4e4a-a7ca-26ed402d3cc6-kube-api-access-z9ccm\") pod \"redhat-operators-966n9\" (UID: \"afe04fa0-ec8f-4e4a-a7ca-26ed402d3cc6\") " pod="openshift-marketplace/redhat-operators-966n9" Dec 11 02:36:54 crc kubenswrapper[4824]: I1211 02:36:54.989425 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afe04fa0-ec8f-4e4a-a7ca-26ed402d3cc6-utilities\") pod \"redhat-operators-966n9\" (UID: \"afe04fa0-ec8f-4e4a-a7ca-26ed402d3cc6\") " pod="openshift-marketplace/redhat-operators-966n9" Dec 11 02:36:54 crc kubenswrapper[4824]: I1211 02:36:54.989468 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afe04fa0-ec8f-4e4a-a7ca-26ed402d3cc6-catalog-content\") pod \"redhat-operators-966n9\" (UID: \"afe04fa0-ec8f-4e4a-a7ca-26ed402d3cc6\") " pod="openshift-marketplace/redhat-operators-966n9" Dec 11 02:36:54 crc kubenswrapper[4824]: I1211 02:36:54.989543 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z9ccm\" (UniqueName: \"kubernetes.io/projected/afe04fa0-ec8f-4e4a-a7ca-26ed402d3cc6-kube-api-access-z9ccm\") pod \"redhat-operators-966n9\" (UID: \"afe04fa0-ec8f-4e4a-a7ca-26ed402d3cc6\") " pod="openshift-marketplace/redhat-operators-966n9" Dec 11 02:36:54 crc kubenswrapper[4824]: I1211 02:36:54.990261 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afe04fa0-ec8f-4e4a-a7ca-26ed402d3cc6-utilities\") pod \"redhat-operators-966n9\" (UID: \"afe04fa0-ec8f-4e4a-a7ca-26ed402d3cc6\") " pod="openshift-marketplace/redhat-operators-966n9" Dec 11 02:36:54 crc kubenswrapper[4824]: I1211 02:36:54.990310 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afe04fa0-ec8f-4e4a-a7ca-26ed402d3cc6-catalog-content\") pod \"redhat-operators-966n9\" (UID: \"afe04fa0-ec8f-4e4a-a7ca-26ed402d3cc6\") " pod="openshift-marketplace/redhat-operators-966n9" Dec 11 02:36:55 crc kubenswrapper[4824]: I1211 02:36:55.011806 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z9ccm\" (UniqueName: \"kubernetes.io/projected/afe04fa0-ec8f-4e4a-a7ca-26ed402d3cc6-kube-api-access-z9ccm\") pod \"redhat-operators-966n9\" (UID: \"afe04fa0-ec8f-4e4a-a7ca-26ed402d3cc6\") " pod="openshift-marketplace/redhat-operators-966n9" Dec 11 02:36:55 crc kubenswrapper[4824]: I1211 02:36:55.131507 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-966n9" Dec 11 02:36:55 crc kubenswrapper[4824]: I1211 02:36:55.643867 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-966n9"] Dec 11 02:36:56 crc kubenswrapper[4824]: I1211 02:36:56.079142 4824 generic.go:334] "Generic (PLEG): container finished" podID="afe04fa0-ec8f-4e4a-a7ca-26ed402d3cc6" containerID="3a94ca43765e4302f256c340b330d441544b451b85fcaae9678e2cfef9ada317" exitCode=0 Dec 11 02:36:56 crc kubenswrapper[4824]: I1211 02:36:56.079248 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-966n9" event={"ID":"afe04fa0-ec8f-4e4a-a7ca-26ed402d3cc6","Type":"ContainerDied","Data":"3a94ca43765e4302f256c340b330d441544b451b85fcaae9678e2cfef9ada317"} Dec 11 02:36:56 crc kubenswrapper[4824]: I1211 02:36:56.080139 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-966n9" event={"ID":"afe04fa0-ec8f-4e4a-a7ca-26ed402d3cc6","Type":"ContainerStarted","Data":"e13954b59555133211c0fd4926c3bb1decee52f664f01c6207fbd478c055fdd8"} Dec 11 02:36:57 crc kubenswrapper[4824]: I1211 02:36:57.099579 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-966n9" event={"ID":"afe04fa0-ec8f-4e4a-a7ca-26ed402d3cc6","Type":"ContainerStarted","Data":"96089da99e04924c9905905f8ec45c7206b8bad9ecd7d09cd9befa6bae221bc7"} Dec 11 02:37:00 crc kubenswrapper[4824]: I1211 02:37:00.640411 4824 generic.go:334] "Generic (PLEG): container finished" podID="afe04fa0-ec8f-4e4a-a7ca-26ed402d3cc6" containerID="96089da99e04924c9905905f8ec45c7206b8bad9ecd7d09cd9befa6bae221bc7" exitCode=0 Dec 11 02:37:00 crc kubenswrapper[4824]: I1211 02:37:00.644633 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-966n9" event={"ID":"afe04fa0-ec8f-4e4a-a7ca-26ed402d3cc6","Type":"ContainerDied","Data":"96089da99e04924c9905905f8ec45c7206b8bad9ecd7d09cd9befa6bae221bc7"} Dec 11 02:37:02 crc kubenswrapper[4824]: I1211 02:37:02.683299 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-966n9" event={"ID":"afe04fa0-ec8f-4e4a-a7ca-26ed402d3cc6","Type":"ContainerStarted","Data":"447adf78c46f4b3abc86b87ffed7efe3a982bfecfb142012a777eb181d957b5b"} Dec 11 02:37:02 crc kubenswrapper[4824]: I1211 02:37:02.711835 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-966n9" podStartSLOduration=3.214440641 podStartE2EDuration="8.711815799s" podCreationTimestamp="2025-12-11 02:36:54 +0000 UTC" firstStartedPulling="2025-12-11 02:36:56.081202086 +0000 UTC m=+2157.770239465" lastFinishedPulling="2025-12-11 02:37:01.578577234 +0000 UTC m=+2163.267614623" observedRunningTime="2025-12-11 02:37:02.701968161 +0000 UTC m=+2164.391005540" watchObservedRunningTime="2025-12-11 02:37:02.711815799 +0000 UTC m=+2164.400853188" Dec 11 02:37:05 crc kubenswrapper[4824]: I1211 02:37:05.132171 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-966n9" Dec 11 02:37:05 crc kubenswrapper[4824]: I1211 02:37:05.132441 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-966n9" Dec 11 02:37:06 crc kubenswrapper[4824]: I1211 02:37:06.197453 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-966n9" podUID="afe04fa0-ec8f-4e4a-a7ca-26ed402d3cc6" containerName="registry-server" probeResult="failure" output=< Dec 11 02:37:06 crc kubenswrapper[4824]: timeout: failed to connect service ":50051" within 1s Dec 11 02:37:06 crc kubenswrapper[4824]: > Dec 11 02:37:07 crc kubenswrapper[4824]: I1211 02:37:07.251796 4824 patch_prober.go:28] interesting pod/machine-config-daemon-gx6xt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 02:37:07 crc kubenswrapper[4824]: I1211 02:37:07.251863 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 02:37:15 crc kubenswrapper[4824]: I1211 02:37:15.172832 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-966n9" Dec 11 02:37:15 crc kubenswrapper[4824]: I1211 02:37:15.224393 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-966n9" Dec 11 02:37:15 crc kubenswrapper[4824]: I1211 02:37:15.421139 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-966n9"] Dec 11 02:37:16 crc kubenswrapper[4824]: I1211 02:37:16.871663 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-966n9" podUID="afe04fa0-ec8f-4e4a-a7ca-26ed402d3cc6" containerName="registry-server" containerID="cri-o://447adf78c46f4b3abc86b87ffed7efe3a982bfecfb142012a777eb181d957b5b" gracePeriod=2 Dec 11 02:37:17 crc kubenswrapper[4824]: I1211 02:37:17.376637 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-966n9" Dec 11 02:37:17 crc kubenswrapper[4824]: I1211 02:37:17.508273 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z9ccm\" (UniqueName: \"kubernetes.io/projected/afe04fa0-ec8f-4e4a-a7ca-26ed402d3cc6-kube-api-access-z9ccm\") pod \"afe04fa0-ec8f-4e4a-a7ca-26ed402d3cc6\" (UID: \"afe04fa0-ec8f-4e4a-a7ca-26ed402d3cc6\") " Dec 11 02:37:17 crc kubenswrapper[4824]: I1211 02:37:17.508367 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afe04fa0-ec8f-4e4a-a7ca-26ed402d3cc6-catalog-content\") pod \"afe04fa0-ec8f-4e4a-a7ca-26ed402d3cc6\" (UID: \"afe04fa0-ec8f-4e4a-a7ca-26ed402d3cc6\") " Dec 11 02:37:17 crc kubenswrapper[4824]: I1211 02:37:17.508424 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afe04fa0-ec8f-4e4a-a7ca-26ed402d3cc6-utilities\") pod \"afe04fa0-ec8f-4e4a-a7ca-26ed402d3cc6\" (UID: \"afe04fa0-ec8f-4e4a-a7ca-26ed402d3cc6\") " Dec 11 02:37:17 crc kubenswrapper[4824]: I1211 02:37:17.509331 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/afe04fa0-ec8f-4e4a-a7ca-26ed402d3cc6-utilities" (OuterVolumeSpecName: "utilities") pod "afe04fa0-ec8f-4e4a-a7ca-26ed402d3cc6" (UID: "afe04fa0-ec8f-4e4a-a7ca-26ed402d3cc6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:37:17 crc kubenswrapper[4824]: I1211 02:37:17.521507 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/afe04fa0-ec8f-4e4a-a7ca-26ed402d3cc6-kube-api-access-z9ccm" (OuterVolumeSpecName: "kube-api-access-z9ccm") pod "afe04fa0-ec8f-4e4a-a7ca-26ed402d3cc6" (UID: "afe04fa0-ec8f-4e4a-a7ca-26ed402d3cc6"). InnerVolumeSpecName "kube-api-access-z9ccm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:37:17 crc kubenswrapper[4824]: I1211 02:37:17.610706 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afe04fa0-ec8f-4e4a-a7ca-26ed402d3cc6-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 02:37:17 crc kubenswrapper[4824]: I1211 02:37:17.610753 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z9ccm\" (UniqueName: \"kubernetes.io/projected/afe04fa0-ec8f-4e4a-a7ca-26ed402d3cc6-kube-api-access-z9ccm\") on node \"crc\" DevicePath \"\"" Dec 11 02:37:17 crc kubenswrapper[4824]: I1211 02:37:17.627843 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/afe04fa0-ec8f-4e4a-a7ca-26ed402d3cc6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "afe04fa0-ec8f-4e4a-a7ca-26ed402d3cc6" (UID: "afe04fa0-ec8f-4e4a-a7ca-26ed402d3cc6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:37:17 crc kubenswrapper[4824]: I1211 02:37:17.713659 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afe04fa0-ec8f-4e4a-a7ca-26ed402d3cc6-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 02:37:17 crc kubenswrapper[4824]: I1211 02:37:17.888358 4824 generic.go:334] "Generic (PLEG): container finished" podID="afe04fa0-ec8f-4e4a-a7ca-26ed402d3cc6" containerID="447adf78c46f4b3abc86b87ffed7efe3a982bfecfb142012a777eb181d957b5b" exitCode=0 Dec 11 02:37:17 crc kubenswrapper[4824]: I1211 02:37:17.888428 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-966n9" event={"ID":"afe04fa0-ec8f-4e4a-a7ca-26ed402d3cc6","Type":"ContainerDied","Data":"447adf78c46f4b3abc86b87ffed7efe3a982bfecfb142012a777eb181d957b5b"} Dec 11 02:37:17 crc kubenswrapper[4824]: I1211 02:37:17.888478 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-966n9" event={"ID":"afe04fa0-ec8f-4e4a-a7ca-26ed402d3cc6","Type":"ContainerDied","Data":"e13954b59555133211c0fd4926c3bb1decee52f664f01c6207fbd478c055fdd8"} Dec 11 02:37:17 crc kubenswrapper[4824]: I1211 02:37:17.888474 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-966n9" Dec 11 02:37:17 crc kubenswrapper[4824]: I1211 02:37:17.888525 4824 scope.go:117] "RemoveContainer" containerID="447adf78c46f4b3abc86b87ffed7efe3a982bfecfb142012a777eb181d957b5b" Dec 11 02:37:17 crc kubenswrapper[4824]: I1211 02:37:17.920396 4824 scope.go:117] "RemoveContainer" containerID="96089da99e04924c9905905f8ec45c7206b8bad9ecd7d09cd9befa6bae221bc7" Dec 11 02:37:17 crc kubenswrapper[4824]: I1211 02:37:17.948165 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-966n9"] Dec 11 02:37:17 crc kubenswrapper[4824]: I1211 02:37:17.966179 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-966n9"] Dec 11 02:37:17 crc kubenswrapper[4824]: I1211 02:37:17.991920 4824 scope.go:117] "RemoveContainer" containerID="3a94ca43765e4302f256c340b330d441544b451b85fcaae9678e2cfef9ada317" Dec 11 02:37:18 crc kubenswrapper[4824]: I1211 02:37:18.034080 4824 scope.go:117] "RemoveContainer" containerID="447adf78c46f4b3abc86b87ffed7efe3a982bfecfb142012a777eb181d957b5b" Dec 11 02:37:18 crc kubenswrapper[4824]: E1211 02:37:18.034763 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"447adf78c46f4b3abc86b87ffed7efe3a982bfecfb142012a777eb181d957b5b\": container with ID starting with 447adf78c46f4b3abc86b87ffed7efe3a982bfecfb142012a777eb181d957b5b not found: ID does not exist" containerID="447adf78c46f4b3abc86b87ffed7efe3a982bfecfb142012a777eb181d957b5b" Dec 11 02:37:18 crc kubenswrapper[4824]: I1211 02:37:18.034821 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"447adf78c46f4b3abc86b87ffed7efe3a982bfecfb142012a777eb181d957b5b"} err="failed to get container status \"447adf78c46f4b3abc86b87ffed7efe3a982bfecfb142012a777eb181d957b5b\": rpc error: code = NotFound desc = could not find container \"447adf78c46f4b3abc86b87ffed7efe3a982bfecfb142012a777eb181d957b5b\": container with ID starting with 447adf78c46f4b3abc86b87ffed7efe3a982bfecfb142012a777eb181d957b5b not found: ID does not exist" Dec 11 02:37:18 crc kubenswrapper[4824]: I1211 02:37:18.034863 4824 scope.go:117] "RemoveContainer" containerID="96089da99e04924c9905905f8ec45c7206b8bad9ecd7d09cd9befa6bae221bc7" Dec 11 02:37:18 crc kubenswrapper[4824]: E1211 02:37:18.035353 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"96089da99e04924c9905905f8ec45c7206b8bad9ecd7d09cd9befa6bae221bc7\": container with ID starting with 96089da99e04924c9905905f8ec45c7206b8bad9ecd7d09cd9befa6bae221bc7 not found: ID does not exist" containerID="96089da99e04924c9905905f8ec45c7206b8bad9ecd7d09cd9befa6bae221bc7" Dec 11 02:37:18 crc kubenswrapper[4824]: I1211 02:37:18.035448 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"96089da99e04924c9905905f8ec45c7206b8bad9ecd7d09cd9befa6bae221bc7"} err="failed to get container status \"96089da99e04924c9905905f8ec45c7206b8bad9ecd7d09cd9befa6bae221bc7\": rpc error: code = NotFound desc = could not find container \"96089da99e04924c9905905f8ec45c7206b8bad9ecd7d09cd9befa6bae221bc7\": container with ID starting with 96089da99e04924c9905905f8ec45c7206b8bad9ecd7d09cd9befa6bae221bc7 not found: ID does not exist" Dec 11 02:37:18 crc kubenswrapper[4824]: I1211 02:37:18.035510 4824 scope.go:117] "RemoveContainer" containerID="3a94ca43765e4302f256c340b330d441544b451b85fcaae9678e2cfef9ada317" Dec 11 02:37:18 crc kubenswrapper[4824]: E1211 02:37:18.035896 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a94ca43765e4302f256c340b330d441544b451b85fcaae9678e2cfef9ada317\": container with ID starting with 3a94ca43765e4302f256c340b330d441544b451b85fcaae9678e2cfef9ada317 not found: ID does not exist" containerID="3a94ca43765e4302f256c340b330d441544b451b85fcaae9678e2cfef9ada317" Dec 11 02:37:18 crc kubenswrapper[4824]: I1211 02:37:18.035931 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a94ca43765e4302f256c340b330d441544b451b85fcaae9678e2cfef9ada317"} err="failed to get container status \"3a94ca43765e4302f256c340b330d441544b451b85fcaae9678e2cfef9ada317\": rpc error: code = NotFound desc = could not find container \"3a94ca43765e4302f256c340b330d441544b451b85fcaae9678e2cfef9ada317\": container with ID starting with 3a94ca43765e4302f256c340b330d441544b451b85fcaae9678e2cfef9ada317 not found: ID does not exist" Dec 11 02:37:18 crc kubenswrapper[4824]: I1211 02:37:18.649607 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="afe04fa0-ec8f-4e4a-a7ca-26ed402d3cc6" path="/var/lib/kubelet/pods/afe04fa0-ec8f-4e4a-a7ca-26ed402d3cc6/volumes" Dec 11 02:37:37 crc kubenswrapper[4824]: I1211 02:37:37.251557 4824 patch_prober.go:28] interesting pod/machine-config-daemon-gx6xt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 02:37:37 crc kubenswrapper[4824]: I1211 02:37:37.252215 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 02:37:37 crc kubenswrapper[4824]: I1211 02:37:37.252279 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" Dec 11 02:37:37 crc kubenswrapper[4824]: I1211 02:37:37.253260 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"592f00198a08d7f30f2bd4761177c27cba23107c7971ce2c5b2441d9ff2d1144"} pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 11 02:37:37 crc kubenswrapper[4824]: I1211 02:37:37.253354 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" containerName="machine-config-daemon" containerID="cri-o://592f00198a08d7f30f2bd4761177c27cba23107c7971ce2c5b2441d9ff2d1144" gracePeriod=600 Dec 11 02:37:37 crc kubenswrapper[4824]: E1211 02:37:37.377880 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 02:37:38 crc kubenswrapper[4824]: I1211 02:37:38.104189 4824 generic.go:334] "Generic (PLEG): container finished" podID="44b156e4-64a4-4d45-aa5e-9b10a862faed" containerID="592f00198a08d7f30f2bd4761177c27cba23107c7971ce2c5b2441d9ff2d1144" exitCode=0 Dec 11 02:37:38 crc kubenswrapper[4824]: I1211 02:37:38.104242 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" event={"ID":"44b156e4-64a4-4d45-aa5e-9b10a862faed","Type":"ContainerDied","Data":"592f00198a08d7f30f2bd4761177c27cba23107c7971ce2c5b2441d9ff2d1144"} Dec 11 02:37:38 crc kubenswrapper[4824]: I1211 02:37:38.104279 4824 scope.go:117] "RemoveContainer" containerID="80aa24371216c8cd3cae76744f65819f6cda361ceaf0439f8ef7d77b23dd30ff" Dec 11 02:37:38 crc kubenswrapper[4824]: I1211 02:37:38.104972 4824 scope.go:117] "RemoveContainer" containerID="592f00198a08d7f30f2bd4761177c27cba23107c7971ce2c5b2441d9ff2d1144" Dec 11 02:37:38 crc kubenswrapper[4824]: E1211 02:37:38.105413 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 02:37:48 crc kubenswrapper[4824]: I1211 02:37:48.644061 4824 scope.go:117] "RemoveContainer" containerID="592f00198a08d7f30f2bd4761177c27cba23107c7971ce2c5b2441d9ff2d1144" Dec 11 02:37:48 crc kubenswrapper[4824]: E1211 02:37:48.645529 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 02:37:59 crc kubenswrapper[4824]: I1211 02:37:59.633324 4824 scope.go:117] "RemoveContainer" containerID="592f00198a08d7f30f2bd4761177c27cba23107c7971ce2c5b2441d9ff2d1144" Dec 11 02:37:59 crc kubenswrapper[4824]: E1211 02:37:59.634306 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 02:38:00 crc kubenswrapper[4824]: I1211 02:38:00.354740 4824 generic.go:334] "Generic (PLEG): container finished" podID="c8ca447b-581a-45ba-af85-a1a4ff83687f" containerID="961c10e9952b0bbcbc72e2fab1193c745c96722120de058bfddd6e33d011e8fa" exitCode=0 Dec 11 02:38:00 crc kubenswrapper[4824]: I1211 02:38:00.354808 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-grms5" event={"ID":"c8ca447b-581a-45ba-af85-a1a4ff83687f","Type":"ContainerDied","Data":"961c10e9952b0bbcbc72e2fab1193c745c96722120de058bfddd6e33d011e8fa"} Dec 11 02:38:01 crc kubenswrapper[4824]: I1211 02:38:01.859280 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-grms5" Dec 11 02:38:01 crc kubenswrapper[4824]: I1211 02:38:01.966706 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/c8ca447b-581a-45ba-af85-a1a4ff83687f-ovncontroller-config-0\") pod \"c8ca447b-581a-45ba-af85-a1a4ff83687f\" (UID: \"c8ca447b-581a-45ba-af85-a1a4ff83687f\") " Dec 11 02:38:01 crc kubenswrapper[4824]: I1211 02:38:01.966931 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c8ca447b-581a-45ba-af85-a1a4ff83687f-ssh-key\") pod \"c8ca447b-581a-45ba-af85-a1a4ff83687f\" (UID: \"c8ca447b-581a-45ba-af85-a1a4ff83687f\") " Dec 11 02:38:01 crc kubenswrapper[4824]: I1211 02:38:01.967251 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8ca447b-581a-45ba-af85-a1a4ff83687f-ovn-combined-ca-bundle\") pod \"c8ca447b-581a-45ba-af85-a1a4ff83687f\" (UID: \"c8ca447b-581a-45ba-af85-a1a4ff83687f\") " Dec 11 02:38:01 crc kubenswrapper[4824]: I1211 02:38:01.967336 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c8ca447b-581a-45ba-af85-a1a4ff83687f-inventory\") pod \"c8ca447b-581a-45ba-af85-a1a4ff83687f\" (UID: \"c8ca447b-581a-45ba-af85-a1a4ff83687f\") " Dec 11 02:38:01 crc kubenswrapper[4824]: I1211 02:38:01.967452 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6clqk\" (UniqueName: \"kubernetes.io/projected/c8ca447b-581a-45ba-af85-a1a4ff83687f-kube-api-access-6clqk\") pod \"c8ca447b-581a-45ba-af85-a1a4ff83687f\" (UID: \"c8ca447b-581a-45ba-af85-a1a4ff83687f\") " Dec 11 02:38:01 crc kubenswrapper[4824]: I1211 02:38:01.972539 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8ca447b-581a-45ba-af85-a1a4ff83687f-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "c8ca447b-581a-45ba-af85-a1a4ff83687f" (UID: "c8ca447b-581a-45ba-af85-a1a4ff83687f"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:38:01 crc kubenswrapper[4824]: I1211 02:38:01.979730 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8ca447b-581a-45ba-af85-a1a4ff83687f-kube-api-access-6clqk" (OuterVolumeSpecName: "kube-api-access-6clqk") pod "c8ca447b-581a-45ba-af85-a1a4ff83687f" (UID: "c8ca447b-581a-45ba-af85-a1a4ff83687f"). InnerVolumeSpecName "kube-api-access-6clqk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:38:02 crc kubenswrapper[4824]: I1211 02:38:02.002752 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c8ca447b-581a-45ba-af85-a1a4ff83687f-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "c8ca447b-581a-45ba-af85-a1a4ff83687f" (UID: "c8ca447b-581a-45ba-af85-a1a4ff83687f"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:38:02 crc kubenswrapper[4824]: I1211 02:38:02.006579 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8ca447b-581a-45ba-af85-a1a4ff83687f-inventory" (OuterVolumeSpecName: "inventory") pod "c8ca447b-581a-45ba-af85-a1a4ff83687f" (UID: "c8ca447b-581a-45ba-af85-a1a4ff83687f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:38:02 crc kubenswrapper[4824]: I1211 02:38:02.010920 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8ca447b-581a-45ba-af85-a1a4ff83687f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c8ca447b-581a-45ba-af85-a1a4ff83687f" (UID: "c8ca447b-581a-45ba-af85-a1a4ff83687f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:38:02 crc kubenswrapper[4824]: I1211 02:38:02.069784 4824 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8ca447b-581a-45ba-af85-a1a4ff83687f-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 02:38:02 crc kubenswrapper[4824]: I1211 02:38:02.069829 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c8ca447b-581a-45ba-af85-a1a4ff83687f-inventory\") on node \"crc\" DevicePath \"\"" Dec 11 02:38:02 crc kubenswrapper[4824]: I1211 02:38:02.069843 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6clqk\" (UniqueName: \"kubernetes.io/projected/c8ca447b-581a-45ba-af85-a1a4ff83687f-kube-api-access-6clqk\") on node \"crc\" DevicePath \"\"" Dec 11 02:38:02 crc kubenswrapper[4824]: I1211 02:38:02.069859 4824 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/c8ca447b-581a-45ba-af85-a1a4ff83687f-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Dec 11 02:38:02 crc kubenswrapper[4824]: I1211 02:38:02.069871 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c8ca447b-581a-45ba-af85-a1a4ff83687f-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 11 02:38:02 crc kubenswrapper[4824]: I1211 02:38:02.375809 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-grms5" event={"ID":"c8ca447b-581a-45ba-af85-a1a4ff83687f","Type":"ContainerDied","Data":"b18a833afad53ae37e49dc03000c4ea5e1dbfdae9f333a1a7daccbb44b28fc84"} Dec 11 02:38:02 crc kubenswrapper[4824]: I1211 02:38:02.375855 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b18a833afad53ae37e49dc03000c4ea5e1dbfdae9f333a1a7daccbb44b28fc84" Dec 11 02:38:02 crc kubenswrapper[4824]: I1211 02:38:02.375861 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-grms5" Dec 11 02:38:02 crc kubenswrapper[4824]: I1211 02:38:02.519972 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8rbcc"] Dec 11 02:38:02 crc kubenswrapper[4824]: E1211 02:38:02.520531 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8ca447b-581a-45ba-af85-a1a4ff83687f" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 11 02:38:02 crc kubenswrapper[4824]: I1211 02:38:02.520553 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8ca447b-581a-45ba-af85-a1a4ff83687f" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 11 02:38:02 crc kubenswrapper[4824]: E1211 02:38:02.520593 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afe04fa0-ec8f-4e4a-a7ca-26ed402d3cc6" containerName="extract-utilities" Dec 11 02:38:02 crc kubenswrapper[4824]: I1211 02:38:02.520602 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="afe04fa0-ec8f-4e4a-a7ca-26ed402d3cc6" containerName="extract-utilities" Dec 11 02:38:02 crc kubenswrapper[4824]: E1211 02:38:02.520619 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afe04fa0-ec8f-4e4a-a7ca-26ed402d3cc6" containerName="registry-server" Dec 11 02:38:02 crc kubenswrapper[4824]: I1211 02:38:02.520627 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="afe04fa0-ec8f-4e4a-a7ca-26ed402d3cc6" containerName="registry-server" Dec 11 02:38:02 crc kubenswrapper[4824]: E1211 02:38:02.520665 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afe04fa0-ec8f-4e4a-a7ca-26ed402d3cc6" containerName="extract-content" Dec 11 02:38:02 crc kubenswrapper[4824]: I1211 02:38:02.520673 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="afe04fa0-ec8f-4e4a-a7ca-26ed402d3cc6" containerName="extract-content" Dec 11 02:38:02 crc kubenswrapper[4824]: I1211 02:38:02.520877 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="afe04fa0-ec8f-4e4a-a7ca-26ed402d3cc6" containerName="registry-server" Dec 11 02:38:02 crc kubenswrapper[4824]: I1211 02:38:02.520896 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8ca447b-581a-45ba-af85-a1a4ff83687f" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 11 02:38:02 crc kubenswrapper[4824]: I1211 02:38:02.521749 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8rbcc" Dec 11 02:38:02 crc kubenswrapper[4824]: I1211 02:38:02.523902 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Dec 11 02:38:02 crc kubenswrapper[4824]: I1211 02:38:02.524317 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 11 02:38:02 crc kubenswrapper[4824]: I1211 02:38:02.535545 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Dec 11 02:38:02 crc kubenswrapper[4824]: I1211 02:38:02.535843 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 11 02:38:02 crc kubenswrapper[4824]: I1211 02:38:02.535976 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 11 02:38:02 crc kubenswrapper[4824]: I1211 02:38:02.536189 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-hs2fv" Dec 11 02:38:02 crc kubenswrapper[4824]: I1211 02:38:02.539833 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8rbcc"] Dec 11 02:38:02 crc kubenswrapper[4824]: I1211 02:38:02.678742 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8736a409-d9d8-437e-965a-630f1ee6ae85-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-8rbcc\" (UID: \"8736a409-d9d8-437e-965a-630f1ee6ae85\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8rbcc" Dec 11 02:38:02 crc kubenswrapper[4824]: I1211 02:38:02.679162 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/8736a409-d9d8-437e-965a-630f1ee6ae85-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-8rbcc\" (UID: \"8736a409-d9d8-437e-965a-630f1ee6ae85\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8rbcc" Dec 11 02:38:02 crc kubenswrapper[4824]: I1211 02:38:02.679216 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v8zpk\" (UniqueName: \"kubernetes.io/projected/8736a409-d9d8-437e-965a-630f1ee6ae85-kube-api-access-v8zpk\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-8rbcc\" (UID: \"8736a409-d9d8-437e-965a-630f1ee6ae85\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8rbcc" Dec 11 02:38:02 crc kubenswrapper[4824]: I1211 02:38:02.679381 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8736a409-d9d8-437e-965a-630f1ee6ae85-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-8rbcc\" (UID: \"8736a409-d9d8-437e-965a-630f1ee6ae85\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8rbcc" Dec 11 02:38:02 crc kubenswrapper[4824]: I1211 02:38:02.679435 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8736a409-d9d8-437e-965a-630f1ee6ae85-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-8rbcc\" (UID: \"8736a409-d9d8-437e-965a-630f1ee6ae85\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8rbcc" Dec 11 02:38:02 crc kubenswrapper[4824]: I1211 02:38:02.679501 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/8736a409-d9d8-437e-965a-630f1ee6ae85-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-8rbcc\" (UID: \"8736a409-d9d8-437e-965a-630f1ee6ae85\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8rbcc" Dec 11 02:38:02 crc kubenswrapper[4824]: I1211 02:38:02.780841 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8736a409-d9d8-437e-965a-630f1ee6ae85-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-8rbcc\" (UID: \"8736a409-d9d8-437e-965a-630f1ee6ae85\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8rbcc" Dec 11 02:38:02 crc kubenswrapper[4824]: I1211 02:38:02.780908 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8736a409-d9d8-437e-965a-630f1ee6ae85-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-8rbcc\" (UID: \"8736a409-d9d8-437e-965a-630f1ee6ae85\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8rbcc" Dec 11 02:38:02 crc kubenswrapper[4824]: I1211 02:38:02.780937 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/8736a409-d9d8-437e-965a-630f1ee6ae85-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-8rbcc\" (UID: \"8736a409-d9d8-437e-965a-630f1ee6ae85\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8rbcc" Dec 11 02:38:02 crc kubenswrapper[4824]: I1211 02:38:02.780997 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8736a409-d9d8-437e-965a-630f1ee6ae85-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-8rbcc\" (UID: \"8736a409-d9d8-437e-965a-630f1ee6ae85\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8rbcc" Dec 11 02:38:02 crc kubenswrapper[4824]: I1211 02:38:02.781027 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/8736a409-d9d8-437e-965a-630f1ee6ae85-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-8rbcc\" (UID: \"8736a409-d9d8-437e-965a-630f1ee6ae85\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8rbcc" Dec 11 02:38:02 crc kubenswrapper[4824]: I1211 02:38:02.781053 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v8zpk\" (UniqueName: \"kubernetes.io/projected/8736a409-d9d8-437e-965a-630f1ee6ae85-kube-api-access-v8zpk\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-8rbcc\" (UID: \"8736a409-d9d8-437e-965a-630f1ee6ae85\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8rbcc" Dec 11 02:38:02 crc kubenswrapper[4824]: I1211 02:38:02.787285 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8736a409-d9d8-437e-965a-630f1ee6ae85-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-8rbcc\" (UID: \"8736a409-d9d8-437e-965a-630f1ee6ae85\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8rbcc" Dec 11 02:38:02 crc kubenswrapper[4824]: I1211 02:38:02.787629 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8736a409-d9d8-437e-965a-630f1ee6ae85-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-8rbcc\" (UID: \"8736a409-d9d8-437e-965a-630f1ee6ae85\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8rbcc" Dec 11 02:38:02 crc kubenswrapper[4824]: I1211 02:38:02.788189 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/8736a409-d9d8-437e-965a-630f1ee6ae85-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-8rbcc\" (UID: \"8736a409-d9d8-437e-965a-630f1ee6ae85\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8rbcc" Dec 11 02:38:02 crc kubenswrapper[4824]: I1211 02:38:02.789007 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8736a409-d9d8-437e-965a-630f1ee6ae85-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-8rbcc\" (UID: \"8736a409-d9d8-437e-965a-630f1ee6ae85\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8rbcc" Dec 11 02:38:02 crc kubenswrapper[4824]: I1211 02:38:02.799940 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/8736a409-d9d8-437e-965a-630f1ee6ae85-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-8rbcc\" (UID: \"8736a409-d9d8-437e-965a-630f1ee6ae85\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8rbcc" Dec 11 02:38:02 crc kubenswrapper[4824]: I1211 02:38:02.802661 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v8zpk\" (UniqueName: \"kubernetes.io/projected/8736a409-d9d8-437e-965a-630f1ee6ae85-kube-api-access-v8zpk\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-8rbcc\" (UID: \"8736a409-d9d8-437e-965a-630f1ee6ae85\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8rbcc" Dec 11 02:38:02 crc kubenswrapper[4824]: I1211 02:38:02.878818 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8rbcc" Dec 11 02:38:03 crc kubenswrapper[4824]: I1211 02:38:03.247600 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8rbcc"] Dec 11 02:38:03 crc kubenswrapper[4824]: W1211 02:38:03.248875 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8736a409_d9d8_437e_965a_630f1ee6ae85.slice/crio-47ab5837ddd4ffc4eaab8e7553057ae72050638c5dbb33aca22a6108d2f3051d WatchSource:0}: Error finding container 47ab5837ddd4ffc4eaab8e7553057ae72050638c5dbb33aca22a6108d2f3051d: Status 404 returned error can't find the container with id 47ab5837ddd4ffc4eaab8e7553057ae72050638c5dbb33aca22a6108d2f3051d Dec 11 02:38:03 crc kubenswrapper[4824]: I1211 02:38:03.390631 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8rbcc" event={"ID":"8736a409-d9d8-437e-965a-630f1ee6ae85","Type":"ContainerStarted","Data":"47ab5837ddd4ffc4eaab8e7553057ae72050638c5dbb33aca22a6108d2f3051d"} Dec 11 02:38:04 crc kubenswrapper[4824]: I1211 02:38:04.401452 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8rbcc" event={"ID":"8736a409-d9d8-437e-965a-630f1ee6ae85","Type":"ContainerStarted","Data":"55511b35433b1305df74789198b7bed5dc03e9fea6db02e2ebfc5c29d8ee92aa"} Dec 11 02:38:04 crc kubenswrapper[4824]: I1211 02:38:04.426347 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8rbcc" podStartSLOduration=2.239497928 podStartE2EDuration="2.426329207s" podCreationTimestamp="2025-12-11 02:38:02 +0000 UTC" firstStartedPulling="2025-12-11 02:38:03.251526073 +0000 UTC m=+2224.940563452" lastFinishedPulling="2025-12-11 02:38:03.438357352 +0000 UTC m=+2225.127394731" observedRunningTime="2025-12-11 02:38:04.420752377 +0000 UTC m=+2226.109789796" watchObservedRunningTime="2025-12-11 02:38:04.426329207 +0000 UTC m=+2226.115366586" Dec 11 02:38:11 crc kubenswrapper[4824]: I1211 02:38:11.633028 4824 scope.go:117] "RemoveContainer" containerID="592f00198a08d7f30f2bd4761177c27cba23107c7971ce2c5b2441d9ff2d1144" Dec 11 02:38:11 crc kubenswrapper[4824]: E1211 02:38:11.633789 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 02:38:22 crc kubenswrapper[4824]: I1211 02:38:22.633370 4824 scope.go:117] "RemoveContainer" containerID="592f00198a08d7f30f2bd4761177c27cba23107c7971ce2c5b2441d9ff2d1144" Dec 11 02:38:22 crc kubenswrapper[4824]: E1211 02:38:22.634428 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 02:38:37 crc kubenswrapper[4824]: I1211 02:38:37.634271 4824 scope.go:117] "RemoveContainer" containerID="592f00198a08d7f30f2bd4761177c27cba23107c7971ce2c5b2441d9ff2d1144" Dec 11 02:38:37 crc kubenswrapper[4824]: E1211 02:38:37.635492 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 02:38:49 crc kubenswrapper[4824]: I1211 02:38:49.633325 4824 scope.go:117] "RemoveContainer" containerID="592f00198a08d7f30f2bd4761177c27cba23107c7971ce2c5b2441d9ff2d1144" Dec 11 02:38:49 crc kubenswrapper[4824]: E1211 02:38:49.634557 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 02:38:59 crc kubenswrapper[4824]: I1211 02:38:59.083161 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8rbcc" event={"ID":"8736a409-d9d8-437e-965a-630f1ee6ae85","Type":"ContainerDied","Data":"55511b35433b1305df74789198b7bed5dc03e9fea6db02e2ebfc5c29d8ee92aa"} Dec 11 02:38:59 crc kubenswrapper[4824]: I1211 02:38:59.083095 4824 generic.go:334] "Generic (PLEG): container finished" podID="8736a409-d9d8-437e-965a-630f1ee6ae85" containerID="55511b35433b1305df74789198b7bed5dc03e9fea6db02e2ebfc5c29d8ee92aa" exitCode=0 Dec 11 02:39:00 crc kubenswrapper[4824]: I1211 02:39:00.513341 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8rbcc" Dec 11 02:39:00 crc kubenswrapper[4824]: I1211 02:39:00.626901 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v8zpk\" (UniqueName: \"kubernetes.io/projected/8736a409-d9d8-437e-965a-630f1ee6ae85-kube-api-access-v8zpk\") pod \"8736a409-d9d8-437e-965a-630f1ee6ae85\" (UID: \"8736a409-d9d8-437e-965a-630f1ee6ae85\") " Dec 11 02:39:00 crc kubenswrapper[4824]: I1211 02:39:00.627012 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8736a409-d9d8-437e-965a-630f1ee6ae85-inventory\") pod \"8736a409-d9d8-437e-965a-630f1ee6ae85\" (UID: \"8736a409-d9d8-437e-965a-630f1ee6ae85\") " Dec 11 02:39:00 crc kubenswrapper[4824]: I1211 02:39:00.627146 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/8736a409-d9d8-437e-965a-630f1ee6ae85-neutron-ovn-metadata-agent-neutron-config-0\") pod \"8736a409-d9d8-437e-965a-630f1ee6ae85\" (UID: \"8736a409-d9d8-437e-965a-630f1ee6ae85\") " Dec 11 02:39:00 crc kubenswrapper[4824]: I1211 02:39:00.627269 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8736a409-d9d8-437e-965a-630f1ee6ae85-ssh-key\") pod \"8736a409-d9d8-437e-965a-630f1ee6ae85\" (UID: \"8736a409-d9d8-437e-965a-630f1ee6ae85\") " Dec 11 02:39:00 crc kubenswrapper[4824]: I1211 02:39:00.627482 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8736a409-d9d8-437e-965a-630f1ee6ae85-neutron-metadata-combined-ca-bundle\") pod \"8736a409-d9d8-437e-965a-630f1ee6ae85\" (UID: \"8736a409-d9d8-437e-965a-630f1ee6ae85\") " Dec 11 02:39:00 crc kubenswrapper[4824]: I1211 02:39:00.627548 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/8736a409-d9d8-437e-965a-630f1ee6ae85-nova-metadata-neutron-config-0\") pod \"8736a409-d9d8-437e-965a-630f1ee6ae85\" (UID: \"8736a409-d9d8-437e-965a-630f1ee6ae85\") " Dec 11 02:39:00 crc kubenswrapper[4824]: I1211 02:39:00.635905 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8736a409-d9d8-437e-965a-630f1ee6ae85-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "8736a409-d9d8-437e-965a-630f1ee6ae85" (UID: "8736a409-d9d8-437e-965a-630f1ee6ae85"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:39:00 crc kubenswrapper[4824]: I1211 02:39:00.636048 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8736a409-d9d8-437e-965a-630f1ee6ae85-kube-api-access-v8zpk" (OuterVolumeSpecName: "kube-api-access-v8zpk") pod "8736a409-d9d8-437e-965a-630f1ee6ae85" (UID: "8736a409-d9d8-437e-965a-630f1ee6ae85"). InnerVolumeSpecName "kube-api-access-v8zpk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:39:00 crc kubenswrapper[4824]: I1211 02:39:00.657695 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8736a409-d9d8-437e-965a-630f1ee6ae85-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "8736a409-d9d8-437e-965a-630f1ee6ae85" (UID: "8736a409-d9d8-437e-965a-630f1ee6ae85"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:39:00 crc kubenswrapper[4824]: I1211 02:39:00.667354 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8736a409-d9d8-437e-965a-630f1ee6ae85-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "8736a409-d9d8-437e-965a-630f1ee6ae85" (UID: "8736a409-d9d8-437e-965a-630f1ee6ae85"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:39:00 crc kubenswrapper[4824]: E1211 02:39:00.675511 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8736a409-d9d8-437e-965a-630f1ee6ae85-nova-metadata-neutron-config-0 podName:8736a409-d9d8-437e-965a-630f1ee6ae85 nodeName:}" failed. No retries permitted until 2025-12-11 02:39:01.175477135 +0000 UTC m=+2282.864514534 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "nova-metadata-neutron-config-0" (UniqueName: "kubernetes.io/secret/8736a409-d9d8-437e-965a-630f1ee6ae85-nova-metadata-neutron-config-0") pod "8736a409-d9d8-437e-965a-630f1ee6ae85" (UID: "8736a409-d9d8-437e-965a-630f1ee6ae85") : error deleting /var/lib/kubelet/pods/8736a409-d9d8-437e-965a-630f1ee6ae85/volume-subpaths: remove /var/lib/kubelet/pods/8736a409-d9d8-437e-965a-630f1ee6ae85/volume-subpaths: no such file or directory Dec 11 02:39:00 crc kubenswrapper[4824]: I1211 02:39:00.678594 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8736a409-d9d8-437e-965a-630f1ee6ae85-inventory" (OuterVolumeSpecName: "inventory") pod "8736a409-d9d8-437e-965a-630f1ee6ae85" (UID: "8736a409-d9d8-437e-965a-630f1ee6ae85"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:39:00 crc kubenswrapper[4824]: I1211 02:39:00.730504 4824 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8736a409-d9d8-437e-965a-630f1ee6ae85-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 02:39:00 crc kubenswrapper[4824]: I1211 02:39:00.730529 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v8zpk\" (UniqueName: \"kubernetes.io/projected/8736a409-d9d8-437e-965a-630f1ee6ae85-kube-api-access-v8zpk\") on node \"crc\" DevicePath \"\"" Dec 11 02:39:00 crc kubenswrapper[4824]: I1211 02:39:00.730542 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8736a409-d9d8-437e-965a-630f1ee6ae85-inventory\") on node \"crc\" DevicePath \"\"" Dec 11 02:39:00 crc kubenswrapper[4824]: I1211 02:39:00.730553 4824 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/8736a409-d9d8-437e-965a-630f1ee6ae85-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 11 02:39:00 crc kubenswrapper[4824]: I1211 02:39:00.730563 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8736a409-d9d8-437e-965a-630f1ee6ae85-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 11 02:39:01 crc kubenswrapper[4824]: I1211 02:39:01.105825 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8rbcc" event={"ID":"8736a409-d9d8-437e-965a-630f1ee6ae85","Type":"ContainerDied","Data":"47ab5837ddd4ffc4eaab8e7553057ae72050638c5dbb33aca22a6108d2f3051d"} Dec 11 02:39:01 crc kubenswrapper[4824]: I1211 02:39:01.105895 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8rbcc" Dec 11 02:39:01 crc kubenswrapper[4824]: I1211 02:39:01.105899 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="47ab5837ddd4ffc4eaab8e7553057ae72050638c5dbb33aca22a6108d2f3051d" Dec 11 02:39:01 crc kubenswrapper[4824]: I1211 02:39:01.242234 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/8736a409-d9d8-437e-965a-630f1ee6ae85-nova-metadata-neutron-config-0\") pod \"8736a409-d9d8-437e-965a-630f1ee6ae85\" (UID: \"8736a409-d9d8-437e-965a-630f1ee6ae85\") " Dec 11 02:39:01 crc kubenswrapper[4824]: I1211 02:39:01.256479 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8736a409-d9d8-437e-965a-630f1ee6ae85-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "8736a409-d9d8-437e-965a-630f1ee6ae85" (UID: "8736a409-d9d8-437e-965a-630f1ee6ae85"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:39:01 crc kubenswrapper[4824]: I1211 02:39:01.273657 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-rs2vt"] Dec 11 02:39:01 crc kubenswrapper[4824]: E1211 02:39:01.274105 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8736a409-d9d8-437e-965a-630f1ee6ae85" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 11 02:39:01 crc kubenswrapper[4824]: I1211 02:39:01.274142 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="8736a409-d9d8-437e-965a-630f1ee6ae85" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 11 02:39:01 crc kubenswrapper[4824]: I1211 02:39:01.274388 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="8736a409-d9d8-437e-965a-630f1ee6ae85" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 11 02:39:01 crc kubenswrapper[4824]: I1211 02:39:01.275047 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-rs2vt" Dec 11 02:39:01 crc kubenswrapper[4824]: I1211 02:39:01.278649 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Dec 11 02:39:01 crc kubenswrapper[4824]: I1211 02:39:01.285788 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-rs2vt"] Dec 11 02:39:01 crc kubenswrapper[4824]: I1211 02:39:01.344766 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bc0c56ce-23bd-4883-b3dd-879d58c540b5-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-rs2vt\" (UID: \"bc0c56ce-23bd-4883-b3dd-879d58c540b5\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-rs2vt" Dec 11 02:39:01 crc kubenswrapper[4824]: I1211 02:39:01.345024 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc0c56ce-23bd-4883-b3dd-879d58c540b5-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-rs2vt\" (UID: \"bc0c56ce-23bd-4883-b3dd-879d58c540b5\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-rs2vt" Dec 11 02:39:01 crc kubenswrapper[4824]: I1211 02:39:01.345274 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/bc0c56ce-23bd-4883-b3dd-879d58c540b5-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-rs2vt\" (UID: \"bc0c56ce-23bd-4883-b3dd-879d58c540b5\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-rs2vt" Dec 11 02:39:01 crc kubenswrapper[4824]: I1211 02:39:01.345482 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bc0c56ce-23bd-4883-b3dd-879d58c540b5-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-rs2vt\" (UID: \"bc0c56ce-23bd-4883-b3dd-879d58c540b5\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-rs2vt" Dec 11 02:39:01 crc kubenswrapper[4824]: I1211 02:39:01.345611 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4g99r\" (UniqueName: \"kubernetes.io/projected/bc0c56ce-23bd-4883-b3dd-879d58c540b5-kube-api-access-4g99r\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-rs2vt\" (UID: \"bc0c56ce-23bd-4883-b3dd-879d58c540b5\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-rs2vt" Dec 11 02:39:01 crc kubenswrapper[4824]: I1211 02:39:01.345841 4824 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/8736a409-d9d8-437e-965a-630f1ee6ae85-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 11 02:39:01 crc kubenswrapper[4824]: I1211 02:39:01.447385 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4g99r\" (UniqueName: \"kubernetes.io/projected/bc0c56ce-23bd-4883-b3dd-879d58c540b5-kube-api-access-4g99r\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-rs2vt\" (UID: \"bc0c56ce-23bd-4883-b3dd-879d58c540b5\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-rs2vt" Dec 11 02:39:01 crc kubenswrapper[4824]: I1211 02:39:01.447700 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bc0c56ce-23bd-4883-b3dd-879d58c540b5-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-rs2vt\" (UID: \"bc0c56ce-23bd-4883-b3dd-879d58c540b5\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-rs2vt" Dec 11 02:39:01 crc kubenswrapper[4824]: I1211 02:39:01.447814 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc0c56ce-23bd-4883-b3dd-879d58c540b5-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-rs2vt\" (UID: \"bc0c56ce-23bd-4883-b3dd-879d58c540b5\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-rs2vt" Dec 11 02:39:01 crc kubenswrapper[4824]: I1211 02:39:01.447981 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/bc0c56ce-23bd-4883-b3dd-879d58c540b5-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-rs2vt\" (UID: \"bc0c56ce-23bd-4883-b3dd-879d58c540b5\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-rs2vt" Dec 11 02:39:01 crc kubenswrapper[4824]: I1211 02:39:01.448086 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bc0c56ce-23bd-4883-b3dd-879d58c540b5-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-rs2vt\" (UID: \"bc0c56ce-23bd-4883-b3dd-879d58c540b5\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-rs2vt" Dec 11 02:39:01 crc kubenswrapper[4824]: I1211 02:39:01.453253 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/bc0c56ce-23bd-4883-b3dd-879d58c540b5-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-rs2vt\" (UID: \"bc0c56ce-23bd-4883-b3dd-879d58c540b5\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-rs2vt" Dec 11 02:39:01 crc kubenswrapper[4824]: I1211 02:39:01.453405 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bc0c56ce-23bd-4883-b3dd-879d58c540b5-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-rs2vt\" (UID: \"bc0c56ce-23bd-4883-b3dd-879d58c540b5\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-rs2vt" Dec 11 02:39:01 crc kubenswrapper[4824]: I1211 02:39:01.456800 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bc0c56ce-23bd-4883-b3dd-879d58c540b5-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-rs2vt\" (UID: \"bc0c56ce-23bd-4883-b3dd-879d58c540b5\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-rs2vt" Dec 11 02:39:01 crc kubenswrapper[4824]: I1211 02:39:01.469413 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc0c56ce-23bd-4883-b3dd-879d58c540b5-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-rs2vt\" (UID: \"bc0c56ce-23bd-4883-b3dd-879d58c540b5\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-rs2vt" Dec 11 02:39:01 crc kubenswrapper[4824]: I1211 02:39:01.473185 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4g99r\" (UniqueName: \"kubernetes.io/projected/bc0c56ce-23bd-4883-b3dd-879d58c540b5-kube-api-access-4g99r\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-rs2vt\" (UID: \"bc0c56ce-23bd-4883-b3dd-879d58c540b5\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-rs2vt" Dec 11 02:39:01 crc kubenswrapper[4824]: I1211 02:39:01.617155 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-rs2vt" Dec 11 02:39:02 crc kubenswrapper[4824]: W1211 02:39:02.235198 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbc0c56ce_23bd_4883_b3dd_879d58c540b5.slice/crio-d2f1391f1843b70a5e71499d37aade96c398a57f20617c37173276cbaaee50b0 WatchSource:0}: Error finding container d2f1391f1843b70a5e71499d37aade96c398a57f20617c37173276cbaaee50b0: Status 404 returned error can't find the container with id d2f1391f1843b70a5e71499d37aade96c398a57f20617c37173276cbaaee50b0 Dec 11 02:39:02 crc kubenswrapper[4824]: I1211 02:39:02.248506 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-rs2vt"] Dec 11 02:39:03 crc kubenswrapper[4824]: I1211 02:39:03.132239 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-rs2vt" event={"ID":"bc0c56ce-23bd-4883-b3dd-879d58c540b5","Type":"ContainerStarted","Data":"c3e2f71a7459ce9e2e5bada6f1423f88116d8fe59308510ff2028de784c27f11"} Dec 11 02:39:03 crc kubenswrapper[4824]: I1211 02:39:03.132646 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-rs2vt" event={"ID":"bc0c56ce-23bd-4883-b3dd-879d58c540b5","Type":"ContainerStarted","Data":"d2f1391f1843b70a5e71499d37aade96c398a57f20617c37173276cbaaee50b0"} Dec 11 02:39:03 crc kubenswrapper[4824]: I1211 02:39:03.163142 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-rs2vt" podStartSLOduration=1.985230864 podStartE2EDuration="2.163094182s" podCreationTimestamp="2025-12-11 02:39:01 +0000 UTC" firstStartedPulling="2025-12-11 02:39:02.237596667 +0000 UTC m=+2283.926634046" lastFinishedPulling="2025-12-11 02:39:02.415459975 +0000 UTC m=+2284.104497364" observedRunningTime="2025-12-11 02:39:03.159751813 +0000 UTC m=+2284.848789212" watchObservedRunningTime="2025-12-11 02:39:03.163094182 +0000 UTC m=+2284.852131601" Dec 11 02:39:04 crc kubenswrapper[4824]: I1211 02:39:04.633565 4824 scope.go:117] "RemoveContainer" containerID="592f00198a08d7f30f2bd4761177c27cba23107c7971ce2c5b2441d9ff2d1144" Dec 11 02:39:04 crc kubenswrapper[4824]: E1211 02:39:04.634415 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 02:39:16 crc kubenswrapper[4824]: I1211 02:39:16.634223 4824 scope.go:117] "RemoveContainer" containerID="592f00198a08d7f30f2bd4761177c27cba23107c7971ce2c5b2441d9ff2d1144" Dec 11 02:39:16 crc kubenswrapper[4824]: E1211 02:39:16.635435 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 02:39:31 crc kubenswrapper[4824]: I1211 02:39:31.632402 4824 scope.go:117] "RemoveContainer" containerID="592f00198a08d7f30f2bd4761177c27cba23107c7971ce2c5b2441d9ff2d1144" Dec 11 02:39:31 crc kubenswrapper[4824]: E1211 02:39:31.633171 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 02:39:46 crc kubenswrapper[4824]: I1211 02:39:46.633103 4824 scope.go:117] "RemoveContainer" containerID="592f00198a08d7f30f2bd4761177c27cba23107c7971ce2c5b2441d9ff2d1144" Dec 11 02:39:46 crc kubenswrapper[4824]: E1211 02:39:46.634546 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 02:40:00 crc kubenswrapper[4824]: I1211 02:40:00.633187 4824 scope.go:117] "RemoveContainer" containerID="592f00198a08d7f30f2bd4761177c27cba23107c7971ce2c5b2441d9ff2d1144" Dec 11 02:40:00 crc kubenswrapper[4824]: E1211 02:40:00.634433 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 02:40:14 crc kubenswrapper[4824]: I1211 02:40:14.632813 4824 scope.go:117] "RemoveContainer" containerID="592f00198a08d7f30f2bd4761177c27cba23107c7971ce2c5b2441d9ff2d1144" Dec 11 02:40:14 crc kubenswrapper[4824]: E1211 02:40:14.634186 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 02:40:29 crc kubenswrapper[4824]: I1211 02:40:29.632463 4824 scope.go:117] "RemoveContainer" containerID="592f00198a08d7f30f2bd4761177c27cba23107c7971ce2c5b2441d9ff2d1144" Dec 11 02:40:29 crc kubenswrapper[4824]: E1211 02:40:29.633346 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 02:40:41 crc kubenswrapper[4824]: I1211 02:40:41.633083 4824 scope.go:117] "RemoveContainer" containerID="592f00198a08d7f30f2bd4761177c27cba23107c7971ce2c5b2441d9ff2d1144" Dec 11 02:40:41 crc kubenswrapper[4824]: E1211 02:40:41.634300 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 02:40:55 crc kubenswrapper[4824]: I1211 02:40:55.633684 4824 scope.go:117] "RemoveContainer" containerID="592f00198a08d7f30f2bd4761177c27cba23107c7971ce2c5b2441d9ff2d1144" Dec 11 02:40:55 crc kubenswrapper[4824]: E1211 02:40:55.634797 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 02:41:06 crc kubenswrapper[4824]: I1211 02:41:06.634468 4824 scope.go:117] "RemoveContainer" containerID="592f00198a08d7f30f2bd4761177c27cba23107c7971ce2c5b2441d9ff2d1144" Dec 11 02:41:06 crc kubenswrapper[4824]: E1211 02:41:06.635279 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 02:41:17 crc kubenswrapper[4824]: I1211 02:41:17.633217 4824 scope.go:117] "RemoveContainer" containerID="592f00198a08d7f30f2bd4761177c27cba23107c7971ce2c5b2441d9ff2d1144" Dec 11 02:41:17 crc kubenswrapper[4824]: E1211 02:41:17.634536 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 02:41:29 crc kubenswrapper[4824]: I1211 02:41:29.632414 4824 scope.go:117] "RemoveContainer" containerID="592f00198a08d7f30f2bd4761177c27cba23107c7971ce2c5b2441d9ff2d1144" Dec 11 02:41:29 crc kubenswrapper[4824]: E1211 02:41:29.633404 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 02:41:41 crc kubenswrapper[4824]: I1211 02:41:41.633377 4824 scope.go:117] "RemoveContainer" containerID="592f00198a08d7f30f2bd4761177c27cba23107c7971ce2c5b2441d9ff2d1144" Dec 11 02:41:41 crc kubenswrapper[4824]: E1211 02:41:41.634391 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 02:41:53 crc kubenswrapper[4824]: I1211 02:41:53.632683 4824 scope.go:117] "RemoveContainer" containerID="592f00198a08d7f30f2bd4761177c27cba23107c7971ce2c5b2441d9ff2d1144" Dec 11 02:41:53 crc kubenswrapper[4824]: E1211 02:41:53.633542 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 02:42:04 crc kubenswrapper[4824]: I1211 02:42:04.632449 4824 scope.go:117] "RemoveContainer" containerID="592f00198a08d7f30f2bd4761177c27cba23107c7971ce2c5b2441d9ff2d1144" Dec 11 02:42:04 crc kubenswrapper[4824]: E1211 02:42:04.633145 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 02:42:16 crc kubenswrapper[4824]: I1211 02:42:16.634729 4824 scope.go:117] "RemoveContainer" containerID="592f00198a08d7f30f2bd4761177c27cba23107c7971ce2c5b2441d9ff2d1144" Dec 11 02:42:16 crc kubenswrapper[4824]: E1211 02:42:16.635938 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 02:42:31 crc kubenswrapper[4824]: I1211 02:42:31.632731 4824 scope.go:117] "RemoveContainer" containerID="592f00198a08d7f30f2bd4761177c27cba23107c7971ce2c5b2441d9ff2d1144" Dec 11 02:42:31 crc kubenswrapper[4824]: E1211 02:42:31.633655 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 02:42:45 crc kubenswrapper[4824]: I1211 02:42:45.634787 4824 scope.go:117] "RemoveContainer" containerID="592f00198a08d7f30f2bd4761177c27cba23107c7971ce2c5b2441d9ff2d1144" Dec 11 02:42:46 crc kubenswrapper[4824]: I1211 02:42:46.690831 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" event={"ID":"44b156e4-64a4-4d45-aa5e-9b10a862faed","Type":"ContainerStarted","Data":"4d2b52538d7857fd4c0ee38baaab7aec6956bb5f853fab54c604059ae739842d"} Dec 11 02:43:32 crc kubenswrapper[4824]: I1211 02:43:32.235764 4824 generic.go:334] "Generic (PLEG): container finished" podID="bc0c56ce-23bd-4883-b3dd-879d58c540b5" containerID="c3e2f71a7459ce9e2e5bada6f1423f88116d8fe59308510ff2028de784c27f11" exitCode=0 Dec 11 02:43:32 crc kubenswrapper[4824]: I1211 02:43:32.235849 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-rs2vt" event={"ID":"bc0c56ce-23bd-4883-b3dd-879d58c540b5","Type":"ContainerDied","Data":"c3e2f71a7459ce9e2e5bada6f1423f88116d8fe59308510ff2028de784c27f11"} Dec 11 02:43:33 crc kubenswrapper[4824]: I1211 02:43:33.789192 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-rs2vt" Dec 11 02:43:33 crc kubenswrapper[4824]: I1211 02:43:33.979770 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bc0c56ce-23bd-4883-b3dd-879d58c540b5-ssh-key\") pod \"bc0c56ce-23bd-4883-b3dd-879d58c540b5\" (UID: \"bc0c56ce-23bd-4883-b3dd-879d58c540b5\") " Dec 11 02:43:33 crc kubenswrapper[4824]: I1211 02:43:33.979914 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4g99r\" (UniqueName: \"kubernetes.io/projected/bc0c56ce-23bd-4883-b3dd-879d58c540b5-kube-api-access-4g99r\") pod \"bc0c56ce-23bd-4883-b3dd-879d58c540b5\" (UID: \"bc0c56ce-23bd-4883-b3dd-879d58c540b5\") " Dec 11 02:43:33 crc kubenswrapper[4824]: I1211 02:43:33.979965 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc0c56ce-23bd-4883-b3dd-879d58c540b5-libvirt-combined-ca-bundle\") pod \"bc0c56ce-23bd-4883-b3dd-879d58c540b5\" (UID: \"bc0c56ce-23bd-4883-b3dd-879d58c540b5\") " Dec 11 02:43:33 crc kubenswrapper[4824]: I1211 02:43:33.979996 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/bc0c56ce-23bd-4883-b3dd-879d58c540b5-libvirt-secret-0\") pod \"bc0c56ce-23bd-4883-b3dd-879d58c540b5\" (UID: \"bc0c56ce-23bd-4883-b3dd-879d58c540b5\") " Dec 11 02:43:33 crc kubenswrapper[4824]: I1211 02:43:33.980180 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bc0c56ce-23bd-4883-b3dd-879d58c540b5-inventory\") pod \"bc0c56ce-23bd-4883-b3dd-879d58c540b5\" (UID: \"bc0c56ce-23bd-4883-b3dd-879d58c540b5\") " Dec 11 02:43:33 crc kubenswrapper[4824]: I1211 02:43:33.988721 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc0c56ce-23bd-4883-b3dd-879d58c540b5-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "bc0c56ce-23bd-4883-b3dd-879d58c540b5" (UID: "bc0c56ce-23bd-4883-b3dd-879d58c540b5"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:43:33 crc kubenswrapper[4824]: I1211 02:43:33.989043 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc0c56ce-23bd-4883-b3dd-879d58c540b5-kube-api-access-4g99r" (OuterVolumeSpecName: "kube-api-access-4g99r") pod "bc0c56ce-23bd-4883-b3dd-879d58c540b5" (UID: "bc0c56ce-23bd-4883-b3dd-879d58c540b5"). InnerVolumeSpecName "kube-api-access-4g99r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:43:34 crc kubenswrapper[4824]: I1211 02:43:34.012961 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc0c56ce-23bd-4883-b3dd-879d58c540b5-inventory" (OuterVolumeSpecName: "inventory") pod "bc0c56ce-23bd-4883-b3dd-879d58c540b5" (UID: "bc0c56ce-23bd-4883-b3dd-879d58c540b5"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:43:34 crc kubenswrapper[4824]: I1211 02:43:34.021556 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc0c56ce-23bd-4883-b3dd-879d58c540b5-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "bc0c56ce-23bd-4883-b3dd-879d58c540b5" (UID: "bc0c56ce-23bd-4883-b3dd-879d58c540b5"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:43:34 crc kubenswrapper[4824]: I1211 02:43:34.042287 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc0c56ce-23bd-4883-b3dd-879d58c540b5-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "bc0c56ce-23bd-4883-b3dd-879d58c540b5" (UID: "bc0c56ce-23bd-4883-b3dd-879d58c540b5"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:43:34 crc kubenswrapper[4824]: I1211 02:43:34.082626 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4g99r\" (UniqueName: \"kubernetes.io/projected/bc0c56ce-23bd-4883-b3dd-879d58c540b5-kube-api-access-4g99r\") on node \"crc\" DevicePath \"\"" Dec 11 02:43:34 crc kubenswrapper[4824]: I1211 02:43:34.082895 4824 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc0c56ce-23bd-4883-b3dd-879d58c540b5-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 02:43:34 crc kubenswrapper[4824]: I1211 02:43:34.083023 4824 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/bc0c56ce-23bd-4883-b3dd-879d58c540b5-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Dec 11 02:43:34 crc kubenswrapper[4824]: I1211 02:43:34.083269 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bc0c56ce-23bd-4883-b3dd-879d58c540b5-inventory\") on node \"crc\" DevicePath \"\"" Dec 11 02:43:34 crc kubenswrapper[4824]: I1211 02:43:34.083402 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bc0c56ce-23bd-4883-b3dd-879d58c540b5-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 11 02:43:34 crc kubenswrapper[4824]: I1211 02:43:34.259314 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-rs2vt" event={"ID":"bc0c56ce-23bd-4883-b3dd-879d58c540b5","Type":"ContainerDied","Data":"d2f1391f1843b70a5e71499d37aade96c398a57f20617c37173276cbaaee50b0"} Dec 11 02:43:34 crc kubenswrapper[4824]: I1211 02:43:34.259632 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d2f1391f1843b70a5e71499d37aade96c398a57f20617c37173276cbaaee50b0" Dec 11 02:43:34 crc kubenswrapper[4824]: I1211 02:43:34.259388 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-rs2vt" Dec 11 02:43:34 crc kubenswrapper[4824]: I1211 02:43:34.383474 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-7x98s"] Dec 11 02:43:34 crc kubenswrapper[4824]: E1211 02:43:34.383989 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc0c56ce-23bd-4883-b3dd-879d58c540b5" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 11 02:43:34 crc kubenswrapper[4824]: I1211 02:43:34.384027 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc0c56ce-23bd-4883-b3dd-879d58c540b5" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 11 02:43:34 crc kubenswrapper[4824]: I1211 02:43:34.384280 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc0c56ce-23bd-4883-b3dd-879d58c540b5" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 11 02:43:34 crc kubenswrapper[4824]: I1211 02:43:34.385032 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7x98s" Dec 11 02:43:34 crc kubenswrapper[4824]: I1211 02:43:34.387640 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Dec 11 02:43:34 crc kubenswrapper[4824]: I1211 02:43:34.387852 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Dec 11 02:43:34 crc kubenswrapper[4824]: I1211 02:43:34.387863 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-hs2fv" Dec 11 02:43:34 crc kubenswrapper[4824]: I1211 02:43:34.388359 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 11 02:43:34 crc kubenswrapper[4824]: I1211 02:43:34.388428 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Dec 11 02:43:34 crc kubenswrapper[4824]: I1211 02:43:34.388736 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 11 02:43:34 crc kubenswrapper[4824]: I1211 02:43:34.391423 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-7x98s"] Dec 11 02:43:34 crc kubenswrapper[4824]: I1211 02:43:34.400632 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 11 02:43:34 crc kubenswrapper[4824]: I1211 02:43:34.490713 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/f844f569-89fd-410c-b2e1-5f9883d9127a-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7x98s\" (UID: \"f844f569-89fd-410c-b2e1-5f9883d9127a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7x98s" Dec 11 02:43:34 crc kubenswrapper[4824]: I1211 02:43:34.491060 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f844f569-89fd-410c-b2e1-5f9883d9127a-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7x98s\" (UID: \"f844f569-89fd-410c-b2e1-5f9883d9127a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7x98s" Dec 11 02:43:34 crc kubenswrapper[4824]: I1211 02:43:34.491248 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/f844f569-89fd-410c-b2e1-5f9883d9127a-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7x98s\" (UID: \"f844f569-89fd-410c-b2e1-5f9883d9127a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7x98s" Dec 11 02:43:34 crc kubenswrapper[4824]: I1211 02:43:34.491421 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f844f569-89fd-410c-b2e1-5f9883d9127a-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7x98s\" (UID: \"f844f569-89fd-410c-b2e1-5f9883d9127a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7x98s" Dec 11 02:43:34 crc kubenswrapper[4824]: I1211 02:43:34.491580 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hrkzr\" (UniqueName: \"kubernetes.io/projected/f844f569-89fd-410c-b2e1-5f9883d9127a-kube-api-access-hrkzr\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7x98s\" (UID: \"f844f569-89fd-410c-b2e1-5f9883d9127a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7x98s" Dec 11 02:43:34 crc kubenswrapper[4824]: I1211 02:43:34.491751 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f844f569-89fd-410c-b2e1-5f9883d9127a-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7x98s\" (UID: \"f844f569-89fd-410c-b2e1-5f9883d9127a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7x98s" Dec 11 02:43:34 crc kubenswrapper[4824]: I1211 02:43:34.491960 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/f844f569-89fd-410c-b2e1-5f9883d9127a-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7x98s\" (UID: \"f844f569-89fd-410c-b2e1-5f9883d9127a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7x98s" Dec 11 02:43:34 crc kubenswrapper[4824]: I1211 02:43:34.492094 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/f844f569-89fd-410c-b2e1-5f9883d9127a-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7x98s\" (UID: \"f844f569-89fd-410c-b2e1-5f9883d9127a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7x98s" Dec 11 02:43:34 crc kubenswrapper[4824]: I1211 02:43:34.492263 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/f844f569-89fd-410c-b2e1-5f9883d9127a-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7x98s\" (UID: \"f844f569-89fd-410c-b2e1-5f9883d9127a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7x98s" Dec 11 02:43:34 crc kubenswrapper[4824]: I1211 02:43:34.594362 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/f844f569-89fd-410c-b2e1-5f9883d9127a-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7x98s\" (UID: \"f844f569-89fd-410c-b2e1-5f9883d9127a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7x98s" Dec 11 02:43:34 crc kubenswrapper[4824]: I1211 02:43:34.594675 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/f844f569-89fd-410c-b2e1-5f9883d9127a-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7x98s\" (UID: \"f844f569-89fd-410c-b2e1-5f9883d9127a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7x98s" Dec 11 02:43:34 crc kubenswrapper[4824]: I1211 02:43:34.594841 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/f844f569-89fd-410c-b2e1-5f9883d9127a-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7x98s\" (UID: \"f844f569-89fd-410c-b2e1-5f9883d9127a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7x98s" Dec 11 02:43:34 crc kubenswrapper[4824]: I1211 02:43:34.594999 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/f844f569-89fd-410c-b2e1-5f9883d9127a-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7x98s\" (UID: \"f844f569-89fd-410c-b2e1-5f9883d9127a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7x98s" Dec 11 02:43:34 crc kubenswrapper[4824]: I1211 02:43:34.595379 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f844f569-89fd-410c-b2e1-5f9883d9127a-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7x98s\" (UID: \"f844f569-89fd-410c-b2e1-5f9883d9127a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7x98s" Dec 11 02:43:34 crc kubenswrapper[4824]: I1211 02:43:34.613259 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/f844f569-89fd-410c-b2e1-5f9883d9127a-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7x98s\" (UID: \"f844f569-89fd-410c-b2e1-5f9883d9127a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7x98s" Dec 11 02:43:34 crc kubenswrapper[4824]: I1211 02:43:34.613929 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/f844f569-89fd-410c-b2e1-5f9883d9127a-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7x98s\" (UID: \"f844f569-89fd-410c-b2e1-5f9883d9127a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7x98s" Dec 11 02:43:34 crc kubenswrapper[4824]: I1211 02:43:34.614558 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/f844f569-89fd-410c-b2e1-5f9883d9127a-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7x98s\" (UID: \"f844f569-89fd-410c-b2e1-5f9883d9127a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7x98s" Dec 11 02:43:34 crc kubenswrapper[4824]: I1211 02:43:34.615139 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/f844f569-89fd-410c-b2e1-5f9883d9127a-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7x98s\" (UID: \"f844f569-89fd-410c-b2e1-5f9883d9127a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7x98s" Dec 11 02:43:34 crc kubenswrapper[4824]: I1211 02:43:34.595530 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/f844f569-89fd-410c-b2e1-5f9883d9127a-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7x98s\" (UID: \"f844f569-89fd-410c-b2e1-5f9883d9127a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7x98s" Dec 11 02:43:34 crc kubenswrapper[4824]: I1211 02:43:34.615513 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f844f569-89fd-410c-b2e1-5f9883d9127a-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7x98s\" (UID: \"f844f569-89fd-410c-b2e1-5f9883d9127a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7x98s" Dec 11 02:43:34 crc kubenswrapper[4824]: I1211 02:43:34.615610 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hrkzr\" (UniqueName: \"kubernetes.io/projected/f844f569-89fd-410c-b2e1-5f9883d9127a-kube-api-access-hrkzr\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7x98s\" (UID: \"f844f569-89fd-410c-b2e1-5f9883d9127a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7x98s" Dec 11 02:43:34 crc kubenswrapper[4824]: I1211 02:43:34.615731 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f844f569-89fd-410c-b2e1-5f9883d9127a-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7x98s\" (UID: \"f844f569-89fd-410c-b2e1-5f9883d9127a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7x98s" Dec 11 02:43:34 crc kubenswrapper[4824]: I1211 02:43:34.617066 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f844f569-89fd-410c-b2e1-5f9883d9127a-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7x98s\" (UID: \"f844f569-89fd-410c-b2e1-5f9883d9127a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7x98s" Dec 11 02:43:34 crc kubenswrapper[4824]: I1211 02:43:34.620069 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f844f569-89fd-410c-b2e1-5f9883d9127a-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7x98s\" (UID: \"f844f569-89fd-410c-b2e1-5f9883d9127a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7x98s" Dec 11 02:43:34 crc kubenswrapper[4824]: I1211 02:43:34.627543 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/f844f569-89fd-410c-b2e1-5f9883d9127a-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7x98s\" (UID: \"f844f569-89fd-410c-b2e1-5f9883d9127a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7x98s" Dec 11 02:43:34 crc kubenswrapper[4824]: I1211 02:43:34.630868 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f844f569-89fd-410c-b2e1-5f9883d9127a-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7x98s\" (UID: \"f844f569-89fd-410c-b2e1-5f9883d9127a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7x98s" Dec 11 02:43:34 crc kubenswrapper[4824]: I1211 02:43:34.635552 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hrkzr\" (UniqueName: \"kubernetes.io/projected/f844f569-89fd-410c-b2e1-5f9883d9127a-kube-api-access-hrkzr\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7x98s\" (UID: \"f844f569-89fd-410c-b2e1-5f9883d9127a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7x98s" Dec 11 02:43:34 crc kubenswrapper[4824]: I1211 02:43:34.715752 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7x98s" Dec 11 02:43:35 crc kubenswrapper[4824]: I1211 02:43:35.237452 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-7x98s"] Dec 11 02:43:35 crc kubenswrapper[4824]: I1211 02:43:35.245390 4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 11 02:43:35 crc kubenswrapper[4824]: I1211 02:43:35.275602 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7x98s" event={"ID":"f844f569-89fd-410c-b2e1-5f9883d9127a","Type":"ContainerStarted","Data":"f4c3c32f1d4f2e6a2e6d615b28fa257f1f82536af9c0bd7c46b3c9bd965c7296"} Dec 11 02:43:36 crc kubenswrapper[4824]: I1211 02:43:36.287561 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7x98s" event={"ID":"f844f569-89fd-410c-b2e1-5f9883d9127a","Type":"ContainerStarted","Data":"d6fc1348714e08242892b205d2c6ee536f9f2a1c7a00a9e0a696dfbc93bb9c78"} Dec 11 02:43:36 crc kubenswrapper[4824]: I1211 02:43:36.305860 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7x98s" podStartSLOduration=2.099437969 podStartE2EDuration="2.305838276s" podCreationTimestamp="2025-12-11 02:43:34 +0000 UTC" firstStartedPulling="2025-12-11 02:43:35.244912409 +0000 UTC m=+2556.933949828" lastFinishedPulling="2025-12-11 02:43:35.451312756 +0000 UTC m=+2557.140350135" observedRunningTime="2025-12-11 02:43:36.303660809 +0000 UTC m=+2557.992698198" watchObservedRunningTime="2025-12-11 02:43:36.305838276 +0000 UTC m=+2557.994875675" Dec 11 02:45:00 crc kubenswrapper[4824]: I1211 02:45:00.462404 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423685-cwtf6"] Dec 11 02:45:00 crc kubenswrapper[4824]: I1211 02:45:00.465175 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423685-cwtf6" Dec 11 02:45:00 crc kubenswrapper[4824]: I1211 02:45:00.469877 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423685-cwtf6"] Dec 11 02:45:00 crc kubenswrapper[4824]: I1211 02:45:00.502264 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 11 02:45:00 crc kubenswrapper[4824]: I1211 02:45:00.502790 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 11 02:45:00 crc kubenswrapper[4824]: I1211 02:45:00.552750 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9drt7\" (UniqueName: \"kubernetes.io/projected/36e68cbe-0b5d-444d-813e-b0e003b9de37-kube-api-access-9drt7\") pod \"collect-profiles-29423685-cwtf6\" (UID: \"36e68cbe-0b5d-444d-813e-b0e003b9de37\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423685-cwtf6" Dec 11 02:45:00 crc kubenswrapper[4824]: I1211 02:45:00.552829 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/36e68cbe-0b5d-444d-813e-b0e003b9de37-secret-volume\") pod \"collect-profiles-29423685-cwtf6\" (UID: \"36e68cbe-0b5d-444d-813e-b0e003b9de37\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423685-cwtf6" Dec 11 02:45:00 crc kubenswrapper[4824]: I1211 02:45:00.552921 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/36e68cbe-0b5d-444d-813e-b0e003b9de37-config-volume\") pod \"collect-profiles-29423685-cwtf6\" (UID: \"36e68cbe-0b5d-444d-813e-b0e003b9de37\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423685-cwtf6" Dec 11 02:45:00 crc kubenswrapper[4824]: I1211 02:45:00.654475 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/36e68cbe-0b5d-444d-813e-b0e003b9de37-config-volume\") pod \"collect-profiles-29423685-cwtf6\" (UID: \"36e68cbe-0b5d-444d-813e-b0e003b9de37\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423685-cwtf6" Dec 11 02:45:00 crc kubenswrapper[4824]: I1211 02:45:00.654559 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9drt7\" (UniqueName: \"kubernetes.io/projected/36e68cbe-0b5d-444d-813e-b0e003b9de37-kube-api-access-9drt7\") pod \"collect-profiles-29423685-cwtf6\" (UID: \"36e68cbe-0b5d-444d-813e-b0e003b9de37\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423685-cwtf6" Dec 11 02:45:00 crc kubenswrapper[4824]: I1211 02:45:00.654607 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/36e68cbe-0b5d-444d-813e-b0e003b9de37-secret-volume\") pod \"collect-profiles-29423685-cwtf6\" (UID: \"36e68cbe-0b5d-444d-813e-b0e003b9de37\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423685-cwtf6" Dec 11 02:45:00 crc kubenswrapper[4824]: I1211 02:45:00.656189 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/36e68cbe-0b5d-444d-813e-b0e003b9de37-config-volume\") pod \"collect-profiles-29423685-cwtf6\" (UID: \"36e68cbe-0b5d-444d-813e-b0e003b9de37\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423685-cwtf6" Dec 11 02:45:00 crc kubenswrapper[4824]: I1211 02:45:00.665296 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/36e68cbe-0b5d-444d-813e-b0e003b9de37-secret-volume\") pod \"collect-profiles-29423685-cwtf6\" (UID: \"36e68cbe-0b5d-444d-813e-b0e003b9de37\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423685-cwtf6" Dec 11 02:45:00 crc kubenswrapper[4824]: I1211 02:45:00.672289 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9drt7\" (UniqueName: \"kubernetes.io/projected/36e68cbe-0b5d-444d-813e-b0e003b9de37-kube-api-access-9drt7\") pod \"collect-profiles-29423685-cwtf6\" (UID: \"36e68cbe-0b5d-444d-813e-b0e003b9de37\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423685-cwtf6" Dec 11 02:45:00 crc kubenswrapper[4824]: I1211 02:45:00.823619 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423685-cwtf6" Dec 11 02:45:01 crc kubenswrapper[4824]: I1211 02:45:01.337306 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423685-cwtf6"] Dec 11 02:45:01 crc kubenswrapper[4824]: W1211 02:45:01.340265 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod36e68cbe_0b5d_444d_813e_b0e003b9de37.slice/crio-e6b45e2b761ddc181db20ef6b6139c9332ccb6c23744d03ac76b3dd6868f5da5 WatchSource:0}: Error finding container e6b45e2b761ddc181db20ef6b6139c9332ccb6c23744d03ac76b3dd6868f5da5: Status 404 returned error can't find the container with id e6b45e2b761ddc181db20ef6b6139c9332ccb6c23744d03ac76b3dd6868f5da5 Dec 11 02:45:01 crc kubenswrapper[4824]: I1211 02:45:01.426774 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423685-cwtf6" event={"ID":"36e68cbe-0b5d-444d-813e-b0e003b9de37","Type":"ContainerStarted","Data":"e6b45e2b761ddc181db20ef6b6139c9332ccb6c23744d03ac76b3dd6868f5da5"} Dec 11 02:45:02 crc kubenswrapper[4824]: I1211 02:45:02.435458 4824 generic.go:334] "Generic (PLEG): container finished" podID="36e68cbe-0b5d-444d-813e-b0e003b9de37" containerID="e2ebc78d22a1a9e70deb4a423472afde25a429236313ba18150c1c51628b06a0" exitCode=0 Dec 11 02:45:02 crc kubenswrapper[4824]: I1211 02:45:02.435651 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423685-cwtf6" event={"ID":"36e68cbe-0b5d-444d-813e-b0e003b9de37","Type":"ContainerDied","Data":"e2ebc78d22a1a9e70deb4a423472afde25a429236313ba18150c1c51628b06a0"} Dec 11 02:45:03 crc kubenswrapper[4824]: I1211 02:45:03.806191 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423685-cwtf6" Dec 11 02:45:03 crc kubenswrapper[4824]: I1211 02:45:03.837516 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/36e68cbe-0b5d-444d-813e-b0e003b9de37-config-volume\") pod \"36e68cbe-0b5d-444d-813e-b0e003b9de37\" (UID: \"36e68cbe-0b5d-444d-813e-b0e003b9de37\") " Dec 11 02:45:03 crc kubenswrapper[4824]: I1211 02:45:03.837652 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9drt7\" (UniqueName: \"kubernetes.io/projected/36e68cbe-0b5d-444d-813e-b0e003b9de37-kube-api-access-9drt7\") pod \"36e68cbe-0b5d-444d-813e-b0e003b9de37\" (UID: \"36e68cbe-0b5d-444d-813e-b0e003b9de37\") " Dec 11 02:45:03 crc kubenswrapper[4824]: I1211 02:45:03.837821 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/36e68cbe-0b5d-444d-813e-b0e003b9de37-secret-volume\") pod \"36e68cbe-0b5d-444d-813e-b0e003b9de37\" (UID: \"36e68cbe-0b5d-444d-813e-b0e003b9de37\") " Dec 11 02:45:03 crc kubenswrapper[4824]: I1211 02:45:03.838279 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/36e68cbe-0b5d-444d-813e-b0e003b9de37-config-volume" (OuterVolumeSpecName: "config-volume") pod "36e68cbe-0b5d-444d-813e-b0e003b9de37" (UID: "36e68cbe-0b5d-444d-813e-b0e003b9de37"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:45:03 crc kubenswrapper[4824]: I1211 02:45:03.838661 4824 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/36e68cbe-0b5d-444d-813e-b0e003b9de37-config-volume\") on node \"crc\" DevicePath \"\"" Dec 11 02:45:03 crc kubenswrapper[4824]: I1211 02:45:03.843835 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36e68cbe-0b5d-444d-813e-b0e003b9de37-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "36e68cbe-0b5d-444d-813e-b0e003b9de37" (UID: "36e68cbe-0b5d-444d-813e-b0e003b9de37"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:45:03 crc kubenswrapper[4824]: I1211 02:45:03.844294 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36e68cbe-0b5d-444d-813e-b0e003b9de37-kube-api-access-9drt7" (OuterVolumeSpecName: "kube-api-access-9drt7") pod "36e68cbe-0b5d-444d-813e-b0e003b9de37" (UID: "36e68cbe-0b5d-444d-813e-b0e003b9de37"). InnerVolumeSpecName "kube-api-access-9drt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:45:03 crc kubenswrapper[4824]: I1211 02:45:03.940137 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9drt7\" (UniqueName: \"kubernetes.io/projected/36e68cbe-0b5d-444d-813e-b0e003b9de37-kube-api-access-9drt7\") on node \"crc\" DevicePath \"\"" Dec 11 02:45:03 crc kubenswrapper[4824]: I1211 02:45:03.940175 4824 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/36e68cbe-0b5d-444d-813e-b0e003b9de37-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 11 02:45:04 crc kubenswrapper[4824]: I1211 02:45:04.459802 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423685-cwtf6" event={"ID":"36e68cbe-0b5d-444d-813e-b0e003b9de37","Type":"ContainerDied","Data":"e6b45e2b761ddc181db20ef6b6139c9332ccb6c23744d03ac76b3dd6868f5da5"} Dec 11 02:45:04 crc kubenswrapper[4824]: I1211 02:45:04.460160 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e6b45e2b761ddc181db20ef6b6139c9332ccb6c23744d03ac76b3dd6868f5da5" Dec 11 02:45:04 crc kubenswrapper[4824]: I1211 02:45:04.459863 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423685-cwtf6" Dec 11 02:45:04 crc kubenswrapper[4824]: I1211 02:45:04.878173 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423640-42vd6"] Dec 11 02:45:04 crc kubenswrapper[4824]: I1211 02:45:04.885803 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423640-42vd6"] Dec 11 02:45:06 crc kubenswrapper[4824]: I1211 02:45:06.644524 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c22bf31-ad33-44cd-a495-05d83688f2bd" path="/var/lib/kubelet/pods/4c22bf31-ad33-44cd-a495-05d83688f2bd/volumes" Dec 11 02:45:07 crc kubenswrapper[4824]: I1211 02:45:07.250894 4824 patch_prober.go:28] interesting pod/machine-config-daemon-gx6xt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 02:45:07 crc kubenswrapper[4824]: I1211 02:45:07.251034 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 02:45:37 crc kubenswrapper[4824]: I1211 02:45:37.250648 4824 patch_prober.go:28] interesting pod/machine-config-daemon-gx6xt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 02:45:37 crc kubenswrapper[4824]: I1211 02:45:37.251269 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 02:46:05 crc kubenswrapper[4824]: I1211 02:46:05.010596 4824 scope.go:117] "RemoveContainer" containerID="f63691527d0e51385a07c889a507a7c4d8c0a08751f37f06bdd5379d9c81e323" Dec 11 02:46:05 crc kubenswrapper[4824]: I1211 02:46:05.873227 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-5qtbv"] Dec 11 02:46:05 crc kubenswrapper[4824]: E1211 02:46:05.873905 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36e68cbe-0b5d-444d-813e-b0e003b9de37" containerName="collect-profiles" Dec 11 02:46:05 crc kubenswrapper[4824]: I1211 02:46:05.873923 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="36e68cbe-0b5d-444d-813e-b0e003b9de37" containerName="collect-profiles" Dec 11 02:46:05 crc kubenswrapper[4824]: I1211 02:46:05.874139 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="36e68cbe-0b5d-444d-813e-b0e003b9de37" containerName="collect-profiles" Dec 11 02:46:05 crc kubenswrapper[4824]: I1211 02:46:05.875419 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5qtbv" Dec 11 02:46:05 crc kubenswrapper[4824]: I1211 02:46:05.885236 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5qtbv"] Dec 11 02:46:05 crc kubenswrapper[4824]: I1211 02:46:05.927005 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f994800f-7f5f-41ca-87d3-df2121b8890f-utilities\") pod \"redhat-marketplace-5qtbv\" (UID: \"f994800f-7f5f-41ca-87d3-df2121b8890f\") " pod="openshift-marketplace/redhat-marketplace-5qtbv" Dec 11 02:46:05 crc kubenswrapper[4824]: I1211 02:46:05.927269 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f994800f-7f5f-41ca-87d3-df2121b8890f-catalog-content\") pod \"redhat-marketplace-5qtbv\" (UID: \"f994800f-7f5f-41ca-87d3-df2121b8890f\") " pod="openshift-marketplace/redhat-marketplace-5qtbv" Dec 11 02:46:05 crc kubenswrapper[4824]: I1211 02:46:05.927505 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xh4f9\" (UniqueName: \"kubernetes.io/projected/f994800f-7f5f-41ca-87d3-df2121b8890f-kube-api-access-xh4f9\") pod \"redhat-marketplace-5qtbv\" (UID: \"f994800f-7f5f-41ca-87d3-df2121b8890f\") " pod="openshift-marketplace/redhat-marketplace-5qtbv" Dec 11 02:46:06 crc kubenswrapper[4824]: I1211 02:46:06.028923 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xh4f9\" (UniqueName: \"kubernetes.io/projected/f994800f-7f5f-41ca-87d3-df2121b8890f-kube-api-access-xh4f9\") pod \"redhat-marketplace-5qtbv\" (UID: \"f994800f-7f5f-41ca-87d3-df2121b8890f\") " pod="openshift-marketplace/redhat-marketplace-5qtbv" Dec 11 02:46:06 crc kubenswrapper[4824]: I1211 02:46:06.029040 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f994800f-7f5f-41ca-87d3-df2121b8890f-utilities\") pod \"redhat-marketplace-5qtbv\" (UID: \"f994800f-7f5f-41ca-87d3-df2121b8890f\") " pod="openshift-marketplace/redhat-marketplace-5qtbv" Dec 11 02:46:06 crc kubenswrapper[4824]: I1211 02:46:06.029131 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f994800f-7f5f-41ca-87d3-df2121b8890f-catalog-content\") pod \"redhat-marketplace-5qtbv\" (UID: \"f994800f-7f5f-41ca-87d3-df2121b8890f\") " pod="openshift-marketplace/redhat-marketplace-5qtbv" Dec 11 02:46:06 crc kubenswrapper[4824]: I1211 02:46:06.029641 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f994800f-7f5f-41ca-87d3-df2121b8890f-utilities\") pod \"redhat-marketplace-5qtbv\" (UID: \"f994800f-7f5f-41ca-87d3-df2121b8890f\") " pod="openshift-marketplace/redhat-marketplace-5qtbv" Dec 11 02:46:06 crc kubenswrapper[4824]: I1211 02:46:06.029707 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f994800f-7f5f-41ca-87d3-df2121b8890f-catalog-content\") pod \"redhat-marketplace-5qtbv\" (UID: \"f994800f-7f5f-41ca-87d3-df2121b8890f\") " pod="openshift-marketplace/redhat-marketplace-5qtbv" Dec 11 02:46:06 crc kubenswrapper[4824]: I1211 02:46:06.048316 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xh4f9\" (UniqueName: \"kubernetes.io/projected/f994800f-7f5f-41ca-87d3-df2121b8890f-kube-api-access-xh4f9\") pod \"redhat-marketplace-5qtbv\" (UID: \"f994800f-7f5f-41ca-87d3-df2121b8890f\") " pod="openshift-marketplace/redhat-marketplace-5qtbv" Dec 11 02:46:06 crc kubenswrapper[4824]: I1211 02:46:06.237657 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5qtbv" Dec 11 02:46:06 crc kubenswrapper[4824]: I1211 02:46:06.765589 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5qtbv"] Dec 11 02:46:07 crc kubenswrapper[4824]: I1211 02:46:07.146986 4824 generic.go:334] "Generic (PLEG): container finished" podID="f994800f-7f5f-41ca-87d3-df2121b8890f" containerID="f592108d6a2cf09a8f61bca8b3de82e4323b15c0ce881471cb7f72337c4d811d" exitCode=0 Dec 11 02:46:07 crc kubenswrapper[4824]: I1211 02:46:07.147182 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5qtbv" event={"ID":"f994800f-7f5f-41ca-87d3-df2121b8890f","Type":"ContainerDied","Data":"f592108d6a2cf09a8f61bca8b3de82e4323b15c0ce881471cb7f72337c4d811d"} Dec 11 02:46:07 crc kubenswrapper[4824]: I1211 02:46:07.147327 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5qtbv" event={"ID":"f994800f-7f5f-41ca-87d3-df2121b8890f","Type":"ContainerStarted","Data":"443384e6385110b15f2813796943a72fd94f9f18a6e3b400520d7601348bc03e"} Dec 11 02:46:07 crc kubenswrapper[4824]: I1211 02:46:07.251224 4824 patch_prober.go:28] interesting pod/machine-config-daemon-gx6xt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 02:46:07 crc kubenswrapper[4824]: I1211 02:46:07.251298 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 02:46:07 crc kubenswrapper[4824]: I1211 02:46:07.251355 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" Dec 11 02:46:07 crc kubenswrapper[4824]: I1211 02:46:07.252226 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4d2b52538d7857fd4c0ee38baaab7aec6956bb5f853fab54c604059ae739842d"} pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 11 02:46:07 crc kubenswrapper[4824]: I1211 02:46:07.252299 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" containerName="machine-config-daemon" containerID="cri-o://4d2b52538d7857fd4c0ee38baaab7aec6956bb5f853fab54c604059ae739842d" gracePeriod=600 Dec 11 02:46:08 crc kubenswrapper[4824]: I1211 02:46:08.163855 4824 generic.go:334] "Generic (PLEG): container finished" podID="44b156e4-64a4-4d45-aa5e-9b10a862faed" containerID="4d2b52538d7857fd4c0ee38baaab7aec6956bb5f853fab54c604059ae739842d" exitCode=0 Dec 11 02:46:08 crc kubenswrapper[4824]: I1211 02:46:08.163919 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" event={"ID":"44b156e4-64a4-4d45-aa5e-9b10a862faed","Type":"ContainerDied","Data":"4d2b52538d7857fd4c0ee38baaab7aec6956bb5f853fab54c604059ae739842d"} Dec 11 02:46:08 crc kubenswrapper[4824]: I1211 02:46:08.164529 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" event={"ID":"44b156e4-64a4-4d45-aa5e-9b10a862faed","Type":"ContainerStarted","Data":"c19d806d2c030c4af9eb013dc9d405f67612520a0b85a1ecd0356d85a1e2fd14"} Dec 11 02:46:08 crc kubenswrapper[4824]: I1211 02:46:08.164561 4824 scope.go:117] "RemoveContainer" containerID="592f00198a08d7f30f2bd4761177c27cba23107c7971ce2c5b2441d9ff2d1144" Dec 11 02:46:09 crc kubenswrapper[4824]: I1211 02:46:09.183011 4824 generic.go:334] "Generic (PLEG): container finished" podID="f994800f-7f5f-41ca-87d3-df2121b8890f" containerID="072bd293338e0e485ab0f7ca61a99d1ff304d70b569e1fb21c2e3a68875c0df5" exitCode=0 Dec 11 02:46:09 crc kubenswrapper[4824]: I1211 02:46:09.183093 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5qtbv" event={"ID":"f994800f-7f5f-41ca-87d3-df2121b8890f","Type":"ContainerDied","Data":"072bd293338e0e485ab0f7ca61a99d1ff304d70b569e1fb21c2e3a68875c0df5"} Dec 11 02:46:10 crc kubenswrapper[4824]: I1211 02:46:10.195006 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5qtbv" event={"ID":"f994800f-7f5f-41ca-87d3-df2121b8890f","Type":"ContainerStarted","Data":"de64accabb5d0166f8f9c48c3cb36b8d2a9607f476f0a7a06a8074a7b6d2b04b"} Dec 11 02:46:10 crc kubenswrapper[4824]: I1211 02:46:10.211414 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-5qtbv" podStartSLOduration=2.616381085 podStartE2EDuration="5.211398676s" podCreationTimestamp="2025-12-11 02:46:05 +0000 UTC" firstStartedPulling="2025-12-11 02:46:07.149435378 +0000 UTC m=+2708.838472767" lastFinishedPulling="2025-12-11 02:46:09.744452969 +0000 UTC m=+2711.433490358" observedRunningTime="2025-12-11 02:46:10.208775047 +0000 UTC m=+2711.897812426" watchObservedRunningTime="2025-12-11 02:46:10.211398676 +0000 UTC m=+2711.900436055" Dec 11 02:46:16 crc kubenswrapper[4824]: I1211 02:46:16.239578 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-5qtbv" Dec 11 02:46:16 crc kubenswrapper[4824]: I1211 02:46:16.241311 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-5qtbv" Dec 11 02:46:16 crc kubenswrapper[4824]: I1211 02:46:16.328329 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-5qtbv" Dec 11 02:46:16 crc kubenswrapper[4824]: I1211 02:46:16.391303 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-5qtbv" Dec 11 02:46:16 crc kubenswrapper[4824]: I1211 02:46:16.581971 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5qtbv"] Dec 11 02:46:18 crc kubenswrapper[4824]: I1211 02:46:18.272523 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-5qtbv" podUID="f994800f-7f5f-41ca-87d3-df2121b8890f" containerName="registry-server" containerID="cri-o://de64accabb5d0166f8f9c48c3cb36b8d2a9607f476f0a7a06a8074a7b6d2b04b" gracePeriod=2 Dec 11 02:46:18 crc kubenswrapper[4824]: E1211 02:46:18.368550 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf994800f_7f5f_41ca_87d3_df2121b8890f.slice/crio-conmon-de64accabb5d0166f8f9c48c3cb36b8d2a9607f476f0a7a06a8074a7b6d2b04b.scope\": RecentStats: unable to find data in memory cache]" Dec 11 02:46:18 crc kubenswrapper[4824]: I1211 02:46:18.802085 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5qtbv" Dec 11 02:46:18 crc kubenswrapper[4824]: I1211 02:46:18.941097 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f994800f-7f5f-41ca-87d3-df2121b8890f-utilities\") pod \"f994800f-7f5f-41ca-87d3-df2121b8890f\" (UID: \"f994800f-7f5f-41ca-87d3-df2121b8890f\") " Dec 11 02:46:18 crc kubenswrapper[4824]: I1211 02:46:18.941565 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xh4f9\" (UniqueName: \"kubernetes.io/projected/f994800f-7f5f-41ca-87d3-df2121b8890f-kube-api-access-xh4f9\") pod \"f994800f-7f5f-41ca-87d3-df2121b8890f\" (UID: \"f994800f-7f5f-41ca-87d3-df2121b8890f\") " Dec 11 02:46:18 crc kubenswrapper[4824]: I1211 02:46:18.941618 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f994800f-7f5f-41ca-87d3-df2121b8890f-catalog-content\") pod \"f994800f-7f5f-41ca-87d3-df2121b8890f\" (UID: \"f994800f-7f5f-41ca-87d3-df2121b8890f\") " Dec 11 02:46:18 crc kubenswrapper[4824]: I1211 02:46:18.942271 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f994800f-7f5f-41ca-87d3-df2121b8890f-utilities" (OuterVolumeSpecName: "utilities") pod "f994800f-7f5f-41ca-87d3-df2121b8890f" (UID: "f994800f-7f5f-41ca-87d3-df2121b8890f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:46:18 crc kubenswrapper[4824]: I1211 02:46:18.953396 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f994800f-7f5f-41ca-87d3-df2121b8890f-kube-api-access-xh4f9" (OuterVolumeSpecName: "kube-api-access-xh4f9") pod "f994800f-7f5f-41ca-87d3-df2121b8890f" (UID: "f994800f-7f5f-41ca-87d3-df2121b8890f"). InnerVolumeSpecName "kube-api-access-xh4f9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:46:18 crc kubenswrapper[4824]: I1211 02:46:18.982012 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f994800f-7f5f-41ca-87d3-df2121b8890f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f994800f-7f5f-41ca-87d3-df2121b8890f" (UID: "f994800f-7f5f-41ca-87d3-df2121b8890f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:46:19 crc kubenswrapper[4824]: I1211 02:46:19.046043 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f994800f-7f5f-41ca-87d3-df2121b8890f-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 02:46:19 crc kubenswrapper[4824]: I1211 02:46:19.046080 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xh4f9\" (UniqueName: \"kubernetes.io/projected/f994800f-7f5f-41ca-87d3-df2121b8890f-kube-api-access-xh4f9\") on node \"crc\" DevicePath \"\"" Dec 11 02:46:19 crc kubenswrapper[4824]: I1211 02:46:19.046092 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f994800f-7f5f-41ca-87d3-df2121b8890f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 02:46:19 crc kubenswrapper[4824]: I1211 02:46:19.283962 4824 generic.go:334] "Generic (PLEG): container finished" podID="f994800f-7f5f-41ca-87d3-df2121b8890f" containerID="de64accabb5d0166f8f9c48c3cb36b8d2a9607f476f0a7a06a8074a7b6d2b04b" exitCode=0 Dec 11 02:46:19 crc kubenswrapper[4824]: I1211 02:46:19.284016 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5qtbv" event={"ID":"f994800f-7f5f-41ca-87d3-df2121b8890f","Type":"ContainerDied","Data":"de64accabb5d0166f8f9c48c3cb36b8d2a9607f476f0a7a06a8074a7b6d2b04b"} Dec 11 02:46:19 crc kubenswrapper[4824]: I1211 02:46:19.284045 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5qtbv" Dec 11 02:46:19 crc kubenswrapper[4824]: I1211 02:46:19.284070 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5qtbv" event={"ID":"f994800f-7f5f-41ca-87d3-df2121b8890f","Type":"ContainerDied","Data":"443384e6385110b15f2813796943a72fd94f9f18a6e3b400520d7601348bc03e"} Dec 11 02:46:19 crc kubenswrapper[4824]: I1211 02:46:19.284094 4824 scope.go:117] "RemoveContainer" containerID="de64accabb5d0166f8f9c48c3cb36b8d2a9607f476f0a7a06a8074a7b6d2b04b" Dec 11 02:46:19 crc kubenswrapper[4824]: I1211 02:46:19.306333 4824 scope.go:117] "RemoveContainer" containerID="072bd293338e0e485ab0f7ca61a99d1ff304d70b569e1fb21c2e3a68875c0df5" Dec 11 02:46:19 crc kubenswrapper[4824]: I1211 02:46:19.316244 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5qtbv"] Dec 11 02:46:19 crc kubenswrapper[4824]: I1211 02:46:19.323361 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-5qtbv"] Dec 11 02:46:19 crc kubenswrapper[4824]: I1211 02:46:19.345376 4824 scope.go:117] "RemoveContainer" containerID="f592108d6a2cf09a8f61bca8b3de82e4323b15c0ce881471cb7f72337c4d811d" Dec 11 02:46:19 crc kubenswrapper[4824]: I1211 02:46:19.379639 4824 scope.go:117] "RemoveContainer" containerID="de64accabb5d0166f8f9c48c3cb36b8d2a9607f476f0a7a06a8074a7b6d2b04b" Dec 11 02:46:19 crc kubenswrapper[4824]: E1211 02:46:19.380350 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de64accabb5d0166f8f9c48c3cb36b8d2a9607f476f0a7a06a8074a7b6d2b04b\": container with ID starting with de64accabb5d0166f8f9c48c3cb36b8d2a9607f476f0a7a06a8074a7b6d2b04b not found: ID does not exist" containerID="de64accabb5d0166f8f9c48c3cb36b8d2a9607f476f0a7a06a8074a7b6d2b04b" Dec 11 02:46:19 crc kubenswrapper[4824]: I1211 02:46:19.380403 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de64accabb5d0166f8f9c48c3cb36b8d2a9607f476f0a7a06a8074a7b6d2b04b"} err="failed to get container status \"de64accabb5d0166f8f9c48c3cb36b8d2a9607f476f0a7a06a8074a7b6d2b04b\": rpc error: code = NotFound desc = could not find container \"de64accabb5d0166f8f9c48c3cb36b8d2a9607f476f0a7a06a8074a7b6d2b04b\": container with ID starting with de64accabb5d0166f8f9c48c3cb36b8d2a9607f476f0a7a06a8074a7b6d2b04b not found: ID does not exist" Dec 11 02:46:19 crc kubenswrapper[4824]: I1211 02:46:19.380435 4824 scope.go:117] "RemoveContainer" containerID="072bd293338e0e485ab0f7ca61a99d1ff304d70b569e1fb21c2e3a68875c0df5" Dec 11 02:46:19 crc kubenswrapper[4824]: E1211 02:46:19.380742 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"072bd293338e0e485ab0f7ca61a99d1ff304d70b569e1fb21c2e3a68875c0df5\": container with ID starting with 072bd293338e0e485ab0f7ca61a99d1ff304d70b569e1fb21c2e3a68875c0df5 not found: ID does not exist" containerID="072bd293338e0e485ab0f7ca61a99d1ff304d70b569e1fb21c2e3a68875c0df5" Dec 11 02:46:19 crc kubenswrapper[4824]: I1211 02:46:19.380784 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"072bd293338e0e485ab0f7ca61a99d1ff304d70b569e1fb21c2e3a68875c0df5"} err="failed to get container status \"072bd293338e0e485ab0f7ca61a99d1ff304d70b569e1fb21c2e3a68875c0df5\": rpc error: code = NotFound desc = could not find container \"072bd293338e0e485ab0f7ca61a99d1ff304d70b569e1fb21c2e3a68875c0df5\": container with ID starting with 072bd293338e0e485ab0f7ca61a99d1ff304d70b569e1fb21c2e3a68875c0df5 not found: ID does not exist" Dec 11 02:46:19 crc kubenswrapper[4824]: I1211 02:46:19.380810 4824 scope.go:117] "RemoveContainer" containerID="f592108d6a2cf09a8f61bca8b3de82e4323b15c0ce881471cb7f72337c4d811d" Dec 11 02:46:19 crc kubenswrapper[4824]: E1211 02:46:19.381074 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f592108d6a2cf09a8f61bca8b3de82e4323b15c0ce881471cb7f72337c4d811d\": container with ID starting with f592108d6a2cf09a8f61bca8b3de82e4323b15c0ce881471cb7f72337c4d811d not found: ID does not exist" containerID="f592108d6a2cf09a8f61bca8b3de82e4323b15c0ce881471cb7f72337c4d811d" Dec 11 02:46:19 crc kubenswrapper[4824]: I1211 02:46:19.381121 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f592108d6a2cf09a8f61bca8b3de82e4323b15c0ce881471cb7f72337c4d811d"} err="failed to get container status \"f592108d6a2cf09a8f61bca8b3de82e4323b15c0ce881471cb7f72337c4d811d\": rpc error: code = NotFound desc = could not find container \"f592108d6a2cf09a8f61bca8b3de82e4323b15c0ce881471cb7f72337c4d811d\": container with ID starting with f592108d6a2cf09a8f61bca8b3de82e4323b15c0ce881471cb7f72337c4d811d not found: ID does not exist" Dec 11 02:46:20 crc kubenswrapper[4824]: I1211 02:46:20.652485 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f994800f-7f5f-41ca-87d3-df2121b8890f" path="/var/lib/kubelet/pods/f994800f-7f5f-41ca-87d3-df2121b8890f/volumes" Dec 11 02:46:42 crc kubenswrapper[4824]: I1211 02:46:42.551232 4824 generic.go:334] "Generic (PLEG): container finished" podID="f844f569-89fd-410c-b2e1-5f9883d9127a" containerID="d6fc1348714e08242892b205d2c6ee536f9f2a1c7a00a9e0a696dfbc93bb9c78" exitCode=0 Dec 11 02:46:42 crc kubenswrapper[4824]: I1211 02:46:42.551332 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7x98s" event={"ID":"f844f569-89fd-410c-b2e1-5f9883d9127a","Type":"ContainerDied","Data":"d6fc1348714e08242892b205d2c6ee536f9f2a1c7a00a9e0a696dfbc93bb9c78"} Dec 11 02:46:44 crc kubenswrapper[4824]: I1211 02:46:44.020205 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7x98s" Dec 11 02:46:44 crc kubenswrapper[4824]: I1211 02:46:44.182060 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f844f569-89fd-410c-b2e1-5f9883d9127a-ssh-key\") pod \"f844f569-89fd-410c-b2e1-5f9883d9127a\" (UID: \"f844f569-89fd-410c-b2e1-5f9883d9127a\") " Dec 11 02:46:44 crc kubenswrapper[4824]: I1211 02:46:44.182177 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/f844f569-89fd-410c-b2e1-5f9883d9127a-nova-cell1-compute-config-0\") pod \"f844f569-89fd-410c-b2e1-5f9883d9127a\" (UID: \"f844f569-89fd-410c-b2e1-5f9883d9127a\") " Dec 11 02:46:44 crc kubenswrapper[4824]: I1211 02:46:44.182264 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/f844f569-89fd-410c-b2e1-5f9883d9127a-nova-extra-config-0\") pod \"f844f569-89fd-410c-b2e1-5f9883d9127a\" (UID: \"f844f569-89fd-410c-b2e1-5f9883d9127a\") " Dec 11 02:46:44 crc kubenswrapper[4824]: I1211 02:46:44.182318 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/f844f569-89fd-410c-b2e1-5f9883d9127a-nova-migration-ssh-key-0\") pod \"f844f569-89fd-410c-b2e1-5f9883d9127a\" (UID: \"f844f569-89fd-410c-b2e1-5f9883d9127a\") " Dec 11 02:46:44 crc kubenswrapper[4824]: I1211 02:46:44.182358 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hrkzr\" (UniqueName: \"kubernetes.io/projected/f844f569-89fd-410c-b2e1-5f9883d9127a-kube-api-access-hrkzr\") pod \"f844f569-89fd-410c-b2e1-5f9883d9127a\" (UID: \"f844f569-89fd-410c-b2e1-5f9883d9127a\") " Dec 11 02:46:44 crc kubenswrapper[4824]: I1211 02:46:44.182387 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/f844f569-89fd-410c-b2e1-5f9883d9127a-nova-migration-ssh-key-1\") pod \"f844f569-89fd-410c-b2e1-5f9883d9127a\" (UID: \"f844f569-89fd-410c-b2e1-5f9883d9127a\") " Dec 11 02:46:44 crc kubenswrapper[4824]: I1211 02:46:44.182421 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/f844f569-89fd-410c-b2e1-5f9883d9127a-nova-cell1-compute-config-1\") pod \"f844f569-89fd-410c-b2e1-5f9883d9127a\" (UID: \"f844f569-89fd-410c-b2e1-5f9883d9127a\") " Dec 11 02:46:44 crc kubenswrapper[4824]: I1211 02:46:44.182475 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f844f569-89fd-410c-b2e1-5f9883d9127a-nova-combined-ca-bundle\") pod \"f844f569-89fd-410c-b2e1-5f9883d9127a\" (UID: \"f844f569-89fd-410c-b2e1-5f9883d9127a\") " Dec 11 02:46:44 crc kubenswrapper[4824]: I1211 02:46:44.182497 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f844f569-89fd-410c-b2e1-5f9883d9127a-inventory\") pod \"f844f569-89fd-410c-b2e1-5f9883d9127a\" (UID: \"f844f569-89fd-410c-b2e1-5f9883d9127a\") " Dec 11 02:46:44 crc kubenswrapper[4824]: I1211 02:46:44.188269 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f844f569-89fd-410c-b2e1-5f9883d9127a-kube-api-access-hrkzr" (OuterVolumeSpecName: "kube-api-access-hrkzr") pod "f844f569-89fd-410c-b2e1-5f9883d9127a" (UID: "f844f569-89fd-410c-b2e1-5f9883d9127a"). InnerVolumeSpecName "kube-api-access-hrkzr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:46:44 crc kubenswrapper[4824]: I1211 02:46:44.188468 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f844f569-89fd-410c-b2e1-5f9883d9127a-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "f844f569-89fd-410c-b2e1-5f9883d9127a" (UID: "f844f569-89fd-410c-b2e1-5f9883d9127a"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:46:44 crc kubenswrapper[4824]: I1211 02:46:44.210761 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f844f569-89fd-410c-b2e1-5f9883d9127a-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "f844f569-89fd-410c-b2e1-5f9883d9127a" (UID: "f844f569-89fd-410c-b2e1-5f9883d9127a"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:46:44 crc kubenswrapper[4824]: I1211 02:46:44.211309 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f844f569-89fd-410c-b2e1-5f9883d9127a-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "f844f569-89fd-410c-b2e1-5f9883d9127a" (UID: "f844f569-89fd-410c-b2e1-5f9883d9127a"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:46:44 crc kubenswrapper[4824]: I1211 02:46:44.213093 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f844f569-89fd-410c-b2e1-5f9883d9127a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f844f569-89fd-410c-b2e1-5f9883d9127a" (UID: "f844f569-89fd-410c-b2e1-5f9883d9127a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:46:44 crc kubenswrapper[4824]: I1211 02:46:44.219261 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f844f569-89fd-410c-b2e1-5f9883d9127a-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "f844f569-89fd-410c-b2e1-5f9883d9127a" (UID: "f844f569-89fd-410c-b2e1-5f9883d9127a"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:46:44 crc kubenswrapper[4824]: I1211 02:46:44.222734 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f844f569-89fd-410c-b2e1-5f9883d9127a-inventory" (OuterVolumeSpecName: "inventory") pod "f844f569-89fd-410c-b2e1-5f9883d9127a" (UID: "f844f569-89fd-410c-b2e1-5f9883d9127a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:46:44 crc kubenswrapper[4824]: I1211 02:46:44.223967 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f844f569-89fd-410c-b2e1-5f9883d9127a-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "f844f569-89fd-410c-b2e1-5f9883d9127a" (UID: "f844f569-89fd-410c-b2e1-5f9883d9127a"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 02:46:44 crc kubenswrapper[4824]: I1211 02:46:44.224318 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f844f569-89fd-410c-b2e1-5f9883d9127a-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "f844f569-89fd-410c-b2e1-5f9883d9127a" (UID: "f844f569-89fd-410c-b2e1-5f9883d9127a"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:46:44 crc kubenswrapper[4824]: I1211 02:46:44.285712 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hrkzr\" (UniqueName: \"kubernetes.io/projected/f844f569-89fd-410c-b2e1-5f9883d9127a-kube-api-access-hrkzr\") on node \"crc\" DevicePath \"\"" Dec 11 02:46:44 crc kubenswrapper[4824]: I1211 02:46:44.285749 4824 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/f844f569-89fd-410c-b2e1-5f9883d9127a-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Dec 11 02:46:44 crc kubenswrapper[4824]: I1211 02:46:44.285762 4824 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/f844f569-89fd-410c-b2e1-5f9883d9127a-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Dec 11 02:46:44 crc kubenswrapper[4824]: I1211 02:46:44.285774 4824 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f844f569-89fd-410c-b2e1-5f9883d9127a-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 02:46:44 crc kubenswrapper[4824]: I1211 02:46:44.285784 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f844f569-89fd-410c-b2e1-5f9883d9127a-inventory\") on node \"crc\" DevicePath \"\"" Dec 11 02:46:44 crc kubenswrapper[4824]: I1211 02:46:44.285792 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f844f569-89fd-410c-b2e1-5f9883d9127a-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 11 02:46:44 crc kubenswrapper[4824]: I1211 02:46:44.285803 4824 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/f844f569-89fd-410c-b2e1-5f9883d9127a-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Dec 11 02:46:44 crc kubenswrapper[4824]: I1211 02:46:44.285813 4824 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/f844f569-89fd-410c-b2e1-5f9883d9127a-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Dec 11 02:46:44 crc kubenswrapper[4824]: I1211 02:46:44.285821 4824 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/f844f569-89fd-410c-b2e1-5f9883d9127a-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Dec 11 02:46:44 crc kubenswrapper[4824]: I1211 02:46:44.575985 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7x98s" event={"ID":"f844f569-89fd-410c-b2e1-5f9883d9127a","Type":"ContainerDied","Data":"f4c3c32f1d4f2e6a2e6d615b28fa257f1f82536af9c0bd7c46b3c9bd965c7296"} Dec 11 02:46:44 crc kubenswrapper[4824]: I1211 02:46:44.576064 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7x98s" Dec 11 02:46:44 crc kubenswrapper[4824]: I1211 02:46:44.576058 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f4c3c32f1d4f2e6a2e6d615b28fa257f1f82536af9c0bd7c46b3c9bd965c7296" Dec 11 02:46:44 crc kubenswrapper[4824]: I1211 02:46:44.697212 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-czhq8"] Dec 11 02:46:44 crc kubenswrapper[4824]: E1211 02:46:44.697812 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f994800f-7f5f-41ca-87d3-df2121b8890f" containerName="registry-server" Dec 11 02:46:44 crc kubenswrapper[4824]: I1211 02:46:44.697833 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f994800f-7f5f-41ca-87d3-df2121b8890f" containerName="registry-server" Dec 11 02:46:44 crc kubenswrapper[4824]: E1211 02:46:44.697899 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f994800f-7f5f-41ca-87d3-df2121b8890f" containerName="extract-utilities" Dec 11 02:46:44 crc kubenswrapper[4824]: I1211 02:46:44.697910 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f994800f-7f5f-41ca-87d3-df2121b8890f" containerName="extract-utilities" Dec 11 02:46:44 crc kubenswrapper[4824]: E1211 02:46:44.697930 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f844f569-89fd-410c-b2e1-5f9883d9127a" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 11 02:46:44 crc kubenswrapper[4824]: I1211 02:46:44.697964 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f844f569-89fd-410c-b2e1-5f9883d9127a" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 11 02:46:44 crc kubenswrapper[4824]: E1211 02:46:44.697983 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f994800f-7f5f-41ca-87d3-df2121b8890f" containerName="extract-content" Dec 11 02:46:44 crc kubenswrapper[4824]: I1211 02:46:44.697991 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f994800f-7f5f-41ca-87d3-df2121b8890f" containerName="extract-content" Dec 11 02:46:44 crc kubenswrapper[4824]: I1211 02:46:44.698233 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f994800f-7f5f-41ca-87d3-df2121b8890f" containerName="registry-server" Dec 11 02:46:44 crc kubenswrapper[4824]: I1211 02:46:44.698256 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f844f569-89fd-410c-b2e1-5f9883d9127a" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 11 02:46:44 crc kubenswrapper[4824]: I1211 02:46:44.698971 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-czhq8" Dec 11 02:46:44 crc kubenswrapper[4824]: I1211 02:46:44.702026 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 11 02:46:44 crc kubenswrapper[4824]: I1211 02:46:44.702204 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 11 02:46:44 crc kubenswrapper[4824]: I1211 02:46:44.734570 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Dec 11 02:46:44 crc kubenswrapper[4824]: I1211 02:46:44.734749 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-hs2fv" Dec 11 02:46:44 crc kubenswrapper[4824]: I1211 02:46:44.734834 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 11 02:46:44 crc kubenswrapper[4824]: I1211 02:46:44.747475 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-czhq8"] Dec 11 02:46:44 crc kubenswrapper[4824]: I1211 02:46:44.802088 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7ed3232-8b86-4804-85e1-0b87715ed1f2-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-czhq8\" (UID: \"f7ed3232-8b86-4804-85e1-0b87715ed1f2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-czhq8" Dec 11 02:46:44 crc kubenswrapper[4824]: I1211 02:46:44.802162 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/f7ed3232-8b86-4804-85e1-0b87715ed1f2-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-czhq8\" (UID: \"f7ed3232-8b86-4804-85e1-0b87715ed1f2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-czhq8" Dec 11 02:46:44 crc kubenswrapper[4824]: I1211 02:46:44.802189 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/f7ed3232-8b86-4804-85e1-0b87715ed1f2-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-czhq8\" (UID: \"f7ed3232-8b86-4804-85e1-0b87715ed1f2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-czhq8" Dec 11 02:46:44 crc kubenswrapper[4824]: I1211 02:46:44.802210 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/f7ed3232-8b86-4804-85e1-0b87715ed1f2-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-czhq8\" (UID: \"f7ed3232-8b86-4804-85e1-0b87715ed1f2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-czhq8" Dec 11 02:46:44 crc kubenswrapper[4824]: I1211 02:46:44.802320 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tk2sm\" (UniqueName: \"kubernetes.io/projected/f7ed3232-8b86-4804-85e1-0b87715ed1f2-kube-api-access-tk2sm\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-czhq8\" (UID: \"f7ed3232-8b86-4804-85e1-0b87715ed1f2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-czhq8" Dec 11 02:46:44 crc kubenswrapper[4824]: I1211 02:46:44.802368 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f7ed3232-8b86-4804-85e1-0b87715ed1f2-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-czhq8\" (UID: \"f7ed3232-8b86-4804-85e1-0b87715ed1f2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-czhq8" Dec 11 02:46:44 crc kubenswrapper[4824]: I1211 02:46:44.802385 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f7ed3232-8b86-4804-85e1-0b87715ed1f2-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-czhq8\" (UID: \"f7ed3232-8b86-4804-85e1-0b87715ed1f2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-czhq8" Dec 11 02:46:44 crc kubenswrapper[4824]: I1211 02:46:44.904256 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tk2sm\" (UniqueName: \"kubernetes.io/projected/f7ed3232-8b86-4804-85e1-0b87715ed1f2-kube-api-access-tk2sm\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-czhq8\" (UID: \"f7ed3232-8b86-4804-85e1-0b87715ed1f2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-czhq8" Dec 11 02:46:44 crc kubenswrapper[4824]: I1211 02:46:44.904353 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f7ed3232-8b86-4804-85e1-0b87715ed1f2-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-czhq8\" (UID: \"f7ed3232-8b86-4804-85e1-0b87715ed1f2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-czhq8" Dec 11 02:46:44 crc kubenswrapper[4824]: I1211 02:46:44.904381 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f7ed3232-8b86-4804-85e1-0b87715ed1f2-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-czhq8\" (UID: \"f7ed3232-8b86-4804-85e1-0b87715ed1f2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-czhq8" Dec 11 02:46:44 crc kubenswrapper[4824]: I1211 02:46:44.904491 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7ed3232-8b86-4804-85e1-0b87715ed1f2-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-czhq8\" (UID: \"f7ed3232-8b86-4804-85e1-0b87715ed1f2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-czhq8" Dec 11 02:46:44 crc kubenswrapper[4824]: I1211 02:46:44.904520 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/f7ed3232-8b86-4804-85e1-0b87715ed1f2-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-czhq8\" (UID: \"f7ed3232-8b86-4804-85e1-0b87715ed1f2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-czhq8" Dec 11 02:46:44 crc kubenswrapper[4824]: I1211 02:46:44.904552 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/f7ed3232-8b86-4804-85e1-0b87715ed1f2-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-czhq8\" (UID: \"f7ed3232-8b86-4804-85e1-0b87715ed1f2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-czhq8" Dec 11 02:46:44 crc kubenswrapper[4824]: I1211 02:46:44.904577 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/f7ed3232-8b86-4804-85e1-0b87715ed1f2-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-czhq8\" (UID: \"f7ed3232-8b86-4804-85e1-0b87715ed1f2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-czhq8" Dec 11 02:46:44 crc kubenswrapper[4824]: I1211 02:46:44.910908 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f7ed3232-8b86-4804-85e1-0b87715ed1f2-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-czhq8\" (UID: \"f7ed3232-8b86-4804-85e1-0b87715ed1f2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-czhq8" Dec 11 02:46:44 crc kubenswrapper[4824]: I1211 02:46:44.910942 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f7ed3232-8b86-4804-85e1-0b87715ed1f2-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-czhq8\" (UID: \"f7ed3232-8b86-4804-85e1-0b87715ed1f2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-czhq8" Dec 11 02:46:44 crc kubenswrapper[4824]: I1211 02:46:44.911019 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/f7ed3232-8b86-4804-85e1-0b87715ed1f2-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-czhq8\" (UID: \"f7ed3232-8b86-4804-85e1-0b87715ed1f2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-czhq8" Dec 11 02:46:44 crc kubenswrapper[4824]: I1211 02:46:44.911464 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/f7ed3232-8b86-4804-85e1-0b87715ed1f2-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-czhq8\" (UID: \"f7ed3232-8b86-4804-85e1-0b87715ed1f2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-czhq8" Dec 11 02:46:44 crc kubenswrapper[4824]: I1211 02:46:44.911534 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7ed3232-8b86-4804-85e1-0b87715ed1f2-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-czhq8\" (UID: \"f7ed3232-8b86-4804-85e1-0b87715ed1f2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-czhq8" Dec 11 02:46:44 crc kubenswrapper[4824]: I1211 02:46:44.911606 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/f7ed3232-8b86-4804-85e1-0b87715ed1f2-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-czhq8\" (UID: \"f7ed3232-8b86-4804-85e1-0b87715ed1f2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-czhq8" Dec 11 02:46:44 crc kubenswrapper[4824]: I1211 02:46:44.922245 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tk2sm\" (UniqueName: \"kubernetes.io/projected/f7ed3232-8b86-4804-85e1-0b87715ed1f2-kube-api-access-tk2sm\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-czhq8\" (UID: \"f7ed3232-8b86-4804-85e1-0b87715ed1f2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-czhq8" Dec 11 02:46:45 crc kubenswrapper[4824]: I1211 02:46:45.062559 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-czhq8" Dec 11 02:46:45 crc kubenswrapper[4824]: I1211 02:46:45.666530 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-czhq8"] Dec 11 02:46:46 crc kubenswrapper[4824]: I1211 02:46:46.609197 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-czhq8" event={"ID":"f7ed3232-8b86-4804-85e1-0b87715ed1f2","Type":"ContainerStarted","Data":"8a568272e2eeedcc9dd89c58d3b0f6d11a516a7929ba3d51210420b385a2d505"} Dec 11 02:46:46 crc kubenswrapper[4824]: I1211 02:46:46.609535 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-czhq8" event={"ID":"f7ed3232-8b86-4804-85e1-0b87715ed1f2","Type":"ContainerStarted","Data":"5921fd3f56bd1a16e691e78dc32d7f295ea0e09bc8974b438e1edcaadf7eb5b9"} Dec 11 02:46:46 crc kubenswrapper[4824]: I1211 02:46:46.637576 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-czhq8" podStartSLOduration=2.419573995 podStartE2EDuration="2.637557804s" podCreationTimestamp="2025-12-11 02:46:44 +0000 UTC" firstStartedPulling="2025-12-11 02:46:45.652751453 +0000 UTC m=+2747.341788842" lastFinishedPulling="2025-12-11 02:46:45.870735232 +0000 UTC m=+2747.559772651" observedRunningTime="2025-12-11 02:46:46.630061078 +0000 UTC m=+2748.319098467" watchObservedRunningTime="2025-12-11 02:46:46.637557804 +0000 UTC m=+2748.326595193" Dec 11 02:47:13 crc kubenswrapper[4824]: I1211 02:47:13.101293 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-qdcv9"] Dec 11 02:47:13 crc kubenswrapper[4824]: I1211 02:47:13.105597 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qdcv9" Dec 11 02:47:13 crc kubenswrapper[4824]: I1211 02:47:13.128789 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qdcv9"] Dec 11 02:47:13 crc kubenswrapper[4824]: I1211 02:47:13.213742 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/42adfb11-2233-42fb-bb0c-a7617d9355ed-catalog-content\") pod \"redhat-operators-qdcv9\" (UID: \"42adfb11-2233-42fb-bb0c-a7617d9355ed\") " pod="openshift-marketplace/redhat-operators-qdcv9" Dec 11 02:47:13 crc kubenswrapper[4824]: I1211 02:47:13.214189 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78nj9\" (UniqueName: \"kubernetes.io/projected/42adfb11-2233-42fb-bb0c-a7617d9355ed-kube-api-access-78nj9\") pod \"redhat-operators-qdcv9\" (UID: \"42adfb11-2233-42fb-bb0c-a7617d9355ed\") " pod="openshift-marketplace/redhat-operators-qdcv9" Dec 11 02:47:13 crc kubenswrapper[4824]: I1211 02:47:13.214379 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/42adfb11-2233-42fb-bb0c-a7617d9355ed-utilities\") pod \"redhat-operators-qdcv9\" (UID: \"42adfb11-2233-42fb-bb0c-a7617d9355ed\") " pod="openshift-marketplace/redhat-operators-qdcv9" Dec 11 02:47:13 crc kubenswrapper[4824]: I1211 02:47:13.316179 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/42adfb11-2233-42fb-bb0c-a7617d9355ed-utilities\") pod \"redhat-operators-qdcv9\" (UID: \"42adfb11-2233-42fb-bb0c-a7617d9355ed\") " pod="openshift-marketplace/redhat-operators-qdcv9" Dec 11 02:47:13 crc kubenswrapper[4824]: I1211 02:47:13.316707 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/42adfb11-2233-42fb-bb0c-a7617d9355ed-utilities\") pod \"redhat-operators-qdcv9\" (UID: \"42adfb11-2233-42fb-bb0c-a7617d9355ed\") " pod="openshift-marketplace/redhat-operators-qdcv9" Dec 11 02:47:13 crc kubenswrapper[4824]: I1211 02:47:13.317058 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/42adfb11-2233-42fb-bb0c-a7617d9355ed-catalog-content\") pod \"redhat-operators-qdcv9\" (UID: \"42adfb11-2233-42fb-bb0c-a7617d9355ed\") " pod="openshift-marketplace/redhat-operators-qdcv9" Dec 11 02:47:13 crc kubenswrapper[4824]: I1211 02:47:13.317344 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78nj9\" (UniqueName: \"kubernetes.io/projected/42adfb11-2233-42fb-bb0c-a7617d9355ed-kube-api-access-78nj9\") pod \"redhat-operators-qdcv9\" (UID: \"42adfb11-2233-42fb-bb0c-a7617d9355ed\") " pod="openshift-marketplace/redhat-operators-qdcv9" Dec 11 02:47:13 crc kubenswrapper[4824]: I1211 02:47:13.317608 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/42adfb11-2233-42fb-bb0c-a7617d9355ed-catalog-content\") pod \"redhat-operators-qdcv9\" (UID: \"42adfb11-2233-42fb-bb0c-a7617d9355ed\") " pod="openshift-marketplace/redhat-operators-qdcv9" Dec 11 02:47:13 crc kubenswrapper[4824]: I1211 02:47:13.338045 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78nj9\" (UniqueName: \"kubernetes.io/projected/42adfb11-2233-42fb-bb0c-a7617d9355ed-kube-api-access-78nj9\") pod \"redhat-operators-qdcv9\" (UID: \"42adfb11-2233-42fb-bb0c-a7617d9355ed\") " pod="openshift-marketplace/redhat-operators-qdcv9" Dec 11 02:47:13 crc kubenswrapper[4824]: I1211 02:47:13.445790 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qdcv9" Dec 11 02:47:13 crc kubenswrapper[4824]: I1211 02:47:13.977862 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qdcv9"] Dec 11 02:47:13 crc kubenswrapper[4824]: W1211 02:47:13.979669 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod42adfb11_2233_42fb_bb0c_a7617d9355ed.slice/crio-c5b709737920993a683bcd21c33f16091e03de01289ad789a37ffa8f95a01784 WatchSource:0}: Error finding container c5b709737920993a683bcd21c33f16091e03de01289ad789a37ffa8f95a01784: Status 404 returned error can't find the container with id c5b709737920993a683bcd21c33f16091e03de01289ad789a37ffa8f95a01784 Dec 11 02:47:14 crc kubenswrapper[4824]: I1211 02:47:14.945198 4824 generic.go:334] "Generic (PLEG): container finished" podID="42adfb11-2233-42fb-bb0c-a7617d9355ed" containerID="2fbc0b2a63630ec699171f1e18e789986ead9fd4f16121b524b0c54ad1a1957e" exitCode=0 Dec 11 02:47:14 crc kubenswrapper[4824]: I1211 02:47:14.945788 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qdcv9" event={"ID":"42adfb11-2233-42fb-bb0c-a7617d9355ed","Type":"ContainerDied","Data":"2fbc0b2a63630ec699171f1e18e789986ead9fd4f16121b524b0c54ad1a1957e"} Dec 11 02:47:14 crc kubenswrapper[4824]: I1211 02:47:14.945819 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qdcv9" event={"ID":"42adfb11-2233-42fb-bb0c-a7617d9355ed","Type":"ContainerStarted","Data":"c5b709737920993a683bcd21c33f16091e03de01289ad789a37ffa8f95a01784"} Dec 11 02:47:15 crc kubenswrapper[4824]: I1211 02:47:15.965697 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qdcv9" event={"ID":"42adfb11-2233-42fb-bb0c-a7617d9355ed","Type":"ContainerStarted","Data":"f81b47f090f0a619047b6f686f6c5741521bc6a70c48a06089225e0fdaa993dc"} Dec 11 02:47:16 crc kubenswrapper[4824]: I1211 02:47:16.980864 4824 generic.go:334] "Generic (PLEG): container finished" podID="42adfb11-2233-42fb-bb0c-a7617d9355ed" containerID="f81b47f090f0a619047b6f686f6c5741521bc6a70c48a06089225e0fdaa993dc" exitCode=0 Dec 11 02:47:16 crc kubenswrapper[4824]: I1211 02:47:16.981316 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qdcv9" event={"ID":"42adfb11-2233-42fb-bb0c-a7617d9355ed","Type":"ContainerDied","Data":"f81b47f090f0a619047b6f686f6c5741521bc6a70c48a06089225e0fdaa993dc"} Dec 11 02:47:19 crc kubenswrapper[4824]: I1211 02:47:19.010870 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qdcv9" event={"ID":"42adfb11-2233-42fb-bb0c-a7617d9355ed","Type":"ContainerStarted","Data":"d4d3e5e8caf348104ae85f637215ac59dafa7f31ddb63f2365e47ed04b3e8cc2"} Dec 11 02:47:19 crc kubenswrapper[4824]: I1211 02:47:19.032197 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-qdcv9" podStartSLOduration=3.203950128 podStartE2EDuration="6.032164388s" podCreationTimestamp="2025-12-11 02:47:13 +0000 UTC" firstStartedPulling="2025-12-11 02:47:14.949506329 +0000 UTC m=+2776.638543728" lastFinishedPulling="2025-12-11 02:47:17.777720559 +0000 UTC m=+2779.466757988" observedRunningTime="2025-12-11 02:47:19.028474488 +0000 UTC m=+2780.717511967" watchObservedRunningTime="2025-12-11 02:47:19.032164388 +0000 UTC m=+2780.721201817" Dec 11 02:47:23 crc kubenswrapper[4824]: I1211 02:47:23.446320 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-qdcv9" Dec 11 02:47:23 crc kubenswrapper[4824]: I1211 02:47:23.449224 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-qdcv9" Dec 11 02:47:24 crc kubenswrapper[4824]: I1211 02:47:24.528109 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-qdcv9" podUID="42adfb11-2233-42fb-bb0c-a7617d9355ed" containerName="registry-server" probeResult="failure" output=< Dec 11 02:47:24 crc kubenswrapper[4824]: timeout: failed to connect service ":50051" within 1s Dec 11 02:47:24 crc kubenswrapper[4824]: > Dec 11 02:47:33 crc kubenswrapper[4824]: I1211 02:47:33.491933 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-qdcv9" Dec 11 02:47:33 crc kubenswrapper[4824]: I1211 02:47:33.545236 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-qdcv9" Dec 11 02:47:33 crc kubenswrapper[4824]: I1211 02:47:33.734839 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qdcv9"] Dec 11 02:47:35 crc kubenswrapper[4824]: I1211 02:47:35.168627 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-qdcv9" podUID="42adfb11-2233-42fb-bb0c-a7617d9355ed" containerName="registry-server" containerID="cri-o://d4d3e5e8caf348104ae85f637215ac59dafa7f31ddb63f2365e47ed04b3e8cc2" gracePeriod=2 Dec 11 02:47:35 crc kubenswrapper[4824]: I1211 02:47:35.757874 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qdcv9" Dec 11 02:47:35 crc kubenswrapper[4824]: I1211 02:47:35.905635 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-78nj9\" (UniqueName: \"kubernetes.io/projected/42adfb11-2233-42fb-bb0c-a7617d9355ed-kube-api-access-78nj9\") pod \"42adfb11-2233-42fb-bb0c-a7617d9355ed\" (UID: \"42adfb11-2233-42fb-bb0c-a7617d9355ed\") " Dec 11 02:47:35 crc kubenswrapper[4824]: I1211 02:47:35.906167 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/42adfb11-2233-42fb-bb0c-a7617d9355ed-catalog-content\") pod \"42adfb11-2233-42fb-bb0c-a7617d9355ed\" (UID: \"42adfb11-2233-42fb-bb0c-a7617d9355ed\") " Dec 11 02:47:35 crc kubenswrapper[4824]: I1211 02:47:35.906366 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/42adfb11-2233-42fb-bb0c-a7617d9355ed-utilities\") pod \"42adfb11-2233-42fb-bb0c-a7617d9355ed\" (UID: \"42adfb11-2233-42fb-bb0c-a7617d9355ed\") " Dec 11 02:47:35 crc kubenswrapper[4824]: I1211 02:47:35.907910 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/42adfb11-2233-42fb-bb0c-a7617d9355ed-utilities" (OuterVolumeSpecName: "utilities") pod "42adfb11-2233-42fb-bb0c-a7617d9355ed" (UID: "42adfb11-2233-42fb-bb0c-a7617d9355ed"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:47:35 crc kubenswrapper[4824]: I1211 02:47:35.916545 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42adfb11-2233-42fb-bb0c-a7617d9355ed-kube-api-access-78nj9" (OuterVolumeSpecName: "kube-api-access-78nj9") pod "42adfb11-2233-42fb-bb0c-a7617d9355ed" (UID: "42adfb11-2233-42fb-bb0c-a7617d9355ed"). InnerVolumeSpecName "kube-api-access-78nj9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:47:36 crc kubenswrapper[4824]: I1211 02:47:36.009575 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-78nj9\" (UniqueName: \"kubernetes.io/projected/42adfb11-2233-42fb-bb0c-a7617d9355ed-kube-api-access-78nj9\") on node \"crc\" DevicePath \"\"" Dec 11 02:47:36 crc kubenswrapper[4824]: I1211 02:47:36.009632 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/42adfb11-2233-42fb-bb0c-a7617d9355ed-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 02:47:36 crc kubenswrapper[4824]: I1211 02:47:36.043856 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/42adfb11-2233-42fb-bb0c-a7617d9355ed-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "42adfb11-2233-42fb-bb0c-a7617d9355ed" (UID: "42adfb11-2233-42fb-bb0c-a7617d9355ed"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:47:36 crc kubenswrapper[4824]: I1211 02:47:36.111938 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/42adfb11-2233-42fb-bb0c-a7617d9355ed-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 02:47:36 crc kubenswrapper[4824]: I1211 02:47:36.182343 4824 generic.go:334] "Generic (PLEG): container finished" podID="42adfb11-2233-42fb-bb0c-a7617d9355ed" containerID="d4d3e5e8caf348104ae85f637215ac59dafa7f31ddb63f2365e47ed04b3e8cc2" exitCode=0 Dec 11 02:47:36 crc kubenswrapper[4824]: I1211 02:47:36.182410 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qdcv9" event={"ID":"42adfb11-2233-42fb-bb0c-a7617d9355ed","Type":"ContainerDied","Data":"d4d3e5e8caf348104ae85f637215ac59dafa7f31ddb63f2365e47ed04b3e8cc2"} Dec 11 02:47:36 crc kubenswrapper[4824]: I1211 02:47:36.182459 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qdcv9" Dec 11 02:47:36 crc kubenswrapper[4824]: I1211 02:47:36.182521 4824 scope.go:117] "RemoveContainer" containerID="d4d3e5e8caf348104ae85f637215ac59dafa7f31ddb63f2365e47ed04b3e8cc2" Dec 11 02:47:36 crc kubenswrapper[4824]: I1211 02:47:36.182464 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qdcv9" event={"ID":"42adfb11-2233-42fb-bb0c-a7617d9355ed","Type":"ContainerDied","Data":"c5b709737920993a683bcd21c33f16091e03de01289ad789a37ffa8f95a01784"} Dec 11 02:47:36 crc kubenswrapper[4824]: I1211 02:47:36.221436 4824 scope.go:117] "RemoveContainer" containerID="f81b47f090f0a619047b6f686f6c5741521bc6a70c48a06089225e0fdaa993dc" Dec 11 02:47:36 crc kubenswrapper[4824]: I1211 02:47:36.223725 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qdcv9"] Dec 11 02:47:36 crc kubenswrapper[4824]: I1211 02:47:36.243003 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-qdcv9"] Dec 11 02:47:36 crc kubenswrapper[4824]: I1211 02:47:36.249308 4824 scope.go:117] "RemoveContainer" containerID="2fbc0b2a63630ec699171f1e18e789986ead9fd4f16121b524b0c54ad1a1957e" Dec 11 02:47:36 crc kubenswrapper[4824]: I1211 02:47:36.296481 4824 scope.go:117] "RemoveContainer" containerID="d4d3e5e8caf348104ae85f637215ac59dafa7f31ddb63f2365e47ed04b3e8cc2" Dec 11 02:47:36 crc kubenswrapper[4824]: E1211 02:47:36.297076 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d4d3e5e8caf348104ae85f637215ac59dafa7f31ddb63f2365e47ed04b3e8cc2\": container with ID starting with d4d3e5e8caf348104ae85f637215ac59dafa7f31ddb63f2365e47ed04b3e8cc2 not found: ID does not exist" containerID="d4d3e5e8caf348104ae85f637215ac59dafa7f31ddb63f2365e47ed04b3e8cc2" Dec 11 02:47:36 crc kubenswrapper[4824]: I1211 02:47:36.297172 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4d3e5e8caf348104ae85f637215ac59dafa7f31ddb63f2365e47ed04b3e8cc2"} err="failed to get container status \"d4d3e5e8caf348104ae85f637215ac59dafa7f31ddb63f2365e47ed04b3e8cc2\": rpc error: code = NotFound desc = could not find container \"d4d3e5e8caf348104ae85f637215ac59dafa7f31ddb63f2365e47ed04b3e8cc2\": container with ID starting with d4d3e5e8caf348104ae85f637215ac59dafa7f31ddb63f2365e47ed04b3e8cc2 not found: ID does not exist" Dec 11 02:47:36 crc kubenswrapper[4824]: I1211 02:47:36.297204 4824 scope.go:117] "RemoveContainer" containerID="f81b47f090f0a619047b6f686f6c5741521bc6a70c48a06089225e0fdaa993dc" Dec 11 02:47:36 crc kubenswrapper[4824]: E1211 02:47:36.297646 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f81b47f090f0a619047b6f686f6c5741521bc6a70c48a06089225e0fdaa993dc\": container with ID starting with f81b47f090f0a619047b6f686f6c5741521bc6a70c48a06089225e0fdaa993dc not found: ID does not exist" containerID="f81b47f090f0a619047b6f686f6c5741521bc6a70c48a06089225e0fdaa993dc" Dec 11 02:47:36 crc kubenswrapper[4824]: I1211 02:47:36.297677 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f81b47f090f0a619047b6f686f6c5741521bc6a70c48a06089225e0fdaa993dc"} err="failed to get container status \"f81b47f090f0a619047b6f686f6c5741521bc6a70c48a06089225e0fdaa993dc\": rpc error: code = NotFound desc = could not find container \"f81b47f090f0a619047b6f686f6c5741521bc6a70c48a06089225e0fdaa993dc\": container with ID starting with f81b47f090f0a619047b6f686f6c5741521bc6a70c48a06089225e0fdaa993dc not found: ID does not exist" Dec 11 02:47:36 crc kubenswrapper[4824]: I1211 02:47:36.297699 4824 scope.go:117] "RemoveContainer" containerID="2fbc0b2a63630ec699171f1e18e789986ead9fd4f16121b524b0c54ad1a1957e" Dec 11 02:47:36 crc kubenswrapper[4824]: E1211 02:47:36.298013 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2fbc0b2a63630ec699171f1e18e789986ead9fd4f16121b524b0c54ad1a1957e\": container with ID starting with 2fbc0b2a63630ec699171f1e18e789986ead9fd4f16121b524b0c54ad1a1957e not found: ID does not exist" containerID="2fbc0b2a63630ec699171f1e18e789986ead9fd4f16121b524b0c54ad1a1957e" Dec 11 02:47:36 crc kubenswrapper[4824]: I1211 02:47:36.298033 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2fbc0b2a63630ec699171f1e18e789986ead9fd4f16121b524b0c54ad1a1957e"} err="failed to get container status \"2fbc0b2a63630ec699171f1e18e789986ead9fd4f16121b524b0c54ad1a1957e\": rpc error: code = NotFound desc = could not find container \"2fbc0b2a63630ec699171f1e18e789986ead9fd4f16121b524b0c54ad1a1957e\": container with ID starting with 2fbc0b2a63630ec699171f1e18e789986ead9fd4f16121b524b0c54ad1a1957e not found: ID does not exist" Dec 11 02:47:36 crc kubenswrapper[4824]: I1211 02:47:36.654185 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="42adfb11-2233-42fb-bb0c-a7617d9355ed" path="/var/lib/kubelet/pods/42adfb11-2233-42fb-bb0c-a7617d9355ed/volumes" Dec 11 02:48:07 crc kubenswrapper[4824]: I1211 02:48:07.251357 4824 patch_prober.go:28] interesting pod/machine-config-daemon-gx6xt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 02:48:07 crc kubenswrapper[4824]: I1211 02:48:07.252224 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 02:48:37 crc kubenswrapper[4824]: I1211 02:48:37.251039 4824 patch_prober.go:28] interesting pod/machine-config-daemon-gx6xt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 02:48:37 crc kubenswrapper[4824]: I1211 02:48:37.253006 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 02:49:07 crc kubenswrapper[4824]: I1211 02:49:07.251065 4824 patch_prober.go:28] interesting pod/machine-config-daemon-gx6xt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 02:49:07 crc kubenswrapper[4824]: I1211 02:49:07.251632 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 02:49:07 crc kubenswrapper[4824]: I1211 02:49:07.251685 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" Dec 11 02:49:07 crc kubenswrapper[4824]: I1211 02:49:07.252577 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c19d806d2c030c4af9eb013dc9d405f67612520a0b85a1ecd0356d85a1e2fd14"} pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 11 02:49:07 crc kubenswrapper[4824]: I1211 02:49:07.252751 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" containerName="machine-config-daemon" containerID="cri-o://c19d806d2c030c4af9eb013dc9d405f67612520a0b85a1ecd0356d85a1e2fd14" gracePeriod=600 Dec 11 02:49:07 crc kubenswrapper[4824]: E1211 02:49:07.377747 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 02:49:08 crc kubenswrapper[4824]: I1211 02:49:08.253320 4824 generic.go:334] "Generic (PLEG): container finished" podID="44b156e4-64a4-4d45-aa5e-9b10a862faed" containerID="c19d806d2c030c4af9eb013dc9d405f67612520a0b85a1ecd0356d85a1e2fd14" exitCode=0 Dec 11 02:49:08 crc kubenswrapper[4824]: I1211 02:49:08.253404 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" event={"ID":"44b156e4-64a4-4d45-aa5e-9b10a862faed","Type":"ContainerDied","Data":"c19d806d2c030c4af9eb013dc9d405f67612520a0b85a1ecd0356d85a1e2fd14"} Dec 11 02:49:08 crc kubenswrapper[4824]: I1211 02:49:08.253509 4824 scope.go:117] "RemoveContainer" containerID="4d2b52538d7857fd4c0ee38baaab7aec6956bb5f853fab54c604059ae739842d" Dec 11 02:49:08 crc kubenswrapper[4824]: I1211 02:49:08.255006 4824 scope.go:117] "RemoveContainer" containerID="c19d806d2c030c4af9eb013dc9d405f67612520a0b85a1ecd0356d85a1e2fd14" Dec 11 02:49:08 crc kubenswrapper[4824]: E1211 02:49:08.255996 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 02:49:21 crc kubenswrapper[4824]: I1211 02:49:21.634218 4824 scope.go:117] "RemoveContainer" containerID="c19d806d2c030c4af9eb013dc9d405f67612520a0b85a1ecd0356d85a1e2fd14" Dec 11 02:49:21 crc kubenswrapper[4824]: E1211 02:49:21.635461 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 02:49:34 crc kubenswrapper[4824]: I1211 02:49:34.632940 4824 scope.go:117] "RemoveContainer" containerID="c19d806d2c030c4af9eb013dc9d405f67612520a0b85a1ecd0356d85a1e2fd14" Dec 11 02:49:34 crc kubenswrapper[4824]: E1211 02:49:34.633961 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 02:49:47 crc kubenswrapper[4824]: I1211 02:49:47.769172 4824 generic.go:334] "Generic (PLEG): container finished" podID="f7ed3232-8b86-4804-85e1-0b87715ed1f2" containerID="8a568272e2eeedcc9dd89c58d3b0f6d11a516a7929ba3d51210420b385a2d505" exitCode=0 Dec 11 02:49:47 crc kubenswrapper[4824]: I1211 02:49:47.769287 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-czhq8" event={"ID":"f7ed3232-8b86-4804-85e1-0b87715ed1f2","Type":"ContainerDied","Data":"8a568272e2eeedcc9dd89c58d3b0f6d11a516a7929ba3d51210420b385a2d505"} Dec 11 02:49:48 crc kubenswrapper[4824]: I1211 02:49:48.640720 4824 scope.go:117] "RemoveContainer" containerID="c19d806d2c030c4af9eb013dc9d405f67612520a0b85a1ecd0356d85a1e2fd14" Dec 11 02:49:48 crc kubenswrapper[4824]: E1211 02:49:48.641148 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 02:49:49 crc kubenswrapper[4824]: I1211 02:49:49.260107 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-czhq8" Dec 11 02:49:49 crc kubenswrapper[4824]: I1211 02:49:49.392316 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/f7ed3232-8b86-4804-85e1-0b87715ed1f2-ceilometer-compute-config-data-0\") pod \"f7ed3232-8b86-4804-85e1-0b87715ed1f2\" (UID: \"f7ed3232-8b86-4804-85e1-0b87715ed1f2\") " Dec 11 02:49:49 crc kubenswrapper[4824]: I1211 02:49:49.392464 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f7ed3232-8b86-4804-85e1-0b87715ed1f2-inventory\") pod \"f7ed3232-8b86-4804-85e1-0b87715ed1f2\" (UID: \"f7ed3232-8b86-4804-85e1-0b87715ed1f2\") " Dec 11 02:49:49 crc kubenswrapper[4824]: I1211 02:49:49.392514 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/f7ed3232-8b86-4804-85e1-0b87715ed1f2-ceilometer-compute-config-data-2\") pod \"f7ed3232-8b86-4804-85e1-0b87715ed1f2\" (UID: \"f7ed3232-8b86-4804-85e1-0b87715ed1f2\") " Dec 11 02:49:49 crc kubenswrapper[4824]: I1211 02:49:49.392545 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk2sm\" (UniqueName: \"kubernetes.io/projected/f7ed3232-8b86-4804-85e1-0b87715ed1f2-kube-api-access-tk2sm\") pod \"f7ed3232-8b86-4804-85e1-0b87715ed1f2\" (UID: \"f7ed3232-8b86-4804-85e1-0b87715ed1f2\") " Dec 11 02:49:49 crc kubenswrapper[4824]: I1211 02:49:49.392600 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/f7ed3232-8b86-4804-85e1-0b87715ed1f2-ceilometer-compute-config-data-1\") pod \"f7ed3232-8b86-4804-85e1-0b87715ed1f2\" (UID: \"f7ed3232-8b86-4804-85e1-0b87715ed1f2\") " Dec 11 02:49:49 crc kubenswrapper[4824]: I1211 02:49:49.392645 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7ed3232-8b86-4804-85e1-0b87715ed1f2-telemetry-combined-ca-bundle\") pod \"f7ed3232-8b86-4804-85e1-0b87715ed1f2\" (UID: \"f7ed3232-8b86-4804-85e1-0b87715ed1f2\") " Dec 11 02:49:49 crc kubenswrapper[4824]: I1211 02:49:49.392726 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f7ed3232-8b86-4804-85e1-0b87715ed1f2-ssh-key\") pod \"f7ed3232-8b86-4804-85e1-0b87715ed1f2\" (UID: \"f7ed3232-8b86-4804-85e1-0b87715ed1f2\") " Dec 11 02:49:49 crc kubenswrapper[4824]: I1211 02:49:49.400271 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7ed3232-8b86-4804-85e1-0b87715ed1f2-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "f7ed3232-8b86-4804-85e1-0b87715ed1f2" (UID: "f7ed3232-8b86-4804-85e1-0b87715ed1f2"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:49:49 crc kubenswrapper[4824]: I1211 02:49:49.410910 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7ed3232-8b86-4804-85e1-0b87715ed1f2-kube-api-access-tk2sm" (OuterVolumeSpecName: "kube-api-access-tk2sm") pod "f7ed3232-8b86-4804-85e1-0b87715ed1f2" (UID: "f7ed3232-8b86-4804-85e1-0b87715ed1f2"). InnerVolumeSpecName "kube-api-access-tk2sm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:49:49 crc kubenswrapper[4824]: I1211 02:49:49.421191 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7ed3232-8b86-4804-85e1-0b87715ed1f2-inventory" (OuterVolumeSpecName: "inventory") pod "f7ed3232-8b86-4804-85e1-0b87715ed1f2" (UID: "f7ed3232-8b86-4804-85e1-0b87715ed1f2"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:49:49 crc kubenswrapper[4824]: I1211 02:49:49.429510 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7ed3232-8b86-4804-85e1-0b87715ed1f2-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "f7ed3232-8b86-4804-85e1-0b87715ed1f2" (UID: "f7ed3232-8b86-4804-85e1-0b87715ed1f2"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:49:49 crc kubenswrapper[4824]: I1211 02:49:49.429867 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7ed3232-8b86-4804-85e1-0b87715ed1f2-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f7ed3232-8b86-4804-85e1-0b87715ed1f2" (UID: "f7ed3232-8b86-4804-85e1-0b87715ed1f2"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:49:49 crc kubenswrapper[4824]: I1211 02:49:49.432948 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7ed3232-8b86-4804-85e1-0b87715ed1f2-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "f7ed3232-8b86-4804-85e1-0b87715ed1f2" (UID: "f7ed3232-8b86-4804-85e1-0b87715ed1f2"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:49:49 crc kubenswrapper[4824]: I1211 02:49:49.454141 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7ed3232-8b86-4804-85e1-0b87715ed1f2-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "f7ed3232-8b86-4804-85e1-0b87715ed1f2" (UID: "f7ed3232-8b86-4804-85e1-0b87715ed1f2"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 02:49:49 crc kubenswrapper[4824]: I1211 02:49:49.495351 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f7ed3232-8b86-4804-85e1-0b87715ed1f2-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 11 02:49:49 crc kubenswrapper[4824]: I1211 02:49:49.495387 4824 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/f7ed3232-8b86-4804-85e1-0b87715ed1f2-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Dec 11 02:49:49 crc kubenswrapper[4824]: I1211 02:49:49.495401 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f7ed3232-8b86-4804-85e1-0b87715ed1f2-inventory\") on node \"crc\" DevicePath \"\"" Dec 11 02:49:49 crc kubenswrapper[4824]: I1211 02:49:49.495414 4824 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/f7ed3232-8b86-4804-85e1-0b87715ed1f2-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Dec 11 02:49:49 crc kubenswrapper[4824]: I1211 02:49:49.495427 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk2sm\" (UniqueName: \"kubernetes.io/projected/f7ed3232-8b86-4804-85e1-0b87715ed1f2-kube-api-access-tk2sm\") on node \"crc\" DevicePath \"\"" Dec 11 02:49:49 crc kubenswrapper[4824]: I1211 02:49:49.495440 4824 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/f7ed3232-8b86-4804-85e1-0b87715ed1f2-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Dec 11 02:49:49 crc kubenswrapper[4824]: I1211 02:49:49.495452 4824 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7ed3232-8b86-4804-85e1-0b87715ed1f2-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 02:49:49 crc kubenswrapper[4824]: I1211 02:49:49.791387 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-czhq8" event={"ID":"f7ed3232-8b86-4804-85e1-0b87715ed1f2","Type":"ContainerDied","Data":"5921fd3f56bd1a16e691e78dc32d7f295ea0e09bc8974b438e1edcaadf7eb5b9"} Dec 11 02:49:49 crc kubenswrapper[4824]: I1211 02:49:49.791651 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5921fd3f56bd1a16e691e78dc32d7f295ea0e09bc8974b438e1edcaadf7eb5b9" Dec 11 02:49:49 crc kubenswrapper[4824]: I1211 02:49:49.791446 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-czhq8" Dec 11 02:50:03 crc kubenswrapper[4824]: I1211 02:50:03.632334 4824 scope.go:117] "RemoveContainer" containerID="c19d806d2c030c4af9eb013dc9d405f67612520a0b85a1ecd0356d85a1e2fd14" Dec 11 02:50:03 crc kubenswrapper[4824]: E1211 02:50:03.633274 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 02:50:13 crc kubenswrapper[4824]: E1211 02:50:13.850797 4824 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.50:35500->38.102.83.50:42353: write tcp 38.102.83.50:35500->38.102.83.50:42353: write: broken pipe Dec 11 02:50:14 crc kubenswrapper[4824]: I1211 02:50:14.632935 4824 scope.go:117] "RemoveContainer" containerID="c19d806d2c030c4af9eb013dc9d405f67612520a0b85a1ecd0356d85a1e2fd14" Dec 11 02:50:14 crc kubenswrapper[4824]: E1211 02:50:14.633483 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 02:50:26 crc kubenswrapper[4824]: I1211 02:50:26.633284 4824 scope.go:117] "RemoveContainer" containerID="c19d806d2c030c4af9eb013dc9d405f67612520a0b85a1ecd0356d85a1e2fd14" Dec 11 02:50:26 crc kubenswrapper[4824]: E1211 02:50:26.634059 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 02:50:39 crc kubenswrapper[4824]: I1211 02:50:39.634143 4824 scope.go:117] "RemoveContainer" containerID="c19d806d2c030c4af9eb013dc9d405f67612520a0b85a1ecd0356d85a1e2fd14" Dec 11 02:50:39 crc kubenswrapper[4824]: E1211 02:50:39.635236 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 02:50:46 crc kubenswrapper[4824]: I1211 02:50:46.683096 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Dec 11 02:50:46 crc kubenswrapper[4824]: E1211 02:50:46.684045 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7ed3232-8b86-4804-85e1-0b87715ed1f2" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 11 02:50:46 crc kubenswrapper[4824]: I1211 02:50:46.684071 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7ed3232-8b86-4804-85e1-0b87715ed1f2" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 11 02:50:46 crc kubenswrapper[4824]: E1211 02:50:46.684135 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42adfb11-2233-42fb-bb0c-a7617d9355ed" containerName="extract-utilities" Dec 11 02:50:46 crc kubenswrapper[4824]: I1211 02:50:46.684149 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="42adfb11-2233-42fb-bb0c-a7617d9355ed" containerName="extract-utilities" Dec 11 02:50:46 crc kubenswrapper[4824]: E1211 02:50:46.684187 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42adfb11-2233-42fb-bb0c-a7617d9355ed" containerName="registry-server" Dec 11 02:50:46 crc kubenswrapper[4824]: I1211 02:50:46.684197 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="42adfb11-2233-42fb-bb0c-a7617d9355ed" containerName="registry-server" Dec 11 02:50:46 crc kubenswrapper[4824]: E1211 02:50:46.684222 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42adfb11-2233-42fb-bb0c-a7617d9355ed" containerName="extract-content" Dec 11 02:50:46 crc kubenswrapper[4824]: I1211 02:50:46.684232 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="42adfb11-2233-42fb-bb0c-a7617d9355ed" containerName="extract-content" Dec 11 02:50:46 crc kubenswrapper[4824]: I1211 02:50:46.684533 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="42adfb11-2233-42fb-bb0c-a7617d9355ed" containerName="registry-server" Dec 11 02:50:46 crc kubenswrapper[4824]: I1211 02:50:46.684555 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7ed3232-8b86-4804-85e1-0b87715ed1f2" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 11 02:50:46 crc kubenswrapper[4824]: I1211 02:50:46.685304 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 11 02:50:46 crc kubenswrapper[4824]: I1211 02:50:46.696482 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Dec 11 02:50:46 crc kubenswrapper[4824]: I1211 02:50:46.697279 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Dec 11 02:50:46 crc kubenswrapper[4824]: I1211 02:50:46.697493 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-nl4zv" Dec 11 02:50:46 crc kubenswrapper[4824]: I1211 02:50:46.698901 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Dec 11 02:50:46 crc kubenswrapper[4824]: I1211 02:50:46.705765 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 11 02:50:46 crc kubenswrapper[4824]: I1211 02:50:46.757075 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ab73a048-199c-464e-98ab-2e87063161b6-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"ab73a048-199c-464e-98ab-2e87063161b6\") " pod="openstack/tempest-tests-tempest" Dec 11 02:50:46 crc kubenswrapper[4824]: I1211 02:50:46.757410 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ab73a048-199c-464e-98ab-2e87063161b6-config-data\") pod \"tempest-tests-tempest\" (UID: \"ab73a048-199c-464e-98ab-2e87063161b6\") " pod="openstack/tempest-tests-tempest" Dec 11 02:50:46 crc kubenswrapper[4824]: I1211 02:50:46.757606 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ab73a048-199c-464e-98ab-2e87063161b6-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"ab73a048-199c-464e-98ab-2e87063161b6\") " pod="openstack/tempest-tests-tempest" Dec 11 02:50:46 crc kubenswrapper[4824]: I1211 02:50:46.860313 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ab73a048-199c-464e-98ab-2e87063161b6-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"ab73a048-199c-464e-98ab-2e87063161b6\") " pod="openstack/tempest-tests-tempest" Dec 11 02:50:46 crc kubenswrapper[4824]: I1211 02:50:46.860386 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ab73a048-199c-464e-98ab-2e87063161b6-config-data\") pod \"tempest-tests-tempest\" (UID: \"ab73a048-199c-464e-98ab-2e87063161b6\") " pod="openstack/tempest-tests-tempest" Dec 11 02:50:46 crc kubenswrapper[4824]: I1211 02:50:46.860411 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/ab73a048-199c-464e-98ab-2e87063161b6-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"ab73a048-199c-464e-98ab-2e87063161b6\") " pod="openstack/tempest-tests-tempest" Dec 11 02:50:47 crc kubenswrapper[4824]: I1211 02:50:46.860652 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ab73a048-199c-464e-98ab-2e87063161b6-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"ab73a048-199c-464e-98ab-2e87063161b6\") " pod="openstack/tempest-tests-tempest" Dec 11 02:50:47 crc kubenswrapper[4824]: I1211 02:50:46.860806 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/ab73a048-199c-464e-98ab-2e87063161b6-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"ab73a048-199c-464e-98ab-2e87063161b6\") " pod="openstack/tempest-tests-tempest" Dec 11 02:50:47 crc kubenswrapper[4824]: I1211 02:50:46.860869 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n8pl5\" (UniqueName: \"kubernetes.io/projected/ab73a048-199c-464e-98ab-2e87063161b6-kube-api-access-n8pl5\") pod \"tempest-tests-tempest\" (UID: \"ab73a048-199c-464e-98ab-2e87063161b6\") " pod="openstack/tempest-tests-tempest" Dec 11 02:50:47 crc kubenswrapper[4824]: I1211 02:50:46.860938 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"tempest-tests-tempest\" (UID: \"ab73a048-199c-464e-98ab-2e87063161b6\") " pod="openstack/tempest-tests-tempest" Dec 11 02:50:47 crc kubenswrapper[4824]: I1211 02:50:46.861033 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ab73a048-199c-464e-98ab-2e87063161b6-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"ab73a048-199c-464e-98ab-2e87063161b6\") " pod="openstack/tempest-tests-tempest" Dec 11 02:50:47 crc kubenswrapper[4824]: I1211 02:50:46.861072 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/ab73a048-199c-464e-98ab-2e87063161b6-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"ab73a048-199c-464e-98ab-2e87063161b6\") " pod="openstack/tempest-tests-tempest" Dec 11 02:50:47 crc kubenswrapper[4824]: I1211 02:50:46.861511 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ab73a048-199c-464e-98ab-2e87063161b6-config-data\") pod \"tempest-tests-tempest\" (UID: \"ab73a048-199c-464e-98ab-2e87063161b6\") " pod="openstack/tempest-tests-tempest" Dec 11 02:50:47 crc kubenswrapper[4824]: I1211 02:50:46.862205 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ab73a048-199c-464e-98ab-2e87063161b6-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"ab73a048-199c-464e-98ab-2e87063161b6\") " pod="openstack/tempest-tests-tempest" Dec 11 02:50:47 crc kubenswrapper[4824]: I1211 02:50:46.867672 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ab73a048-199c-464e-98ab-2e87063161b6-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"ab73a048-199c-464e-98ab-2e87063161b6\") " pod="openstack/tempest-tests-tempest" Dec 11 02:50:47 crc kubenswrapper[4824]: I1211 02:50:46.964568 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ab73a048-199c-464e-98ab-2e87063161b6-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"ab73a048-199c-464e-98ab-2e87063161b6\") " pod="openstack/tempest-tests-tempest" Dec 11 02:50:47 crc kubenswrapper[4824]: I1211 02:50:46.964643 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/ab73a048-199c-464e-98ab-2e87063161b6-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"ab73a048-199c-464e-98ab-2e87063161b6\") " pod="openstack/tempest-tests-tempest" Dec 11 02:50:47 crc kubenswrapper[4824]: I1211 02:50:46.964880 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/ab73a048-199c-464e-98ab-2e87063161b6-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"ab73a048-199c-464e-98ab-2e87063161b6\") " pod="openstack/tempest-tests-tempest" Dec 11 02:50:47 crc kubenswrapper[4824]: I1211 02:50:46.964933 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n8pl5\" (UniqueName: \"kubernetes.io/projected/ab73a048-199c-464e-98ab-2e87063161b6-kube-api-access-n8pl5\") pod \"tempest-tests-tempest\" (UID: \"ab73a048-199c-464e-98ab-2e87063161b6\") " pod="openstack/tempest-tests-tempest" Dec 11 02:50:47 crc kubenswrapper[4824]: I1211 02:50:46.964999 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"tempest-tests-tempest\" (UID: \"ab73a048-199c-464e-98ab-2e87063161b6\") " pod="openstack/tempest-tests-tempest" Dec 11 02:50:47 crc kubenswrapper[4824]: I1211 02:50:46.965102 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/ab73a048-199c-464e-98ab-2e87063161b6-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"ab73a048-199c-464e-98ab-2e87063161b6\") " pod="openstack/tempest-tests-tempest" Dec 11 02:50:47 crc kubenswrapper[4824]: I1211 02:50:46.965689 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/ab73a048-199c-464e-98ab-2e87063161b6-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"ab73a048-199c-464e-98ab-2e87063161b6\") " pod="openstack/tempest-tests-tempest" Dec 11 02:50:47 crc kubenswrapper[4824]: I1211 02:50:46.965942 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"tempest-tests-tempest\" (UID: \"ab73a048-199c-464e-98ab-2e87063161b6\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/tempest-tests-tempest" Dec 11 02:50:47 crc kubenswrapper[4824]: I1211 02:50:46.966105 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/ab73a048-199c-464e-98ab-2e87063161b6-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"ab73a048-199c-464e-98ab-2e87063161b6\") " pod="openstack/tempest-tests-tempest" Dec 11 02:50:47 crc kubenswrapper[4824]: I1211 02:50:46.969378 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ab73a048-199c-464e-98ab-2e87063161b6-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"ab73a048-199c-464e-98ab-2e87063161b6\") " pod="openstack/tempest-tests-tempest" Dec 11 02:50:47 crc kubenswrapper[4824]: I1211 02:50:46.976325 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/ab73a048-199c-464e-98ab-2e87063161b6-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"ab73a048-199c-464e-98ab-2e87063161b6\") " pod="openstack/tempest-tests-tempest" Dec 11 02:50:47 crc kubenswrapper[4824]: I1211 02:50:46.985841 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n8pl5\" (UniqueName: \"kubernetes.io/projected/ab73a048-199c-464e-98ab-2e87063161b6-kube-api-access-n8pl5\") pod \"tempest-tests-tempest\" (UID: \"ab73a048-199c-464e-98ab-2e87063161b6\") " pod="openstack/tempest-tests-tempest" Dec 11 02:50:47 crc kubenswrapper[4824]: I1211 02:50:47.023549 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"tempest-tests-tempest\" (UID: \"ab73a048-199c-464e-98ab-2e87063161b6\") " pod="openstack/tempest-tests-tempest" Dec 11 02:50:47 crc kubenswrapper[4824]: I1211 02:50:47.326482 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 11 02:50:47 crc kubenswrapper[4824]: I1211 02:50:47.928741 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 11 02:50:47 crc kubenswrapper[4824]: I1211 02:50:47.943703 4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 11 02:50:48 crc kubenswrapper[4824]: I1211 02:50:48.919171 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"ab73a048-199c-464e-98ab-2e87063161b6","Type":"ContainerStarted","Data":"5f19dd57346c6c5329a1a25379c65c995b28041e84abd07d0f7209e8e0744941"} Dec 11 02:50:51 crc kubenswrapper[4824]: I1211 02:50:51.632550 4824 scope.go:117] "RemoveContainer" containerID="c19d806d2c030c4af9eb013dc9d405f67612520a0b85a1ecd0356d85a1e2fd14" Dec 11 02:50:51 crc kubenswrapper[4824]: E1211 02:50:51.633048 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 02:51:03 crc kubenswrapper[4824]: I1211 02:51:03.632173 4824 scope.go:117] "RemoveContainer" containerID="c19d806d2c030c4af9eb013dc9d405f67612520a0b85a1ecd0356d85a1e2fd14" Dec 11 02:51:03 crc kubenswrapper[4824]: E1211 02:51:03.632976 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 02:51:16 crc kubenswrapper[4824]: I1211 02:51:16.634876 4824 scope.go:117] "RemoveContainer" containerID="c19d806d2c030c4af9eb013dc9d405f67612520a0b85a1ecd0356d85a1e2fd14" Dec 11 02:51:16 crc kubenswrapper[4824]: E1211 02:51:16.635808 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 02:51:24 crc kubenswrapper[4824]: E1211 02:51:24.140618 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Dec 11 02:51:24 crc kubenswrapper[4824]: E1211 02:51:24.141423 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-n8pl5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(ab73a048-199c-464e-98ab-2e87063161b6): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 11 02:51:24 crc kubenswrapper[4824]: E1211 02:51:24.142666 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="ab73a048-199c-464e-98ab-2e87063161b6" Dec 11 02:51:24 crc kubenswrapper[4824]: E1211 02:51:24.303968 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="ab73a048-199c-464e-98ab-2e87063161b6" Dec 11 02:51:30 crc kubenswrapper[4824]: I1211 02:51:30.634330 4824 scope.go:117] "RemoveContainer" containerID="c19d806d2c030c4af9eb013dc9d405f67612520a0b85a1ecd0356d85a1e2fd14" Dec 11 02:51:30 crc kubenswrapper[4824]: E1211 02:51:30.635074 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 02:51:38 crc kubenswrapper[4824]: I1211 02:51:38.158809 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Dec 11 02:51:39 crc kubenswrapper[4824]: I1211 02:51:39.495080 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"ab73a048-199c-464e-98ab-2e87063161b6","Type":"ContainerStarted","Data":"47d6cc75b297e87d539ca6fb5a3d045b63a034458b297206ee08a75db9a2f4d2"} Dec 11 02:51:39 crc kubenswrapper[4824]: I1211 02:51:39.523706 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=4.312463276 podStartE2EDuration="54.523685016s" podCreationTimestamp="2025-12-11 02:50:45 +0000 UTC" firstStartedPulling="2025-12-11 02:50:47.943234476 +0000 UTC m=+2989.632271895" lastFinishedPulling="2025-12-11 02:51:38.154456216 +0000 UTC m=+3039.843493635" observedRunningTime="2025-12-11 02:51:39.513495701 +0000 UTC m=+3041.202533080" watchObservedRunningTime="2025-12-11 02:51:39.523685016 +0000 UTC m=+3041.212722395" Dec 11 02:51:43 crc kubenswrapper[4824]: I1211 02:51:43.633310 4824 scope.go:117] "RemoveContainer" containerID="c19d806d2c030c4af9eb013dc9d405f67612520a0b85a1ecd0356d85a1e2fd14" Dec 11 02:51:43 crc kubenswrapper[4824]: E1211 02:51:43.634390 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 02:51:55 crc kubenswrapper[4824]: I1211 02:51:55.632557 4824 scope.go:117] "RemoveContainer" containerID="c19d806d2c030c4af9eb013dc9d405f67612520a0b85a1ecd0356d85a1e2fd14" Dec 11 02:51:55 crc kubenswrapper[4824]: E1211 02:51:55.633387 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 02:52:09 crc kubenswrapper[4824]: I1211 02:52:09.633180 4824 scope.go:117] "RemoveContainer" containerID="c19d806d2c030c4af9eb013dc9d405f67612520a0b85a1ecd0356d85a1e2fd14" Dec 11 02:52:09 crc kubenswrapper[4824]: E1211 02:52:09.634426 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 02:52:16 crc kubenswrapper[4824]: I1211 02:52:16.581071 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-jlblg"] Dec 11 02:52:16 crc kubenswrapper[4824]: I1211 02:52:16.586756 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jlblg" Dec 11 02:52:16 crc kubenswrapper[4824]: I1211 02:52:16.607186 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jlblg"] Dec 11 02:52:16 crc kubenswrapper[4824]: I1211 02:52:16.678168 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f4d993a-bd6a-4ba8-bbbb-d389c2b33015-catalog-content\") pod \"community-operators-jlblg\" (UID: \"5f4d993a-bd6a-4ba8-bbbb-d389c2b33015\") " pod="openshift-marketplace/community-operators-jlblg" Dec 11 02:52:16 crc kubenswrapper[4824]: I1211 02:52:16.678258 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f4d993a-bd6a-4ba8-bbbb-d389c2b33015-utilities\") pod \"community-operators-jlblg\" (UID: \"5f4d993a-bd6a-4ba8-bbbb-d389c2b33015\") " pod="openshift-marketplace/community-operators-jlblg" Dec 11 02:52:16 crc kubenswrapper[4824]: I1211 02:52:16.678442 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-542dn\" (UniqueName: \"kubernetes.io/projected/5f4d993a-bd6a-4ba8-bbbb-d389c2b33015-kube-api-access-542dn\") pod \"community-operators-jlblg\" (UID: \"5f4d993a-bd6a-4ba8-bbbb-d389c2b33015\") " pod="openshift-marketplace/community-operators-jlblg" Dec 11 02:52:16 crc kubenswrapper[4824]: I1211 02:52:16.768647 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-5m84m"] Dec 11 02:52:16 crc kubenswrapper[4824]: I1211 02:52:16.770541 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5m84m" Dec 11 02:52:16 crc kubenswrapper[4824]: I1211 02:52:16.780088 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5m84m"] Dec 11 02:52:16 crc kubenswrapper[4824]: I1211 02:52:16.781802 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-542dn\" (UniqueName: \"kubernetes.io/projected/5f4d993a-bd6a-4ba8-bbbb-d389c2b33015-kube-api-access-542dn\") pod \"community-operators-jlblg\" (UID: \"5f4d993a-bd6a-4ba8-bbbb-d389c2b33015\") " pod="openshift-marketplace/community-operators-jlblg" Dec 11 02:52:16 crc kubenswrapper[4824]: I1211 02:52:16.782218 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f4d993a-bd6a-4ba8-bbbb-d389c2b33015-catalog-content\") pod \"community-operators-jlblg\" (UID: \"5f4d993a-bd6a-4ba8-bbbb-d389c2b33015\") " pod="openshift-marketplace/community-operators-jlblg" Dec 11 02:52:16 crc kubenswrapper[4824]: I1211 02:52:16.782492 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f4d993a-bd6a-4ba8-bbbb-d389c2b33015-utilities\") pod \"community-operators-jlblg\" (UID: \"5f4d993a-bd6a-4ba8-bbbb-d389c2b33015\") " pod="openshift-marketplace/community-operators-jlblg" Dec 11 02:52:16 crc kubenswrapper[4824]: I1211 02:52:16.783987 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f4d993a-bd6a-4ba8-bbbb-d389c2b33015-utilities\") pod \"community-operators-jlblg\" (UID: \"5f4d993a-bd6a-4ba8-bbbb-d389c2b33015\") " pod="openshift-marketplace/community-operators-jlblg" Dec 11 02:52:16 crc kubenswrapper[4824]: I1211 02:52:16.784473 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f4d993a-bd6a-4ba8-bbbb-d389c2b33015-catalog-content\") pod \"community-operators-jlblg\" (UID: \"5f4d993a-bd6a-4ba8-bbbb-d389c2b33015\") " pod="openshift-marketplace/community-operators-jlblg" Dec 11 02:52:16 crc kubenswrapper[4824]: I1211 02:52:16.829755 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-542dn\" (UniqueName: \"kubernetes.io/projected/5f4d993a-bd6a-4ba8-bbbb-d389c2b33015-kube-api-access-542dn\") pod \"community-operators-jlblg\" (UID: \"5f4d993a-bd6a-4ba8-bbbb-d389c2b33015\") " pod="openshift-marketplace/community-operators-jlblg" Dec 11 02:52:16 crc kubenswrapper[4824]: I1211 02:52:16.884678 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b33086d6-9b3e-4673-bd80-52913963f7b3-catalog-content\") pod \"certified-operators-5m84m\" (UID: \"b33086d6-9b3e-4673-bd80-52913963f7b3\") " pod="openshift-marketplace/certified-operators-5m84m" Dec 11 02:52:16 crc kubenswrapper[4824]: I1211 02:52:16.884751 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b33086d6-9b3e-4673-bd80-52913963f7b3-utilities\") pod \"certified-operators-5m84m\" (UID: \"b33086d6-9b3e-4673-bd80-52913963f7b3\") " pod="openshift-marketplace/certified-operators-5m84m" Dec 11 02:52:16 crc kubenswrapper[4824]: I1211 02:52:16.884795 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vp8tx\" (UniqueName: \"kubernetes.io/projected/b33086d6-9b3e-4673-bd80-52913963f7b3-kube-api-access-vp8tx\") pod \"certified-operators-5m84m\" (UID: \"b33086d6-9b3e-4673-bd80-52913963f7b3\") " pod="openshift-marketplace/certified-operators-5m84m" Dec 11 02:52:16 crc kubenswrapper[4824]: I1211 02:52:16.957799 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jlblg" Dec 11 02:52:16 crc kubenswrapper[4824]: I1211 02:52:16.986355 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b33086d6-9b3e-4673-bd80-52913963f7b3-utilities\") pod \"certified-operators-5m84m\" (UID: \"b33086d6-9b3e-4673-bd80-52913963f7b3\") " pod="openshift-marketplace/certified-operators-5m84m" Dec 11 02:52:16 crc kubenswrapper[4824]: I1211 02:52:16.986558 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vp8tx\" (UniqueName: \"kubernetes.io/projected/b33086d6-9b3e-4673-bd80-52913963f7b3-kube-api-access-vp8tx\") pod \"certified-operators-5m84m\" (UID: \"b33086d6-9b3e-4673-bd80-52913963f7b3\") " pod="openshift-marketplace/certified-operators-5m84m" Dec 11 02:52:16 crc kubenswrapper[4824]: I1211 02:52:16.986759 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b33086d6-9b3e-4673-bd80-52913963f7b3-catalog-content\") pod \"certified-operators-5m84m\" (UID: \"b33086d6-9b3e-4673-bd80-52913963f7b3\") " pod="openshift-marketplace/certified-operators-5m84m" Dec 11 02:52:16 crc kubenswrapper[4824]: I1211 02:52:16.986852 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b33086d6-9b3e-4673-bd80-52913963f7b3-utilities\") pod \"certified-operators-5m84m\" (UID: \"b33086d6-9b3e-4673-bd80-52913963f7b3\") " pod="openshift-marketplace/certified-operators-5m84m" Dec 11 02:52:16 crc kubenswrapper[4824]: I1211 02:52:16.987142 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b33086d6-9b3e-4673-bd80-52913963f7b3-catalog-content\") pod \"certified-operators-5m84m\" (UID: \"b33086d6-9b3e-4673-bd80-52913963f7b3\") " pod="openshift-marketplace/certified-operators-5m84m" Dec 11 02:52:17 crc kubenswrapper[4824]: I1211 02:52:17.007242 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vp8tx\" (UniqueName: \"kubernetes.io/projected/b33086d6-9b3e-4673-bd80-52913963f7b3-kube-api-access-vp8tx\") pod \"certified-operators-5m84m\" (UID: \"b33086d6-9b3e-4673-bd80-52913963f7b3\") " pod="openshift-marketplace/certified-operators-5m84m" Dec 11 02:52:17 crc kubenswrapper[4824]: I1211 02:52:17.104551 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5m84m" Dec 11 02:52:17 crc kubenswrapper[4824]: I1211 02:52:17.518966 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jlblg"] Dec 11 02:52:17 crc kubenswrapper[4824]: I1211 02:52:17.659573 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5m84m"] Dec 11 02:52:17 crc kubenswrapper[4824]: W1211 02:52:17.669156 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb33086d6_9b3e_4673_bd80_52913963f7b3.slice/crio-4a350bdf94c5c4940b0e55ed4021f18afb5f9c52bfa692f385c8a3bd2bce18e9 WatchSource:0}: Error finding container 4a350bdf94c5c4940b0e55ed4021f18afb5f9c52bfa692f385c8a3bd2bce18e9: Status 404 returned error can't find the container with id 4a350bdf94c5c4940b0e55ed4021f18afb5f9c52bfa692f385c8a3bd2bce18e9 Dec 11 02:52:17 crc kubenswrapper[4824]: E1211 02:52:17.856622 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5f4d993a_bd6a_4ba8_bbbb_d389c2b33015.slice/crio-conmon-90227acee0755a7f586dada4f49bd7e9dc65c73563cfa3361d8efc5f45cd2c36.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5f4d993a_bd6a_4ba8_bbbb_d389c2b33015.slice/crio-90227acee0755a7f586dada4f49bd7e9dc65c73563cfa3361d8efc5f45cd2c36.scope\": RecentStats: unable to find data in memory cache]" Dec 11 02:52:17 crc kubenswrapper[4824]: I1211 02:52:17.918307 4824 generic.go:334] "Generic (PLEG): container finished" podID="b33086d6-9b3e-4673-bd80-52913963f7b3" containerID="46c54a93056956ed7c077a56eabc854b3acee4d2cdd3391b09a4061e1d7f6376" exitCode=0 Dec 11 02:52:17 crc kubenswrapper[4824]: I1211 02:52:17.918554 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5m84m" event={"ID":"b33086d6-9b3e-4673-bd80-52913963f7b3","Type":"ContainerDied","Data":"46c54a93056956ed7c077a56eabc854b3acee4d2cdd3391b09a4061e1d7f6376"} Dec 11 02:52:17 crc kubenswrapper[4824]: I1211 02:52:17.918589 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5m84m" event={"ID":"b33086d6-9b3e-4673-bd80-52913963f7b3","Type":"ContainerStarted","Data":"4a350bdf94c5c4940b0e55ed4021f18afb5f9c52bfa692f385c8a3bd2bce18e9"} Dec 11 02:52:17 crc kubenswrapper[4824]: I1211 02:52:17.920508 4824 generic.go:334] "Generic (PLEG): container finished" podID="5f4d993a-bd6a-4ba8-bbbb-d389c2b33015" containerID="90227acee0755a7f586dada4f49bd7e9dc65c73563cfa3361d8efc5f45cd2c36" exitCode=0 Dec 11 02:52:17 crc kubenswrapper[4824]: I1211 02:52:17.920527 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jlblg" event={"ID":"5f4d993a-bd6a-4ba8-bbbb-d389c2b33015","Type":"ContainerDied","Data":"90227acee0755a7f586dada4f49bd7e9dc65c73563cfa3361d8efc5f45cd2c36"} Dec 11 02:52:17 crc kubenswrapper[4824]: I1211 02:52:17.920540 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jlblg" event={"ID":"5f4d993a-bd6a-4ba8-bbbb-d389c2b33015","Type":"ContainerStarted","Data":"341c56e28334be1bc0f4b2893a0cde177feb5154a79b6e4ee1d19db4adc36928"} Dec 11 02:52:18 crc kubenswrapper[4824]: I1211 02:52:18.933985 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5m84m" event={"ID":"b33086d6-9b3e-4673-bd80-52913963f7b3","Type":"ContainerStarted","Data":"4ec219f3c5938ba8e5ae00dd84ff18ff78aa4a373addf7839364e7f80267b362"} Dec 11 02:52:19 crc kubenswrapper[4824]: I1211 02:52:19.950182 4824 generic.go:334] "Generic (PLEG): container finished" podID="b33086d6-9b3e-4673-bd80-52913963f7b3" containerID="4ec219f3c5938ba8e5ae00dd84ff18ff78aa4a373addf7839364e7f80267b362" exitCode=0 Dec 11 02:52:19 crc kubenswrapper[4824]: I1211 02:52:19.950477 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5m84m" event={"ID":"b33086d6-9b3e-4673-bd80-52913963f7b3","Type":"ContainerDied","Data":"4ec219f3c5938ba8e5ae00dd84ff18ff78aa4a373addf7839364e7f80267b362"} Dec 11 02:52:19 crc kubenswrapper[4824]: I1211 02:52:19.956504 4824 generic.go:334] "Generic (PLEG): container finished" podID="5f4d993a-bd6a-4ba8-bbbb-d389c2b33015" containerID="3990facf3638a7d494066f06707c242937f30a0dd7a09d75b9cfdfdd91904195" exitCode=0 Dec 11 02:52:19 crc kubenswrapper[4824]: I1211 02:52:19.956570 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jlblg" event={"ID":"5f4d993a-bd6a-4ba8-bbbb-d389c2b33015","Type":"ContainerDied","Data":"3990facf3638a7d494066f06707c242937f30a0dd7a09d75b9cfdfdd91904195"} Dec 11 02:52:21 crc kubenswrapper[4824]: I1211 02:52:21.983004 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5m84m" event={"ID":"b33086d6-9b3e-4673-bd80-52913963f7b3","Type":"ContainerStarted","Data":"b6b060cf2556c7d0029e47cf28061fc9adb4c9e1bb6d37c038ffcbf88e2fe2bf"} Dec 11 02:52:21 crc kubenswrapper[4824]: I1211 02:52:21.988452 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jlblg" event={"ID":"5f4d993a-bd6a-4ba8-bbbb-d389c2b33015","Type":"ContainerStarted","Data":"479dde395c206033464c506f4e0db10b950394c8cc165998491a253be1696a10"} Dec 11 02:52:22 crc kubenswrapper[4824]: I1211 02:52:22.027715 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-5m84m" podStartSLOduration=2.634842145 podStartE2EDuration="6.027669847s" podCreationTimestamp="2025-12-11 02:52:16 +0000 UTC" firstStartedPulling="2025-12-11 02:52:17.9205015 +0000 UTC m=+3079.609538879" lastFinishedPulling="2025-12-11 02:52:21.313329212 +0000 UTC m=+3083.002366581" observedRunningTime="2025-12-11 02:52:22.009242558 +0000 UTC m=+3083.698279967" watchObservedRunningTime="2025-12-11 02:52:22.027669847 +0000 UTC m=+3083.716707246" Dec 11 02:52:22 crc kubenswrapper[4824]: I1211 02:52:22.047208 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-jlblg" podStartSLOduration=2.731190482 podStartE2EDuration="6.047175455s" podCreationTimestamp="2025-12-11 02:52:16 +0000 UTC" firstStartedPulling="2025-12-11 02:52:17.921507207 +0000 UTC m=+3079.610544586" lastFinishedPulling="2025-12-11 02:52:21.23749216 +0000 UTC m=+3082.926529559" observedRunningTime="2025-12-11 02:52:22.04219702 +0000 UTC m=+3083.731234499" watchObservedRunningTime="2025-12-11 02:52:22.047175455 +0000 UTC m=+3083.736212914" Dec 11 02:52:23 crc kubenswrapper[4824]: I1211 02:52:23.645458 4824 scope.go:117] "RemoveContainer" containerID="c19d806d2c030c4af9eb013dc9d405f67612520a0b85a1ecd0356d85a1e2fd14" Dec 11 02:52:23 crc kubenswrapper[4824]: E1211 02:52:23.647709 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 02:52:26 crc kubenswrapper[4824]: I1211 02:52:26.958024 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-jlblg" Dec 11 02:52:26 crc kubenswrapper[4824]: I1211 02:52:26.958887 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-jlblg" Dec 11 02:52:27 crc kubenswrapper[4824]: I1211 02:52:27.047300 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-jlblg" Dec 11 02:52:27 crc kubenswrapper[4824]: I1211 02:52:27.106658 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-5m84m" Dec 11 02:52:27 crc kubenswrapper[4824]: I1211 02:52:27.106739 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-5m84m" Dec 11 02:52:27 crc kubenswrapper[4824]: I1211 02:52:27.137708 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-jlblg" Dec 11 02:52:27 crc kubenswrapper[4824]: I1211 02:52:27.167394 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-5m84m" Dec 11 02:52:27 crc kubenswrapper[4824]: I1211 02:52:27.355605 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jlblg"] Dec 11 02:52:28 crc kubenswrapper[4824]: I1211 02:52:28.105388 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-5m84m" Dec 11 02:52:29 crc kubenswrapper[4824]: I1211 02:52:29.056012 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-jlblg" podUID="5f4d993a-bd6a-4ba8-bbbb-d389c2b33015" containerName="registry-server" containerID="cri-o://479dde395c206033464c506f4e0db10b950394c8cc165998491a253be1696a10" gracePeriod=2 Dec 11 02:52:29 crc kubenswrapper[4824]: I1211 02:52:29.553988 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5m84m"] Dec 11 02:52:29 crc kubenswrapper[4824]: I1211 02:52:29.608784 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jlblg" Dec 11 02:52:29 crc kubenswrapper[4824]: I1211 02:52:29.768661 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f4d993a-bd6a-4ba8-bbbb-d389c2b33015-utilities\") pod \"5f4d993a-bd6a-4ba8-bbbb-d389c2b33015\" (UID: \"5f4d993a-bd6a-4ba8-bbbb-d389c2b33015\") " Dec 11 02:52:29 crc kubenswrapper[4824]: I1211 02:52:29.768830 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f4d993a-bd6a-4ba8-bbbb-d389c2b33015-catalog-content\") pod \"5f4d993a-bd6a-4ba8-bbbb-d389c2b33015\" (UID: \"5f4d993a-bd6a-4ba8-bbbb-d389c2b33015\") " Dec 11 02:52:29 crc kubenswrapper[4824]: I1211 02:52:29.768913 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-542dn\" (UniqueName: \"kubernetes.io/projected/5f4d993a-bd6a-4ba8-bbbb-d389c2b33015-kube-api-access-542dn\") pod \"5f4d993a-bd6a-4ba8-bbbb-d389c2b33015\" (UID: \"5f4d993a-bd6a-4ba8-bbbb-d389c2b33015\") " Dec 11 02:52:29 crc kubenswrapper[4824]: I1211 02:52:29.774456 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f4d993a-bd6a-4ba8-bbbb-d389c2b33015-kube-api-access-542dn" (OuterVolumeSpecName: "kube-api-access-542dn") pod "5f4d993a-bd6a-4ba8-bbbb-d389c2b33015" (UID: "5f4d993a-bd6a-4ba8-bbbb-d389c2b33015"). InnerVolumeSpecName "kube-api-access-542dn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:52:29 crc kubenswrapper[4824]: I1211 02:52:29.778315 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5f4d993a-bd6a-4ba8-bbbb-d389c2b33015-utilities" (OuterVolumeSpecName: "utilities") pod "5f4d993a-bd6a-4ba8-bbbb-d389c2b33015" (UID: "5f4d993a-bd6a-4ba8-bbbb-d389c2b33015"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:52:29 crc kubenswrapper[4824]: I1211 02:52:29.819648 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5f4d993a-bd6a-4ba8-bbbb-d389c2b33015-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5f4d993a-bd6a-4ba8-bbbb-d389c2b33015" (UID: "5f4d993a-bd6a-4ba8-bbbb-d389c2b33015"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:52:29 crc kubenswrapper[4824]: I1211 02:52:29.870994 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-542dn\" (UniqueName: \"kubernetes.io/projected/5f4d993a-bd6a-4ba8-bbbb-d389c2b33015-kube-api-access-542dn\") on node \"crc\" DevicePath \"\"" Dec 11 02:52:29 crc kubenswrapper[4824]: I1211 02:52:29.871023 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f4d993a-bd6a-4ba8-bbbb-d389c2b33015-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 02:52:29 crc kubenswrapper[4824]: I1211 02:52:29.871032 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f4d993a-bd6a-4ba8-bbbb-d389c2b33015-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 02:52:30 crc kubenswrapper[4824]: I1211 02:52:30.070948 4824 generic.go:334] "Generic (PLEG): container finished" podID="5f4d993a-bd6a-4ba8-bbbb-d389c2b33015" containerID="479dde395c206033464c506f4e0db10b950394c8cc165998491a253be1696a10" exitCode=0 Dec 11 02:52:30 crc kubenswrapper[4824]: I1211 02:52:30.071036 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jlblg" Dec 11 02:52:30 crc kubenswrapper[4824]: I1211 02:52:30.071037 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jlblg" event={"ID":"5f4d993a-bd6a-4ba8-bbbb-d389c2b33015","Type":"ContainerDied","Data":"479dde395c206033464c506f4e0db10b950394c8cc165998491a253be1696a10"} Dec 11 02:52:30 crc kubenswrapper[4824]: I1211 02:52:30.071167 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jlblg" event={"ID":"5f4d993a-bd6a-4ba8-bbbb-d389c2b33015","Type":"ContainerDied","Data":"341c56e28334be1bc0f4b2893a0cde177feb5154a79b6e4ee1d19db4adc36928"} Dec 11 02:52:30 crc kubenswrapper[4824]: I1211 02:52:30.071197 4824 scope.go:117] "RemoveContainer" containerID="479dde395c206033464c506f4e0db10b950394c8cc165998491a253be1696a10" Dec 11 02:52:30 crc kubenswrapper[4824]: I1211 02:52:30.071889 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-5m84m" podUID="b33086d6-9b3e-4673-bd80-52913963f7b3" containerName="registry-server" containerID="cri-o://b6b060cf2556c7d0029e47cf28061fc9adb4c9e1bb6d37c038ffcbf88e2fe2bf" gracePeriod=2 Dec 11 02:52:30 crc kubenswrapper[4824]: I1211 02:52:30.136188 4824 scope.go:117] "RemoveContainer" containerID="3990facf3638a7d494066f06707c242937f30a0dd7a09d75b9cfdfdd91904195" Dec 11 02:52:30 crc kubenswrapper[4824]: I1211 02:52:30.141292 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jlblg"] Dec 11 02:52:30 crc kubenswrapper[4824]: I1211 02:52:30.151177 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-jlblg"] Dec 11 02:52:30 crc kubenswrapper[4824]: I1211 02:52:30.177224 4824 scope.go:117] "RemoveContainer" containerID="90227acee0755a7f586dada4f49bd7e9dc65c73563cfa3361d8efc5f45cd2c36" Dec 11 02:52:30 crc kubenswrapper[4824]: I1211 02:52:30.339915 4824 scope.go:117] "RemoveContainer" containerID="479dde395c206033464c506f4e0db10b950394c8cc165998491a253be1696a10" Dec 11 02:52:30 crc kubenswrapper[4824]: E1211 02:52:30.340678 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"479dde395c206033464c506f4e0db10b950394c8cc165998491a253be1696a10\": container with ID starting with 479dde395c206033464c506f4e0db10b950394c8cc165998491a253be1696a10 not found: ID does not exist" containerID="479dde395c206033464c506f4e0db10b950394c8cc165998491a253be1696a10" Dec 11 02:52:30 crc kubenswrapper[4824]: I1211 02:52:30.340727 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"479dde395c206033464c506f4e0db10b950394c8cc165998491a253be1696a10"} err="failed to get container status \"479dde395c206033464c506f4e0db10b950394c8cc165998491a253be1696a10\": rpc error: code = NotFound desc = could not find container \"479dde395c206033464c506f4e0db10b950394c8cc165998491a253be1696a10\": container with ID starting with 479dde395c206033464c506f4e0db10b950394c8cc165998491a253be1696a10 not found: ID does not exist" Dec 11 02:52:30 crc kubenswrapper[4824]: I1211 02:52:30.340759 4824 scope.go:117] "RemoveContainer" containerID="3990facf3638a7d494066f06707c242937f30a0dd7a09d75b9cfdfdd91904195" Dec 11 02:52:30 crc kubenswrapper[4824]: E1211 02:52:30.342155 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3990facf3638a7d494066f06707c242937f30a0dd7a09d75b9cfdfdd91904195\": container with ID starting with 3990facf3638a7d494066f06707c242937f30a0dd7a09d75b9cfdfdd91904195 not found: ID does not exist" containerID="3990facf3638a7d494066f06707c242937f30a0dd7a09d75b9cfdfdd91904195" Dec 11 02:52:30 crc kubenswrapper[4824]: I1211 02:52:30.342189 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3990facf3638a7d494066f06707c242937f30a0dd7a09d75b9cfdfdd91904195"} err="failed to get container status \"3990facf3638a7d494066f06707c242937f30a0dd7a09d75b9cfdfdd91904195\": rpc error: code = NotFound desc = could not find container \"3990facf3638a7d494066f06707c242937f30a0dd7a09d75b9cfdfdd91904195\": container with ID starting with 3990facf3638a7d494066f06707c242937f30a0dd7a09d75b9cfdfdd91904195 not found: ID does not exist" Dec 11 02:52:30 crc kubenswrapper[4824]: I1211 02:52:30.342216 4824 scope.go:117] "RemoveContainer" containerID="90227acee0755a7f586dada4f49bd7e9dc65c73563cfa3361d8efc5f45cd2c36" Dec 11 02:52:30 crc kubenswrapper[4824]: E1211 02:52:30.342497 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90227acee0755a7f586dada4f49bd7e9dc65c73563cfa3361d8efc5f45cd2c36\": container with ID starting with 90227acee0755a7f586dada4f49bd7e9dc65c73563cfa3361d8efc5f45cd2c36 not found: ID does not exist" containerID="90227acee0755a7f586dada4f49bd7e9dc65c73563cfa3361d8efc5f45cd2c36" Dec 11 02:52:30 crc kubenswrapper[4824]: I1211 02:52:30.342531 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90227acee0755a7f586dada4f49bd7e9dc65c73563cfa3361d8efc5f45cd2c36"} err="failed to get container status \"90227acee0755a7f586dada4f49bd7e9dc65c73563cfa3361d8efc5f45cd2c36\": rpc error: code = NotFound desc = could not find container \"90227acee0755a7f586dada4f49bd7e9dc65c73563cfa3361d8efc5f45cd2c36\": container with ID starting with 90227acee0755a7f586dada4f49bd7e9dc65c73563cfa3361d8efc5f45cd2c36 not found: ID does not exist" Dec 11 02:52:30 crc kubenswrapper[4824]: I1211 02:52:30.604806 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5m84m" Dec 11 02:52:30 crc kubenswrapper[4824]: I1211 02:52:30.633689 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b33086d6-9b3e-4673-bd80-52913963f7b3-catalog-content\") pod \"b33086d6-9b3e-4673-bd80-52913963f7b3\" (UID: \"b33086d6-9b3e-4673-bd80-52913963f7b3\") " Dec 11 02:52:30 crc kubenswrapper[4824]: I1211 02:52:30.633822 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vp8tx\" (UniqueName: \"kubernetes.io/projected/b33086d6-9b3e-4673-bd80-52913963f7b3-kube-api-access-vp8tx\") pod \"b33086d6-9b3e-4673-bd80-52913963f7b3\" (UID: \"b33086d6-9b3e-4673-bd80-52913963f7b3\") " Dec 11 02:52:30 crc kubenswrapper[4824]: I1211 02:52:30.634063 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b33086d6-9b3e-4673-bd80-52913963f7b3-utilities\") pod \"b33086d6-9b3e-4673-bd80-52913963f7b3\" (UID: \"b33086d6-9b3e-4673-bd80-52913963f7b3\") " Dec 11 02:52:30 crc kubenswrapper[4824]: I1211 02:52:30.634978 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b33086d6-9b3e-4673-bd80-52913963f7b3-utilities" (OuterVolumeSpecName: "utilities") pod "b33086d6-9b3e-4673-bd80-52913963f7b3" (UID: "b33086d6-9b3e-4673-bd80-52913963f7b3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:52:30 crc kubenswrapper[4824]: I1211 02:52:30.641315 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b33086d6-9b3e-4673-bd80-52913963f7b3-kube-api-access-vp8tx" (OuterVolumeSpecName: "kube-api-access-vp8tx") pod "b33086d6-9b3e-4673-bd80-52913963f7b3" (UID: "b33086d6-9b3e-4673-bd80-52913963f7b3"). InnerVolumeSpecName "kube-api-access-vp8tx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:52:30 crc kubenswrapper[4824]: I1211 02:52:30.651933 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5f4d993a-bd6a-4ba8-bbbb-d389c2b33015" path="/var/lib/kubelet/pods/5f4d993a-bd6a-4ba8-bbbb-d389c2b33015/volumes" Dec 11 02:52:30 crc kubenswrapper[4824]: I1211 02:52:30.699091 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b33086d6-9b3e-4673-bd80-52913963f7b3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b33086d6-9b3e-4673-bd80-52913963f7b3" (UID: "b33086d6-9b3e-4673-bd80-52913963f7b3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:52:30 crc kubenswrapper[4824]: I1211 02:52:30.736402 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vp8tx\" (UniqueName: \"kubernetes.io/projected/b33086d6-9b3e-4673-bd80-52913963f7b3-kube-api-access-vp8tx\") on node \"crc\" DevicePath \"\"" Dec 11 02:52:30 crc kubenswrapper[4824]: I1211 02:52:30.736444 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b33086d6-9b3e-4673-bd80-52913963f7b3-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 02:52:30 crc kubenswrapper[4824]: I1211 02:52:30.736453 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b33086d6-9b3e-4673-bd80-52913963f7b3-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 02:52:31 crc kubenswrapper[4824]: I1211 02:52:31.091734 4824 generic.go:334] "Generic (PLEG): container finished" podID="b33086d6-9b3e-4673-bd80-52913963f7b3" containerID="b6b060cf2556c7d0029e47cf28061fc9adb4c9e1bb6d37c038ffcbf88e2fe2bf" exitCode=0 Dec 11 02:52:31 crc kubenswrapper[4824]: I1211 02:52:31.091861 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5m84m" event={"ID":"b33086d6-9b3e-4673-bd80-52913963f7b3","Type":"ContainerDied","Data":"b6b060cf2556c7d0029e47cf28061fc9adb4c9e1bb6d37c038ffcbf88e2fe2bf"} Dec 11 02:52:31 crc kubenswrapper[4824]: I1211 02:52:31.092103 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5m84m" event={"ID":"b33086d6-9b3e-4673-bd80-52913963f7b3","Type":"ContainerDied","Data":"4a350bdf94c5c4940b0e55ed4021f18afb5f9c52bfa692f385c8a3bd2bce18e9"} Dec 11 02:52:31 crc kubenswrapper[4824]: I1211 02:52:31.091892 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5m84m" Dec 11 02:52:31 crc kubenswrapper[4824]: I1211 02:52:31.092191 4824 scope.go:117] "RemoveContainer" containerID="b6b060cf2556c7d0029e47cf28061fc9adb4c9e1bb6d37c038ffcbf88e2fe2bf" Dec 11 02:52:31 crc kubenswrapper[4824]: I1211 02:52:31.116319 4824 scope.go:117] "RemoveContainer" containerID="4ec219f3c5938ba8e5ae00dd84ff18ff78aa4a373addf7839364e7f80267b362" Dec 11 02:52:31 crc kubenswrapper[4824]: I1211 02:52:31.148188 4824 scope.go:117] "RemoveContainer" containerID="46c54a93056956ed7c077a56eabc854b3acee4d2cdd3391b09a4061e1d7f6376" Dec 11 02:52:31 crc kubenswrapper[4824]: I1211 02:52:31.160622 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5m84m"] Dec 11 02:52:31 crc kubenswrapper[4824]: I1211 02:52:31.179656 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-5m84m"] Dec 11 02:52:31 crc kubenswrapper[4824]: I1211 02:52:31.183593 4824 scope.go:117] "RemoveContainer" containerID="b6b060cf2556c7d0029e47cf28061fc9adb4c9e1bb6d37c038ffcbf88e2fe2bf" Dec 11 02:52:31 crc kubenswrapper[4824]: E1211 02:52:31.184308 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b6b060cf2556c7d0029e47cf28061fc9adb4c9e1bb6d37c038ffcbf88e2fe2bf\": container with ID starting with b6b060cf2556c7d0029e47cf28061fc9adb4c9e1bb6d37c038ffcbf88e2fe2bf not found: ID does not exist" containerID="b6b060cf2556c7d0029e47cf28061fc9adb4c9e1bb6d37c038ffcbf88e2fe2bf" Dec 11 02:52:31 crc kubenswrapper[4824]: I1211 02:52:31.184345 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b6b060cf2556c7d0029e47cf28061fc9adb4c9e1bb6d37c038ffcbf88e2fe2bf"} err="failed to get container status \"b6b060cf2556c7d0029e47cf28061fc9adb4c9e1bb6d37c038ffcbf88e2fe2bf\": rpc error: code = NotFound desc = could not find container \"b6b060cf2556c7d0029e47cf28061fc9adb4c9e1bb6d37c038ffcbf88e2fe2bf\": container with ID starting with b6b060cf2556c7d0029e47cf28061fc9adb4c9e1bb6d37c038ffcbf88e2fe2bf not found: ID does not exist" Dec 11 02:52:31 crc kubenswrapper[4824]: I1211 02:52:31.184369 4824 scope.go:117] "RemoveContainer" containerID="4ec219f3c5938ba8e5ae00dd84ff18ff78aa4a373addf7839364e7f80267b362" Dec 11 02:52:31 crc kubenswrapper[4824]: E1211 02:52:31.184843 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4ec219f3c5938ba8e5ae00dd84ff18ff78aa4a373addf7839364e7f80267b362\": container with ID starting with 4ec219f3c5938ba8e5ae00dd84ff18ff78aa4a373addf7839364e7f80267b362 not found: ID does not exist" containerID="4ec219f3c5938ba8e5ae00dd84ff18ff78aa4a373addf7839364e7f80267b362" Dec 11 02:52:31 crc kubenswrapper[4824]: I1211 02:52:31.184869 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ec219f3c5938ba8e5ae00dd84ff18ff78aa4a373addf7839364e7f80267b362"} err="failed to get container status \"4ec219f3c5938ba8e5ae00dd84ff18ff78aa4a373addf7839364e7f80267b362\": rpc error: code = NotFound desc = could not find container \"4ec219f3c5938ba8e5ae00dd84ff18ff78aa4a373addf7839364e7f80267b362\": container with ID starting with 4ec219f3c5938ba8e5ae00dd84ff18ff78aa4a373addf7839364e7f80267b362 not found: ID does not exist" Dec 11 02:52:31 crc kubenswrapper[4824]: I1211 02:52:31.184887 4824 scope.go:117] "RemoveContainer" containerID="46c54a93056956ed7c077a56eabc854b3acee4d2cdd3391b09a4061e1d7f6376" Dec 11 02:52:31 crc kubenswrapper[4824]: E1211 02:52:31.185200 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"46c54a93056956ed7c077a56eabc854b3acee4d2cdd3391b09a4061e1d7f6376\": container with ID starting with 46c54a93056956ed7c077a56eabc854b3acee4d2cdd3391b09a4061e1d7f6376 not found: ID does not exist" containerID="46c54a93056956ed7c077a56eabc854b3acee4d2cdd3391b09a4061e1d7f6376" Dec 11 02:52:31 crc kubenswrapper[4824]: I1211 02:52:31.185252 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46c54a93056956ed7c077a56eabc854b3acee4d2cdd3391b09a4061e1d7f6376"} err="failed to get container status \"46c54a93056956ed7c077a56eabc854b3acee4d2cdd3391b09a4061e1d7f6376\": rpc error: code = NotFound desc = could not find container \"46c54a93056956ed7c077a56eabc854b3acee4d2cdd3391b09a4061e1d7f6376\": container with ID starting with 46c54a93056956ed7c077a56eabc854b3acee4d2cdd3391b09a4061e1d7f6376 not found: ID does not exist" Dec 11 02:52:32 crc kubenswrapper[4824]: I1211 02:52:32.645794 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b33086d6-9b3e-4673-bd80-52913963f7b3" path="/var/lib/kubelet/pods/b33086d6-9b3e-4673-bd80-52913963f7b3/volumes" Dec 11 02:52:34 crc kubenswrapper[4824]: I1211 02:52:34.633379 4824 scope.go:117] "RemoveContainer" containerID="c19d806d2c030c4af9eb013dc9d405f67612520a0b85a1ecd0356d85a1e2fd14" Dec 11 02:52:34 crc kubenswrapper[4824]: E1211 02:52:34.633871 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 02:52:45 crc kubenswrapper[4824]: I1211 02:52:45.633253 4824 scope.go:117] "RemoveContainer" containerID="c19d806d2c030c4af9eb013dc9d405f67612520a0b85a1ecd0356d85a1e2fd14" Dec 11 02:52:45 crc kubenswrapper[4824]: E1211 02:52:45.634278 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 02:52:56 crc kubenswrapper[4824]: I1211 02:52:56.638314 4824 scope.go:117] "RemoveContainer" containerID="c19d806d2c030c4af9eb013dc9d405f67612520a0b85a1ecd0356d85a1e2fd14" Dec 11 02:52:56 crc kubenswrapper[4824]: E1211 02:52:56.639560 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 02:53:10 crc kubenswrapper[4824]: I1211 02:53:10.634184 4824 scope.go:117] "RemoveContainer" containerID="c19d806d2c030c4af9eb013dc9d405f67612520a0b85a1ecd0356d85a1e2fd14" Dec 11 02:53:10 crc kubenswrapper[4824]: E1211 02:53:10.635406 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 02:53:25 crc kubenswrapper[4824]: I1211 02:53:25.662615 4824 scope.go:117] "RemoveContainer" containerID="c19d806d2c030c4af9eb013dc9d405f67612520a0b85a1ecd0356d85a1e2fd14" Dec 11 02:53:25 crc kubenswrapper[4824]: E1211 02:53:25.665650 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 02:53:40 crc kubenswrapper[4824]: I1211 02:53:40.633084 4824 scope.go:117] "RemoveContainer" containerID="c19d806d2c030c4af9eb013dc9d405f67612520a0b85a1ecd0356d85a1e2fd14" Dec 11 02:53:40 crc kubenswrapper[4824]: E1211 02:53:40.633739 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 02:53:51 crc kubenswrapper[4824]: I1211 02:53:51.636618 4824 scope.go:117] "RemoveContainer" containerID="c19d806d2c030c4af9eb013dc9d405f67612520a0b85a1ecd0356d85a1e2fd14" Dec 11 02:53:51 crc kubenswrapper[4824]: E1211 02:53:51.638243 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 02:54:03 crc kubenswrapper[4824]: I1211 02:54:03.633275 4824 scope.go:117] "RemoveContainer" containerID="c19d806d2c030c4af9eb013dc9d405f67612520a0b85a1ecd0356d85a1e2fd14" Dec 11 02:54:03 crc kubenswrapper[4824]: E1211 02:54:03.634422 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 02:54:16 crc kubenswrapper[4824]: I1211 02:54:16.633921 4824 scope.go:117] "RemoveContainer" containerID="c19d806d2c030c4af9eb013dc9d405f67612520a0b85a1ecd0356d85a1e2fd14" Dec 11 02:54:17 crc kubenswrapper[4824]: I1211 02:54:17.268364 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" event={"ID":"44b156e4-64a4-4d45-aa5e-9b10a862faed","Type":"ContainerStarted","Data":"b4f54810df237c174e9f9c1b3d2b74efaf5ad20b26965b89c1000a0ab8e02ced"} Dec 11 02:56:37 crc kubenswrapper[4824]: I1211 02:56:37.251555 4824 patch_prober.go:28] interesting pod/machine-config-daemon-gx6xt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 02:56:37 crc kubenswrapper[4824]: I1211 02:56:37.252041 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 02:56:38 crc kubenswrapper[4824]: I1211 02:56:38.068570 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-2bncf"] Dec 11 02:56:38 crc kubenswrapper[4824]: E1211 02:56:38.069204 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f4d993a-bd6a-4ba8-bbbb-d389c2b33015" containerName="extract-content" Dec 11 02:56:38 crc kubenswrapper[4824]: I1211 02:56:38.069220 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f4d993a-bd6a-4ba8-bbbb-d389c2b33015" containerName="extract-content" Dec 11 02:56:38 crc kubenswrapper[4824]: E1211 02:56:38.069248 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b33086d6-9b3e-4673-bd80-52913963f7b3" containerName="registry-server" Dec 11 02:56:38 crc kubenswrapper[4824]: I1211 02:56:38.069255 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="b33086d6-9b3e-4673-bd80-52913963f7b3" containerName="registry-server" Dec 11 02:56:38 crc kubenswrapper[4824]: E1211 02:56:38.069268 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b33086d6-9b3e-4673-bd80-52913963f7b3" containerName="extract-content" Dec 11 02:56:38 crc kubenswrapper[4824]: I1211 02:56:38.069274 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="b33086d6-9b3e-4673-bd80-52913963f7b3" containerName="extract-content" Dec 11 02:56:38 crc kubenswrapper[4824]: E1211 02:56:38.069288 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b33086d6-9b3e-4673-bd80-52913963f7b3" containerName="extract-utilities" Dec 11 02:56:38 crc kubenswrapper[4824]: I1211 02:56:38.069294 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="b33086d6-9b3e-4673-bd80-52913963f7b3" containerName="extract-utilities" Dec 11 02:56:38 crc kubenswrapper[4824]: E1211 02:56:38.069307 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f4d993a-bd6a-4ba8-bbbb-d389c2b33015" containerName="extract-utilities" Dec 11 02:56:38 crc kubenswrapper[4824]: I1211 02:56:38.069313 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f4d993a-bd6a-4ba8-bbbb-d389c2b33015" containerName="extract-utilities" Dec 11 02:56:38 crc kubenswrapper[4824]: E1211 02:56:38.069326 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f4d993a-bd6a-4ba8-bbbb-d389c2b33015" containerName="registry-server" Dec 11 02:56:38 crc kubenswrapper[4824]: I1211 02:56:38.069331 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f4d993a-bd6a-4ba8-bbbb-d389c2b33015" containerName="registry-server" Dec 11 02:56:38 crc kubenswrapper[4824]: I1211 02:56:38.069740 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="b33086d6-9b3e-4673-bd80-52913963f7b3" containerName="registry-server" Dec 11 02:56:38 crc kubenswrapper[4824]: I1211 02:56:38.070159 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f4d993a-bd6a-4ba8-bbbb-d389c2b33015" containerName="registry-server" Dec 11 02:56:38 crc kubenswrapper[4824]: I1211 02:56:38.071936 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2bncf" Dec 11 02:56:38 crc kubenswrapper[4824]: I1211 02:56:38.079358 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2bncf"] Dec 11 02:56:38 crc kubenswrapper[4824]: I1211 02:56:38.201752 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d343c6a4-1223-4557-ba6c-d92afde1afc9-catalog-content\") pod \"redhat-marketplace-2bncf\" (UID: \"d343c6a4-1223-4557-ba6c-d92afde1afc9\") " pod="openshift-marketplace/redhat-marketplace-2bncf" Dec 11 02:56:38 crc kubenswrapper[4824]: I1211 02:56:38.201930 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dqrk2\" (UniqueName: \"kubernetes.io/projected/d343c6a4-1223-4557-ba6c-d92afde1afc9-kube-api-access-dqrk2\") pod \"redhat-marketplace-2bncf\" (UID: \"d343c6a4-1223-4557-ba6c-d92afde1afc9\") " pod="openshift-marketplace/redhat-marketplace-2bncf" Dec 11 02:56:38 crc kubenswrapper[4824]: I1211 02:56:38.202147 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d343c6a4-1223-4557-ba6c-d92afde1afc9-utilities\") pod \"redhat-marketplace-2bncf\" (UID: \"d343c6a4-1223-4557-ba6c-d92afde1afc9\") " pod="openshift-marketplace/redhat-marketplace-2bncf" Dec 11 02:56:38 crc kubenswrapper[4824]: I1211 02:56:38.304370 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d343c6a4-1223-4557-ba6c-d92afde1afc9-catalog-content\") pod \"redhat-marketplace-2bncf\" (UID: \"d343c6a4-1223-4557-ba6c-d92afde1afc9\") " pod="openshift-marketplace/redhat-marketplace-2bncf" Dec 11 02:56:38 crc kubenswrapper[4824]: I1211 02:56:38.304463 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dqrk2\" (UniqueName: \"kubernetes.io/projected/d343c6a4-1223-4557-ba6c-d92afde1afc9-kube-api-access-dqrk2\") pod \"redhat-marketplace-2bncf\" (UID: \"d343c6a4-1223-4557-ba6c-d92afde1afc9\") " pod="openshift-marketplace/redhat-marketplace-2bncf" Dec 11 02:56:38 crc kubenswrapper[4824]: I1211 02:56:38.304528 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d343c6a4-1223-4557-ba6c-d92afde1afc9-utilities\") pod \"redhat-marketplace-2bncf\" (UID: \"d343c6a4-1223-4557-ba6c-d92afde1afc9\") " pod="openshift-marketplace/redhat-marketplace-2bncf" Dec 11 02:56:38 crc kubenswrapper[4824]: I1211 02:56:38.304908 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d343c6a4-1223-4557-ba6c-d92afde1afc9-catalog-content\") pod \"redhat-marketplace-2bncf\" (UID: \"d343c6a4-1223-4557-ba6c-d92afde1afc9\") " pod="openshift-marketplace/redhat-marketplace-2bncf" Dec 11 02:56:38 crc kubenswrapper[4824]: I1211 02:56:38.305020 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d343c6a4-1223-4557-ba6c-d92afde1afc9-utilities\") pod \"redhat-marketplace-2bncf\" (UID: \"d343c6a4-1223-4557-ba6c-d92afde1afc9\") " pod="openshift-marketplace/redhat-marketplace-2bncf" Dec 11 02:56:38 crc kubenswrapper[4824]: I1211 02:56:38.332916 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dqrk2\" (UniqueName: \"kubernetes.io/projected/d343c6a4-1223-4557-ba6c-d92afde1afc9-kube-api-access-dqrk2\") pod \"redhat-marketplace-2bncf\" (UID: \"d343c6a4-1223-4557-ba6c-d92afde1afc9\") " pod="openshift-marketplace/redhat-marketplace-2bncf" Dec 11 02:56:38 crc kubenswrapper[4824]: I1211 02:56:38.403941 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2bncf" Dec 11 02:56:38 crc kubenswrapper[4824]: I1211 02:56:38.861992 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2bncf"] Dec 11 02:56:39 crc kubenswrapper[4824]: I1211 02:56:39.844047 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2bncf" event={"ID":"d343c6a4-1223-4557-ba6c-d92afde1afc9","Type":"ContainerDied","Data":"41c194011c3db553c2bcdff851affe196be2d21ca38f077812ea6e5bcb333009"} Dec 11 02:56:39 crc kubenswrapper[4824]: I1211 02:56:39.843975 4824 generic.go:334] "Generic (PLEG): container finished" podID="d343c6a4-1223-4557-ba6c-d92afde1afc9" containerID="41c194011c3db553c2bcdff851affe196be2d21ca38f077812ea6e5bcb333009" exitCode=0 Dec 11 02:56:39 crc kubenswrapper[4824]: I1211 02:56:39.844497 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2bncf" event={"ID":"d343c6a4-1223-4557-ba6c-d92afde1afc9","Type":"ContainerStarted","Data":"580e8747320051a91e27de711cf99c7d9ddf67f9f6362a4f7c27b9ce09369be9"} Dec 11 02:56:39 crc kubenswrapper[4824]: I1211 02:56:39.871047 4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 11 02:56:40 crc kubenswrapper[4824]: I1211 02:56:40.857493 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2bncf" event={"ID":"d343c6a4-1223-4557-ba6c-d92afde1afc9","Type":"ContainerStarted","Data":"858c897d5d8072e8c1ef62b86c4d6ade2e99d3634f71860fccaa6b7735505422"} Dec 11 02:56:41 crc kubenswrapper[4824]: I1211 02:56:41.874494 4824 generic.go:334] "Generic (PLEG): container finished" podID="d343c6a4-1223-4557-ba6c-d92afde1afc9" containerID="858c897d5d8072e8c1ef62b86c4d6ade2e99d3634f71860fccaa6b7735505422" exitCode=0 Dec 11 02:56:41 crc kubenswrapper[4824]: I1211 02:56:41.874603 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2bncf" event={"ID":"d343c6a4-1223-4557-ba6c-d92afde1afc9","Type":"ContainerDied","Data":"858c897d5d8072e8c1ef62b86c4d6ade2e99d3634f71860fccaa6b7735505422"} Dec 11 02:56:42 crc kubenswrapper[4824]: I1211 02:56:42.895921 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2bncf" event={"ID":"d343c6a4-1223-4557-ba6c-d92afde1afc9","Type":"ContainerStarted","Data":"cc7ec1a9ff3364ae1780b5e730d6af73271866d00a7bbde1bca00af920ffbd81"} Dec 11 02:56:42 crc kubenswrapper[4824]: I1211 02:56:42.918783 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-2bncf" podStartSLOduration=2.414452552 podStartE2EDuration="4.918756615s" podCreationTimestamp="2025-12-11 02:56:38 +0000 UTC" firstStartedPulling="2025-12-11 02:56:39.868863284 +0000 UTC m=+3341.557900703" lastFinishedPulling="2025-12-11 02:56:42.373167377 +0000 UTC m=+3344.062204766" observedRunningTime="2025-12-11 02:56:42.91416871 +0000 UTC m=+3344.603206089" watchObservedRunningTime="2025-12-11 02:56:42.918756615 +0000 UTC m=+3344.607794004" Dec 11 02:56:48 crc kubenswrapper[4824]: I1211 02:56:48.405041 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-2bncf" Dec 11 02:56:48 crc kubenswrapper[4824]: I1211 02:56:48.405552 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-2bncf" Dec 11 02:56:48 crc kubenswrapper[4824]: I1211 02:56:48.472938 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-2bncf" Dec 11 02:56:49 crc kubenswrapper[4824]: I1211 02:56:49.021067 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-2bncf" Dec 11 02:56:49 crc kubenswrapper[4824]: I1211 02:56:49.086640 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2bncf"] Dec 11 02:56:50 crc kubenswrapper[4824]: I1211 02:56:50.976201 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-2bncf" podUID="d343c6a4-1223-4557-ba6c-d92afde1afc9" containerName="registry-server" containerID="cri-o://cc7ec1a9ff3364ae1780b5e730d6af73271866d00a7bbde1bca00af920ffbd81" gracePeriod=2 Dec 11 02:56:51 crc kubenswrapper[4824]: I1211 02:56:51.988933 4824 generic.go:334] "Generic (PLEG): container finished" podID="d343c6a4-1223-4557-ba6c-d92afde1afc9" containerID="cc7ec1a9ff3364ae1780b5e730d6af73271866d00a7bbde1bca00af920ffbd81" exitCode=0 Dec 11 02:56:51 crc kubenswrapper[4824]: I1211 02:56:51.989424 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2bncf" event={"ID":"d343c6a4-1223-4557-ba6c-d92afde1afc9","Type":"ContainerDied","Data":"cc7ec1a9ff3364ae1780b5e730d6af73271866d00a7bbde1bca00af920ffbd81"} Dec 11 02:56:52 crc kubenswrapper[4824]: I1211 02:56:52.115697 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2bncf" Dec 11 02:56:52 crc kubenswrapper[4824]: I1211 02:56:52.233015 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dqrk2\" (UniqueName: \"kubernetes.io/projected/d343c6a4-1223-4557-ba6c-d92afde1afc9-kube-api-access-dqrk2\") pod \"d343c6a4-1223-4557-ba6c-d92afde1afc9\" (UID: \"d343c6a4-1223-4557-ba6c-d92afde1afc9\") " Dec 11 02:56:52 crc kubenswrapper[4824]: I1211 02:56:52.233210 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d343c6a4-1223-4557-ba6c-d92afde1afc9-catalog-content\") pod \"d343c6a4-1223-4557-ba6c-d92afde1afc9\" (UID: \"d343c6a4-1223-4557-ba6c-d92afde1afc9\") " Dec 11 02:56:52 crc kubenswrapper[4824]: I1211 02:56:52.233324 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d343c6a4-1223-4557-ba6c-d92afde1afc9-utilities\") pod \"d343c6a4-1223-4557-ba6c-d92afde1afc9\" (UID: \"d343c6a4-1223-4557-ba6c-d92afde1afc9\") " Dec 11 02:56:52 crc kubenswrapper[4824]: I1211 02:56:52.234341 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d343c6a4-1223-4557-ba6c-d92afde1afc9-utilities" (OuterVolumeSpecName: "utilities") pod "d343c6a4-1223-4557-ba6c-d92afde1afc9" (UID: "d343c6a4-1223-4557-ba6c-d92afde1afc9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:56:52 crc kubenswrapper[4824]: I1211 02:56:52.238074 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d343c6a4-1223-4557-ba6c-d92afde1afc9-kube-api-access-dqrk2" (OuterVolumeSpecName: "kube-api-access-dqrk2") pod "d343c6a4-1223-4557-ba6c-d92afde1afc9" (UID: "d343c6a4-1223-4557-ba6c-d92afde1afc9"). InnerVolumeSpecName "kube-api-access-dqrk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:56:52 crc kubenswrapper[4824]: I1211 02:56:52.266125 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d343c6a4-1223-4557-ba6c-d92afde1afc9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d343c6a4-1223-4557-ba6c-d92afde1afc9" (UID: "d343c6a4-1223-4557-ba6c-d92afde1afc9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:56:52 crc kubenswrapper[4824]: I1211 02:56:52.335750 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d343c6a4-1223-4557-ba6c-d92afde1afc9-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 02:56:52 crc kubenswrapper[4824]: I1211 02:56:52.335790 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dqrk2\" (UniqueName: \"kubernetes.io/projected/d343c6a4-1223-4557-ba6c-d92afde1afc9-kube-api-access-dqrk2\") on node \"crc\" DevicePath \"\"" Dec 11 02:56:52 crc kubenswrapper[4824]: I1211 02:56:52.335803 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d343c6a4-1223-4557-ba6c-d92afde1afc9-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 02:56:53 crc kubenswrapper[4824]: I1211 02:56:53.008790 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2bncf" event={"ID":"d343c6a4-1223-4557-ba6c-d92afde1afc9","Type":"ContainerDied","Data":"580e8747320051a91e27de711cf99c7d9ddf67f9f6362a4f7c27b9ce09369be9"} Dec 11 02:56:53 crc kubenswrapper[4824]: I1211 02:56:53.008845 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2bncf" Dec 11 02:56:53 crc kubenswrapper[4824]: I1211 02:56:53.008892 4824 scope.go:117] "RemoveContainer" containerID="cc7ec1a9ff3364ae1780b5e730d6af73271866d00a7bbde1bca00af920ffbd81" Dec 11 02:56:53 crc kubenswrapper[4824]: I1211 02:56:53.048009 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2bncf"] Dec 11 02:56:53 crc kubenswrapper[4824]: I1211 02:56:53.062415 4824 scope.go:117] "RemoveContainer" containerID="858c897d5d8072e8c1ef62b86c4d6ade2e99d3634f71860fccaa6b7735505422" Dec 11 02:56:53 crc kubenswrapper[4824]: I1211 02:56:53.073492 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-2bncf"] Dec 11 02:56:53 crc kubenswrapper[4824]: I1211 02:56:53.104389 4824 scope.go:117] "RemoveContainer" containerID="41c194011c3db553c2bcdff851affe196be2d21ca38f077812ea6e5bcb333009" Dec 11 02:56:54 crc kubenswrapper[4824]: I1211 02:56:54.659508 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d343c6a4-1223-4557-ba6c-d92afde1afc9" path="/var/lib/kubelet/pods/d343c6a4-1223-4557-ba6c-d92afde1afc9/volumes" Dec 11 02:57:07 crc kubenswrapper[4824]: I1211 02:57:07.251328 4824 patch_prober.go:28] interesting pod/machine-config-daemon-gx6xt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 02:57:07 crc kubenswrapper[4824]: I1211 02:57:07.251914 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 02:57:37 crc kubenswrapper[4824]: I1211 02:57:37.253482 4824 patch_prober.go:28] interesting pod/machine-config-daemon-gx6xt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 02:57:37 crc kubenswrapper[4824]: I1211 02:57:37.254805 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 02:57:37 crc kubenswrapper[4824]: I1211 02:57:37.254890 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" Dec 11 02:57:37 crc kubenswrapper[4824]: I1211 02:57:37.256287 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b4f54810df237c174e9f9c1b3d2b74efaf5ad20b26965b89c1000a0ab8e02ced"} pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 11 02:57:37 crc kubenswrapper[4824]: I1211 02:57:37.256389 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" containerName="machine-config-daemon" containerID="cri-o://b4f54810df237c174e9f9c1b3d2b74efaf5ad20b26965b89c1000a0ab8e02ced" gracePeriod=600 Dec 11 02:57:37 crc kubenswrapper[4824]: I1211 02:57:37.490198 4824 generic.go:334] "Generic (PLEG): container finished" podID="44b156e4-64a4-4d45-aa5e-9b10a862faed" containerID="b4f54810df237c174e9f9c1b3d2b74efaf5ad20b26965b89c1000a0ab8e02ced" exitCode=0 Dec 11 02:57:37 crc kubenswrapper[4824]: I1211 02:57:37.490266 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" event={"ID":"44b156e4-64a4-4d45-aa5e-9b10a862faed","Type":"ContainerDied","Data":"b4f54810df237c174e9f9c1b3d2b74efaf5ad20b26965b89c1000a0ab8e02ced"} Dec 11 02:57:37 crc kubenswrapper[4824]: I1211 02:57:37.490321 4824 scope.go:117] "RemoveContainer" containerID="c19d806d2c030c4af9eb013dc9d405f67612520a0b85a1ecd0356d85a1e2fd14" Dec 11 02:57:38 crc kubenswrapper[4824]: I1211 02:57:38.505102 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" event={"ID":"44b156e4-64a4-4d45-aa5e-9b10a862faed","Type":"ContainerStarted","Data":"8dcf42d380c145f29ce6b988ed66e3ef33605f29467da12e5fe7eac51562e700"} Dec 11 02:58:02 crc kubenswrapper[4824]: I1211 02:58:02.375254 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-wjmvj"] Dec 11 02:58:02 crc kubenswrapper[4824]: E1211 02:58:02.376191 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d343c6a4-1223-4557-ba6c-d92afde1afc9" containerName="extract-content" Dec 11 02:58:02 crc kubenswrapper[4824]: I1211 02:58:02.376203 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d343c6a4-1223-4557-ba6c-d92afde1afc9" containerName="extract-content" Dec 11 02:58:02 crc kubenswrapper[4824]: E1211 02:58:02.376218 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d343c6a4-1223-4557-ba6c-d92afde1afc9" containerName="registry-server" Dec 11 02:58:02 crc kubenswrapper[4824]: I1211 02:58:02.376224 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d343c6a4-1223-4557-ba6c-d92afde1afc9" containerName="registry-server" Dec 11 02:58:02 crc kubenswrapper[4824]: E1211 02:58:02.376241 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d343c6a4-1223-4557-ba6c-d92afde1afc9" containerName="extract-utilities" Dec 11 02:58:02 crc kubenswrapper[4824]: I1211 02:58:02.376247 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d343c6a4-1223-4557-ba6c-d92afde1afc9" containerName="extract-utilities" Dec 11 02:58:02 crc kubenswrapper[4824]: I1211 02:58:02.376438 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="d343c6a4-1223-4557-ba6c-d92afde1afc9" containerName="registry-server" Dec 11 02:58:02 crc kubenswrapper[4824]: I1211 02:58:02.377744 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wjmvj" Dec 11 02:58:02 crc kubenswrapper[4824]: I1211 02:58:02.394051 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wjmvj"] Dec 11 02:58:02 crc kubenswrapper[4824]: I1211 02:58:02.471818 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/351ca498-5ccd-4d80-a99f-3d56f7c5dbf2-catalog-content\") pod \"redhat-operators-wjmvj\" (UID: \"351ca498-5ccd-4d80-a99f-3d56f7c5dbf2\") " pod="openshift-marketplace/redhat-operators-wjmvj" Dec 11 02:58:02 crc kubenswrapper[4824]: I1211 02:58:02.472191 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/351ca498-5ccd-4d80-a99f-3d56f7c5dbf2-utilities\") pod \"redhat-operators-wjmvj\" (UID: \"351ca498-5ccd-4d80-a99f-3d56f7c5dbf2\") " pod="openshift-marketplace/redhat-operators-wjmvj" Dec 11 02:58:02 crc kubenswrapper[4824]: I1211 02:58:02.472260 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9gbv2\" (UniqueName: \"kubernetes.io/projected/351ca498-5ccd-4d80-a99f-3d56f7c5dbf2-kube-api-access-9gbv2\") pod \"redhat-operators-wjmvj\" (UID: \"351ca498-5ccd-4d80-a99f-3d56f7c5dbf2\") " pod="openshift-marketplace/redhat-operators-wjmvj" Dec 11 02:58:02 crc kubenswrapper[4824]: I1211 02:58:02.574109 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/351ca498-5ccd-4d80-a99f-3d56f7c5dbf2-utilities\") pod \"redhat-operators-wjmvj\" (UID: \"351ca498-5ccd-4d80-a99f-3d56f7c5dbf2\") " pod="openshift-marketplace/redhat-operators-wjmvj" Dec 11 02:58:02 crc kubenswrapper[4824]: I1211 02:58:02.574278 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9gbv2\" (UniqueName: \"kubernetes.io/projected/351ca498-5ccd-4d80-a99f-3d56f7c5dbf2-kube-api-access-9gbv2\") pod \"redhat-operators-wjmvj\" (UID: \"351ca498-5ccd-4d80-a99f-3d56f7c5dbf2\") " pod="openshift-marketplace/redhat-operators-wjmvj" Dec 11 02:58:02 crc kubenswrapper[4824]: I1211 02:58:02.574385 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/351ca498-5ccd-4d80-a99f-3d56f7c5dbf2-catalog-content\") pod \"redhat-operators-wjmvj\" (UID: \"351ca498-5ccd-4d80-a99f-3d56f7c5dbf2\") " pod="openshift-marketplace/redhat-operators-wjmvj" Dec 11 02:58:02 crc kubenswrapper[4824]: I1211 02:58:02.574646 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/351ca498-5ccd-4d80-a99f-3d56f7c5dbf2-utilities\") pod \"redhat-operators-wjmvj\" (UID: \"351ca498-5ccd-4d80-a99f-3d56f7c5dbf2\") " pod="openshift-marketplace/redhat-operators-wjmvj" Dec 11 02:58:02 crc kubenswrapper[4824]: I1211 02:58:02.574751 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/351ca498-5ccd-4d80-a99f-3d56f7c5dbf2-catalog-content\") pod \"redhat-operators-wjmvj\" (UID: \"351ca498-5ccd-4d80-a99f-3d56f7c5dbf2\") " pod="openshift-marketplace/redhat-operators-wjmvj" Dec 11 02:58:02 crc kubenswrapper[4824]: I1211 02:58:02.597625 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9gbv2\" (UniqueName: \"kubernetes.io/projected/351ca498-5ccd-4d80-a99f-3d56f7c5dbf2-kube-api-access-9gbv2\") pod \"redhat-operators-wjmvj\" (UID: \"351ca498-5ccd-4d80-a99f-3d56f7c5dbf2\") " pod="openshift-marketplace/redhat-operators-wjmvj" Dec 11 02:58:02 crc kubenswrapper[4824]: I1211 02:58:02.699019 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wjmvj" Dec 11 02:58:03 crc kubenswrapper[4824]: I1211 02:58:03.174131 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wjmvj"] Dec 11 02:58:03 crc kubenswrapper[4824]: I1211 02:58:03.843104 4824 generic.go:334] "Generic (PLEG): container finished" podID="351ca498-5ccd-4d80-a99f-3d56f7c5dbf2" containerID="b974d64d344ae1c505231c82a9c1971b9c64d1aaa2defeb8f9294067b00df493" exitCode=0 Dec 11 02:58:03 crc kubenswrapper[4824]: I1211 02:58:03.843147 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wjmvj" event={"ID":"351ca498-5ccd-4d80-a99f-3d56f7c5dbf2","Type":"ContainerDied","Data":"b974d64d344ae1c505231c82a9c1971b9c64d1aaa2defeb8f9294067b00df493"} Dec 11 02:58:03 crc kubenswrapper[4824]: I1211 02:58:03.843480 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wjmvj" event={"ID":"351ca498-5ccd-4d80-a99f-3d56f7c5dbf2","Type":"ContainerStarted","Data":"391fa6fa852feb82f5582f4920385604f4e929aec3b4821af457423fc4dde7d2"} Dec 11 02:58:13 crc kubenswrapper[4824]: I1211 02:58:13.947719 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wjmvj" event={"ID":"351ca498-5ccd-4d80-a99f-3d56f7c5dbf2","Type":"ContainerStarted","Data":"e9577ae6aec1f8e4d22e6e9a3b491c3ad4c9989263d17e7ec6caf11fa6465426"} Dec 11 02:58:16 crc kubenswrapper[4824]: I1211 02:58:16.975514 4824 generic.go:334] "Generic (PLEG): container finished" podID="351ca498-5ccd-4d80-a99f-3d56f7c5dbf2" containerID="e9577ae6aec1f8e4d22e6e9a3b491c3ad4c9989263d17e7ec6caf11fa6465426" exitCode=0 Dec 11 02:58:16 crc kubenswrapper[4824]: I1211 02:58:16.975576 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wjmvj" event={"ID":"351ca498-5ccd-4d80-a99f-3d56f7c5dbf2","Type":"ContainerDied","Data":"e9577ae6aec1f8e4d22e6e9a3b491c3ad4c9989263d17e7ec6caf11fa6465426"} Dec 11 02:58:17 crc kubenswrapper[4824]: I1211 02:58:17.994300 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wjmvj" event={"ID":"351ca498-5ccd-4d80-a99f-3d56f7c5dbf2","Type":"ContainerStarted","Data":"1437a04f3a9ed181acfda33405d431396c09df6c7841227327da7d0409db5539"} Dec 11 02:58:18 crc kubenswrapper[4824]: I1211 02:58:18.028154 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-wjmvj" podStartSLOduration=2.224172541 podStartE2EDuration="16.028133567s" podCreationTimestamp="2025-12-11 02:58:02 +0000 UTC" firstStartedPulling="2025-12-11 02:58:03.84472264 +0000 UTC m=+3425.533760019" lastFinishedPulling="2025-12-11 02:58:17.648683666 +0000 UTC m=+3439.337721045" observedRunningTime="2025-12-11 02:58:18.022044752 +0000 UTC m=+3439.711082131" watchObservedRunningTime="2025-12-11 02:58:18.028133567 +0000 UTC m=+3439.717171286" Dec 11 02:58:22 crc kubenswrapper[4824]: I1211 02:58:22.699767 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-wjmvj" Dec 11 02:58:22 crc kubenswrapper[4824]: I1211 02:58:22.700426 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-wjmvj" Dec 11 02:58:23 crc kubenswrapper[4824]: I1211 02:58:23.784241 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-wjmvj" podUID="351ca498-5ccd-4d80-a99f-3d56f7c5dbf2" containerName="registry-server" probeResult="failure" output=< Dec 11 02:58:23 crc kubenswrapper[4824]: timeout: failed to connect service ":50051" within 1s Dec 11 02:58:23 crc kubenswrapper[4824]: > Dec 11 02:58:32 crc kubenswrapper[4824]: I1211 02:58:32.756616 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-wjmvj" Dec 11 02:58:32 crc kubenswrapper[4824]: I1211 02:58:32.829234 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-wjmvj" Dec 11 02:58:33 crc kubenswrapper[4824]: I1211 02:58:33.418140 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wjmvj"] Dec 11 02:58:33 crc kubenswrapper[4824]: I1211 02:58:33.579994 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-df294"] Dec 11 02:58:33 crc kubenswrapper[4824]: I1211 02:58:33.580565 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-df294" podUID="582641c9-665f-4d61-9374-d50d139762ab" containerName="registry-server" containerID="cri-o://b49fa7d611d18c7a26fada0e82cea44d92b8ccecf9ca8b3a58290dfcb4d65125" gracePeriod=2 Dec 11 02:58:34 crc kubenswrapper[4824]: I1211 02:58:34.111277 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-df294" Dec 11 02:58:34 crc kubenswrapper[4824]: I1211 02:58:34.164509 4824 generic.go:334] "Generic (PLEG): container finished" podID="582641c9-665f-4d61-9374-d50d139762ab" containerID="b49fa7d611d18c7a26fada0e82cea44d92b8ccecf9ca8b3a58290dfcb4d65125" exitCode=0 Dec 11 02:58:34 crc kubenswrapper[4824]: I1211 02:58:34.164590 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-df294" Dec 11 02:58:34 crc kubenswrapper[4824]: I1211 02:58:34.164640 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-df294" event={"ID":"582641c9-665f-4d61-9374-d50d139762ab","Type":"ContainerDied","Data":"b49fa7d611d18c7a26fada0e82cea44d92b8ccecf9ca8b3a58290dfcb4d65125"} Dec 11 02:58:34 crc kubenswrapper[4824]: I1211 02:58:34.164678 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-df294" event={"ID":"582641c9-665f-4d61-9374-d50d139762ab","Type":"ContainerDied","Data":"d59c1b9a1b23fee8e1fca1b79490ab8f0d72576f8d5021cc0f03a041d36b58d1"} Dec 11 02:58:34 crc kubenswrapper[4824]: I1211 02:58:34.164699 4824 scope.go:117] "RemoveContainer" containerID="b49fa7d611d18c7a26fada0e82cea44d92b8ccecf9ca8b3a58290dfcb4d65125" Dec 11 02:58:34 crc kubenswrapper[4824]: I1211 02:58:34.197576 4824 scope.go:117] "RemoveContainer" containerID="974bc6bccabbf3e088c5ae9fbfd3f06cc72e3d912dcccb4739b8ab20633e3ad1" Dec 11 02:58:34 crc kubenswrapper[4824]: I1211 02:58:34.236686 4824 scope.go:117] "RemoveContainer" containerID="40bfe40ea92f2b6e73afe1db9e6c91dc3ffa2799d2c8e505adecb83a2c6deb57" Dec 11 02:58:34 crc kubenswrapper[4824]: I1211 02:58:34.237675 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/582641c9-665f-4d61-9374-d50d139762ab-utilities\") pod \"582641c9-665f-4d61-9374-d50d139762ab\" (UID: \"582641c9-665f-4d61-9374-d50d139762ab\") " Dec 11 02:58:34 crc kubenswrapper[4824]: I1211 02:58:34.237751 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/582641c9-665f-4d61-9374-d50d139762ab-catalog-content\") pod \"582641c9-665f-4d61-9374-d50d139762ab\" (UID: \"582641c9-665f-4d61-9374-d50d139762ab\") " Dec 11 02:58:34 crc kubenswrapper[4824]: I1211 02:58:34.237771 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vw8xj\" (UniqueName: \"kubernetes.io/projected/582641c9-665f-4d61-9374-d50d139762ab-kube-api-access-vw8xj\") pod \"582641c9-665f-4d61-9374-d50d139762ab\" (UID: \"582641c9-665f-4d61-9374-d50d139762ab\") " Dec 11 02:58:34 crc kubenswrapper[4824]: I1211 02:58:34.238441 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/582641c9-665f-4d61-9374-d50d139762ab-utilities" (OuterVolumeSpecName: "utilities") pod "582641c9-665f-4d61-9374-d50d139762ab" (UID: "582641c9-665f-4d61-9374-d50d139762ab"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:58:34 crc kubenswrapper[4824]: I1211 02:58:34.238769 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/582641c9-665f-4d61-9374-d50d139762ab-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 02:58:34 crc kubenswrapper[4824]: I1211 02:58:34.254363 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/582641c9-665f-4d61-9374-d50d139762ab-kube-api-access-vw8xj" (OuterVolumeSpecName: "kube-api-access-vw8xj") pod "582641c9-665f-4d61-9374-d50d139762ab" (UID: "582641c9-665f-4d61-9374-d50d139762ab"). InnerVolumeSpecName "kube-api-access-vw8xj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 02:58:34 crc kubenswrapper[4824]: I1211 02:58:34.319146 4824 scope.go:117] "RemoveContainer" containerID="b49fa7d611d18c7a26fada0e82cea44d92b8ccecf9ca8b3a58290dfcb4d65125" Dec 11 02:58:34 crc kubenswrapper[4824]: E1211 02:58:34.319597 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b49fa7d611d18c7a26fada0e82cea44d92b8ccecf9ca8b3a58290dfcb4d65125\": container with ID starting with b49fa7d611d18c7a26fada0e82cea44d92b8ccecf9ca8b3a58290dfcb4d65125 not found: ID does not exist" containerID="b49fa7d611d18c7a26fada0e82cea44d92b8ccecf9ca8b3a58290dfcb4d65125" Dec 11 02:58:34 crc kubenswrapper[4824]: I1211 02:58:34.319642 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b49fa7d611d18c7a26fada0e82cea44d92b8ccecf9ca8b3a58290dfcb4d65125"} err="failed to get container status \"b49fa7d611d18c7a26fada0e82cea44d92b8ccecf9ca8b3a58290dfcb4d65125\": rpc error: code = NotFound desc = could not find container \"b49fa7d611d18c7a26fada0e82cea44d92b8ccecf9ca8b3a58290dfcb4d65125\": container with ID starting with b49fa7d611d18c7a26fada0e82cea44d92b8ccecf9ca8b3a58290dfcb4d65125 not found: ID does not exist" Dec 11 02:58:34 crc kubenswrapper[4824]: I1211 02:58:34.319667 4824 scope.go:117] "RemoveContainer" containerID="974bc6bccabbf3e088c5ae9fbfd3f06cc72e3d912dcccb4739b8ab20633e3ad1" Dec 11 02:58:34 crc kubenswrapper[4824]: E1211 02:58:34.320157 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"974bc6bccabbf3e088c5ae9fbfd3f06cc72e3d912dcccb4739b8ab20633e3ad1\": container with ID starting with 974bc6bccabbf3e088c5ae9fbfd3f06cc72e3d912dcccb4739b8ab20633e3ad1 not found: ID does not exist" containerID="974bc6bccabbf3e088c5ae9fbfd3f06cc72e3d912dcccb4739b8ab20633e3ad1" Dec 11 02:58:34 crc kubenswrapper[4824]: I1211 02:58:34.320188 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"974bc6bccabbf3e088c5ae9fbfd3f06cc72e3d912dcccb4739b8ab20633e3ad1"} err="failed to get container status \"974bc6bccabbf3e088c5ae9fbfd3f06cc72e3d912dcccb4739b8ab20633e3ad1\": rpc error: code = NotFound desc = could not find container \"974bc6bccabbf3e088c5ae9fbfd3f06cc72e3d912dcccb4739b8ab20633e3ad1\": container with ID starting with 974bc6bccabbf3e088c5ae9fbfd3f06cc72e3d912dcccb4739b8ab20633e3ad1 not found: ID does not exist" Dec 11 02:58:34 crc kubenswrapper[4824]: I1211 02:58:34.320209 4824 scope.go:117] "RemoveContainer" containerID="40bfe40ea92f2b6e73afe1db9e6c91dc3ffa2799d2c8e505adecb83a2c6deb57" Dec 11 02:58:34 crc kubenswrapper[4824]: E1211 02:58:34.320574 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"40bfe40ea92f2b6e73afe1db9e6c91dc3ffa2799d2c8e505adecb83a2c6deb57\": container with ID starting with 40bfe40ea92f2b6e73afe1db9e6c91dc3ffa2799d2c8e505adecb83a2c6deb57 not found: ID does not exist" containerID="40bfe40ea92f2b6e73afe1db9e6c91dc3ffa2799d2c8e505adecb83a2c6deb57" Dec 11 02:58:34 crc kubenswrapper[4824]: I1211 02:58:34.320631 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"40bfe40ea92f2b6e73afe1db9e6c91dc3ffa2799d2c8e505adecb83a2c6deb57"} err="failed to get container status \"40bfe40ea92f2b6e73afe1db9e6c91dc3ffa2799d2c8e505adecb83a2c6deb57\": rpc error: code = NotFound desc = could not find container \"40bfe40ea92f2b6e73afe1db9e6c91dc3ffa2799d2c8e505adecb83a2c6deb57\": container with ID starting with 40bfe40ea92f2b6e73afe1db9e6c91dc3ffa2799d2c8e505adecb83a2c6deb57 not found: ID does not exist" Dec 11 02:58:34 crc kubenswrapper[4824]: I1211 02:58:34.340302 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vw8xj\" (UniqueName: \"kubernetes.io/projected/582641c9-665f-4d61-9374-d50d139762ab-kube-api-access-vw8xj\") on node \"crc\" DevicePath \"\"" Dec 11 02:58:34 crc kubenswrapper[4824]: I1211 02:58:34.369209 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/582641c9-665f-4d61-9374-d50d139762ab-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "582641c9-665f-4d61-9374-d50d139762ab" (UID: "582641c9-665f-4d61-9374-d50d139762ab"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 02:58:34 crc kubenswrapper[4824]: I1211 02:58:34.441758 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/582641c9-665f-4d61-9374-d50d139762ab-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 02:58:34 crc kubenswrapper[4824]: I1211 02:58:34.502893 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-df294"] Dec 11 02:58:34 crc kubenswrapper[4824]: I1211 02:58:34.509847 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-df294"] Dec 11 02:58:34 crc kubenswrapper[4824]: I1211 02:58:34.643537 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="582641c9-665f-4d61-9374-d50d139762ab" path="/var/lib/kubelet/pods/582641c9-665f-4d61-9374-d50d139762ab/volumes" Dec 11 02:59:37 crc kubenswrapper[4824]: I1211 02:59:37.251239 4824 patch_prober.go:28] interesting pod/machine-config-daemon-gx6xt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 02:59:37 crc kubenswrapper[4824]: I1211 02:59:37.251770 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 03:00:00 crc kubenswrapper[4824]: I1211 03:00:00.170716 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423700-x7r27"] Dec 11 03:00:00 crc kubenswrapper[4824]: E1211 03:00:00.171907 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="582641c9-665f-4d61-9374-d50d139762ab" containerName="extract-utilities" Dec 11 03:00:00 crc kubenswrapper[4824]: I1211 03:00:00.171932 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="582641c9-665f-4d61-9374-d50d139762ab" containerName="extract-utilities" Dec 11 03:00:00 crc kubenswrapper[4824]: E1211 03:00:00.171963 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="582641c9-665f-4d61-9374-d50d139762ab" containerName="extract-content" Dec 11 03:00:00 crc kubenswrapper[4824]: I1211 03:00:00.171974 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="582641c9-665f-4d61-9374-d50d139762ab" containerName="extract-content" Dec 11 03:00:00 crc kubenswrapper[4824]: E1211 03:00:00.172013 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="582641c9-665f-4d61-9374-d50d139762ab" containerName="registry-server" Dec 11 03:00:00 crc kubenswrapper[4824]: I1211 03:00:00.172025 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="582641c9-665f-4d61-9374-d50d139762ab" containerName="registry-server" Dec 11 03:00:00 crc kubenswrapper[4824]: I1211 03:00:00.172397 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="582641c9-665f-4d61-9374-d50d139762ab" containerName="registry-server" Dec 11 03:00:00 crc kubenswrapper[4824]: I1211 03:00:00.173419 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423700-x7r27" Dec 11 03:00:00 crc kubenswrapper[4824]: I1211 03:00:00.178214 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 11 03:00:00 crc kubenswrapper[4824]: I1211 03:00:00.178735 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 11 03:00:00 crc kubenswrapper[4824]: I1211 03:00:00.198328 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423700-x7r27"] Dec 11 03:00:00 crc kubenswrapper[4824]: I1211 03:00:00.273294 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/428cdc13-10d3-48d0-92d6-cb953b0a0215-config-volume\") pod \"collect-profiles-29423700-x7r27\" (UID: \"428cdc13-10d3-48d0-92d6-cb953b0a0215\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423700-x7r27" Dec 11 03:00:00 crc kubenswrapper[4824]: I1211 03:00:00.273616 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nfjlm\" (UniqueName: \"kubernetes.io/projected/428cdc13-10d3-48d0-92d6-cb953b0a0215-kube-api-access-nfjlm\") pod \"collect-profiles-29423700-x7r27\" (UID: \"428cdc13-10d3-48d0-92d6-cb953b0a0215\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423700-x7r27" Dec 11 03:00:00 crc kubenswrapper[4824]: I1211 03:00:00.273894 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/428cdc13-10d3-48d0-92d6-cb953b0a0215-secret-volume\") pod \"collect-profiles-29423700-x7r27\" (UID: \"428cdc13-10d3-48d0-92d6-cb953b0a0215\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423700-x7r27" Dec 11 03:00:00 crc kubenswrapper[4824]: I1211 03:00:00.376225 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/428cdc13-10d3-48d0-92d6-cb953b0a0215-config-volume\") pod \"collect-profiles-29423700-x7r27\" (UID: \"428cdc13-10d3-48d0-92d6-cb953b0a0215\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423700-x7r27" Dec 11 03:00:00 crc kubenswrapper[4824]: I1211 03:00:00.376332 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nfjlm\" (UniqueName: \"kubernetes.io/projected/428cdc13-10d3-48d0-92d6-cb953b0a0215-kube-api-access-nfjlm\") pod \"collect-profiles-29423700-x7r27\" (UID: \"428cdc13-10d3-48d0-92d6-cb953b0a0215\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423700-x7r27" Dec 11 03:00:00 crc kubenswrapper[4824]: I1211 03:00:00.376447 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/428cdc13-10d3-48d0-92d6-cb953b0a0215-secret-volume\") pod \"collect-profiles-29423700-x7r27\" (UID: \"428cdc13-10d3-48d0-92d6-cb953b0a0215\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423700-x7r27" Dec 11 03:00:00 crc kubenswrapper[4824]: I1211 03:00:00.377640 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/428cdc13-10d3-48d0-92d6-cb953b0a0215-config-volume\") pod \"collect-profiles-29423700-x7r27\" (UID: \"428cdc13-10d3-48d0-92d6-cb953b0a0215\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423700-x7r27" Dec 11 03:00:00 crc kubenswrapper[4824]: I1211 03:00:00.391534 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/428cdc13-10d3-48d0-92d6-cb953b0a0215-secret-volume\") pod \"collect-profiles-29423700-x7r27\" (UID: \"428cdc13-10d3-48d0-92d6-cb953b0a0215\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423700-x7r27" Dec 11 03:00:00 crc kubenswrapper[4824]: I1211 03:00:00.409535 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nfjlm\" (UniqueName: \"kubernetes.io/projected/428cdc13-10d3-48d0-92d6-cb953b0a0215-kube-api-access-nfjlm\") pod \"collect-profiles-29423700-x7r27\" (UID: \"428cdc13-10d3-48d0-92d6-cb953b0a0215\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423700-x7r27" Dec 11 03:00:00 crc kubenswrapper[4824]: I1211 03:00:00.531067 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423700-x7r27" Dec 11 03:00:01 crc kubenswrapper[4824]: I1211 03:00:01.082748 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423700-x7r27"] Dec 11 03:00:01 crc kubenswrapper[4824]: I1211 03:00:01.124970 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423700-x7r27" event={"ID":"428cdc13-10d3-48d0-92d6-cb953b0a0215","Type":"ContainerStarted","Data":"42055af341de12d26cda2d9d80d6fce784b18c8b8abb7f70fb0e75e9aab2767d"} Dec 11 03:00:02 crc kubenswrapper[4824]: I1211 03:00:02.138783 4824 generic.go:334] "Generic (PLEG): container finished" podID="428cdc13-10d3-48d0-92d6-cb953b0a0215" containerID="f4a5b4a82f3eef6270ab7b20711812365bd02d39e1d489fdec673588c7e61222" exitCode=0 Dec 11 03:00:02 crc kubenswrapper[4824]: I1211 03:00:02.138935 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423700-x7r27" event={"ID":"428cdc13-10d3-48d0-92d6-cb953b0a0215","Type":"ContainerDied","Data":"f4a5b4a82f3eef6270ab7b20711812365bd02d39e1d489fdec673588c7e61222"} Dec 11 03:00:03 crc kubenswrapper[4824]: I1211 03:00:03.605541 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423700-x7r27" Dec 11 03:00:03 crc kubenswrapper[4824]: I1211 03:00:03.744786 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/428cdc13-10d3-48d0-92d6-cb953b0a0215-config-volume\") pod \"428cdc13-10d3-48d0-92d6-cb953b0a0215\" (UID: \"428cdc13-10d3-48d0-92d6-cb953b0a0215\") " Dec 11 03:00:03 crc kubenswrapper[4824]: I1211 03:00:03.744995 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nfjlm\" (UniqueName: \"kubernetes.io/projected/428cdc13-10d3-48d0-92d6-cb953b0a0215-kube-api-access-nfjlm\") pod \"428cdc13-10d3-48d0-92d6-cb953b0a0215\" (UID: \"428cdc13-10d3-48d0-92d6-cb953b0a0215\") " Dec 11 03:00:03 crc kubenswrapper[4824]: I1211 03:00:03.745049 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/428cdc13-10d3-48d0-92d6-cb953b0a0215-secret-volume\") pod \"428cdc13-10d3-48d0-92d6-cb953b0a0215\" (UID: \"428cdc13-10d3-48d0-92d6-cb953b0a0215\") " Dec 11 03:00:03 crc kubenswrapper[4824]: I1211 03:00:03.745875 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/428cdc13-10d3-48d0-92d6-cb953b0a0215-config-volume" (OuterVolumeSpecName: "config-volume") pod "428cdc13-10d3-48d0-92d6-cb953b0a0215" (UID: "428cdc13-10d3-48d0-92d6-cb953b0a0215"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 03:00:03 crc kubenswrapper[4824]: I1211 03:00:03.750424 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/428cdc13-10d3-48d0-92d6-cb953b0a0215-kube-api-access-nfjlm" (OuterVolumeSpecName: "kube-api-access-nfjlm") pod "428cdc13-10d3-48d0-92d6-cb953b0a0215" (UID: "428cdc13-10d3-48d0-92d6-cb953b0a0215"). InnerVolumeSpecName "kube-api-access-nfjlm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 03:00:03 crc kubenswrapper[4824]: I1211 03:00:03.750471 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/428cdc13-10d3-48d0-92d6-cb953b0a0215-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "428cdc13-10d3-48d0-92d6-cb953b0a0215" (UID: "428cdc13-10d3-48d0-92d6-cb953b0a0215"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 03:00:03 crc kubenswrapper[4824]: I1211 03:00:03.846950 4824 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/428cdc13-10d3-48d0-92d6-cb953b0a0215-config-volume\") on node \"crc\" DevicePath \"\"" Dec 11 03:00:03 crc kubenswrapper[4824]: I1211 03:00:03.846987 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nfjlm\" (UniqueName: \"kubernetes.io/projected/428cdc13-10d3-48d0-92d6-cb953b0a0215-kube-api-access-nfjlm\") on node \"crc\" DevicePath \"\"" Dec 11 03:00:03 crc kubenswrapper[4824]: I1211 03:00:03.847001 4824 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/428cdc13-10d3-48d0-92d6-cb953b0a0215-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 11 03:00:04 crc kubenswrapper[4824]: I1211 03:00:04.164154 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423700-x7r27" event={"ID":"428cdc13-10d3-48d0-92d6-cb953b0a0215","Type":"ContainerDied","Data":"42055af341de12d26cda2d9d80d6fce784b18c8b8abb7f70fb0e75e9aab2767d"} Dec 11 03:00:04 crc kubenswrapper[4824]: I1211 03:00:04.164208 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="42055af341de12d26cda2d9d80d6fce784b18c8b8abb7f70fb0e75e9aab2767d" Dec 11 03:00:04 crc kubenswrapper[4824]: I1211 03:00:04.164647 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423700-x7r27" Dec 11 03:00:04 crc kubenswrapper[4824]: I1211 03:00:04.704802 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423655-kzzdn"] Dec 11 03:00:04 crc kubenswrapper[4824]: I1211 03:00:04.717817 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423655-kzzdn"] Dec 11 03:00:06 crc kubenswrapper[4824]: I1211 03:00:06.649085 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4f5cf3cd-83ba-4408-b280-6bbcdd085ab1" path="/var/lib/kubelet/pods/4f5cf3cd-83ba-4408-b280-6bbcdd085ab1/volumes" Dec 11 03:00:07 crc kubenswrapper[4824]: I1211 03:00:07.251899 4824 patch_prober.go:28] interesting pod/machine-config-daemon-gx6xt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 03:00:07 crc kubenswrapper[4824]: I1211 03:00:07.251983 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 03:00:37 crc kubenswrapper[4824]: I1211 03:00:37.251406 4824 patch_prober.go:28] interesting pod/machine-config-daemon-gx6xt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 03:00:37 crc kubenswrapper[4824]: I1211 03:00:37.252068 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 03:00:37 crc kubenswrapper[4824]: I1211 03:00:37.252163 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" Dec 11 03:00:37 crc kubenswrapper[4824]: I1211 03:00:37.256370 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8dcf42d380c145f29ce6b988ed66e3ef33605f29467da12e5fe7eac51562e700"} pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 11 03:00:37 crc kubenswrapper[4824]: I1211 03:00:37.256484 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" containerName="machine-config-daemon" containerID="cri-o://8dcf42d380c145f29ce6b988ed66e3ef33605f29467da12e5fe7eac51562e700" gracePeriod=600 Dec 11 03:00:37 crc kubenswrapper[4824]: E1211 03:00:37.382185 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 03:00:37 crc kubenswrapper[4824]: I1211 03:00:37.554397 4824 generic.go:334] "Generic (PLEG): container finished" podID="44b156e4-64a4-4d45-aa5e-9b10a862faed" containerID="8dcf42d380c145f29ce6b988ed66e3ef33605f29467da12e5fe7eac51562e700" exitCode=0 Dec 11 03:00:37 crc kubenswrapper[4824]: I1211 03:00:37.554464 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" event={"ID":"44b156e4-64a4-4d45-aa5e-9b10a862faed","Type":"ContainerDied","Data":"8dcf42d380c145f29ce6b988ed66e3ef33605f29467da12e5fe7eac51562e700"} Dec 11 03:00:37 crc kubenswrapper[4824]: I1211 03:00:37.554514 4824 scope.go:117] "RemoveContainer" containerID="b4f54810df237c174e9f9c1b3d2b74efaf5ad20b26965b89c1000a0ab8e02ced" Dec 11 03:00:37 crc kubenswrapper[4824]: I1211 03:00:37.555459 4824 scope.go:117] "RemoveContainer" containerID="8dcf42d380c145f29ce6b988ed66e3ef33605f29467da12e5fe7eac51562e700" Dec 11 03:00:37 crc kubenswrapper[4824]: E1211 03:00:37.555910 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 03:00:50 crc kubenswrapper[4824]: I1211 03:00:50.632781 4824 scope.go:117] "RemoveContainer" containerID="8dcf42d380c145f29ce6b988ed66e3ef33605f29467da12e5fe7eac51562e700" Dec 11 03:00:50 crc kubenswrapper[4824]: E1211 03:00:50.633366 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 03:01:00 crc kubenswrapper[4824]: I1211 03:01:00.169437 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29423701-hlhbx"] Dec 11 03:01:00 crc kubenswrapper[4824]: E1211 03:01:00.171756 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="428cdc13-10d3-48d0-92d6-cb953b0a0215" containerName="collect-profiles" Dec 11 03:01:00 crc kubenswrapper[4824]: I1211 03:01:00.171856 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="428cdc13-10d3-48d0-92d6-cb953b0a0215" containerName="collect-profiles" Dec 11 03:01:00 crc kubenswrapper[4824]: I1211 03:01:00.172183 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="428cdc13-10d3-48d0-92d6-cb953b0a0215" containerName="collect-profiles" Dec 11 03:01:00 crc kubenswrapper[4824]: I1211 03:01:00.173006 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29423701-hlhbx" Dec 11 03:01:00 crc kubenswrapper[4824]: I1211 03:01:00.190013 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29423701-hlhbx"] Dec 11 03:01:00 crc kubenswrapper[4824]: I1211 03:01:00.287393 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-srj6q\" (UniqueName: \"kubernetes.io/projected/a20415c4-3d24-4424-b32f-9e88eb6305e3-kube-api-access-srj6q\") pod \"keystone-cron-29423701-hlhbx\" (UID: \"a20415c4-3d24-4424-b32f-9e88eb6305e3\") " pod="openstack/keystone-cron-29423701-hlhbx" Dec 11 03:01:00 crc kubenswrapper[4824]: I1211 03:01:00.287462 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a20415c4-3d24-4424-b32f-9e88eb6305e3-fernet-keys\") pod \"keystone-cron-29423701-hlhbx\" (UID: \"a20415c4-3d24-4424-b32f-9e88eb6305e3\") " pod="openstack/keystone-cron-29423701-hlhbx" Dec 11 03:01:00 crc kubenswrapper[4824]: I1211 03:01:00.287678 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a20415c4-3d24-4424-b32f-9e88eb6305e3-config-data\") pod \"keystone-cron-29423701-hlhbx\" (UID: \"a20415c4-3d24-4424-b32f-9e88eb6305e3\") " pod="openstack/keystone-cron-29423701-hlhbx" Dec 11 03:01:00 crc kubenswrapper[4824]: I1211 03:01:00.287811 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a20415c4-3d24-4424-b32f-9e88eb6305e3-combined-ca-bundle\") pod \"keystone-cron-29423701-hlhbx\" (UID: \"a20415c4-3d24-4424-b32f-9e88eb6305e3\") " pod="openstack/keystone-cron-29423701-hlhbx" Dec 11 03:01:00 crc kubenswrapper[4824]: I1211 03:01:00.389860 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a20415c4-3d24-4424-b32f-9e88eb6305e3-fernet-keys\") pod \"keystone-cron-29423701-hlhbx\" (UID: \"a20415c4-3d24-4424-b32f-9e88eb6305e3\") " pod="openstack/keystone-cron-29423701-hlhbx" Dec 11 03:01:00 crc kubenswrapper[4824]: I1211 03:01:00.390302 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-srj6q\" (UniqueName: \"kubernetes.io/projected/a20415c4-3d24-4424-b32f-9e88eb6305e3-kube-api-access-srj6q\") pod \"keystone-cron-29423701-hlhbx\" (UID: \"a20415c4-3d24-4424-b32f-9e88eb6305e3\") " pod="openstack/keystone-cron-29423701-hlhbx" Dec 11 03:01:00 crc kubenswrapper[4824]: I1211 03:01:00.390619 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a20415c4-3d24-4424-b32f-9e88eb6305e3-config-data\") pod \"keystone-cron-29423701-hlhbx\" (UID: \"a20415c4-3d24-4424-b32f-9e88eb6305e3\") " pod="openstack/keystone-cron-29423701-hlhbx" Dec 11 03:01:00 crc kubenswrapper[4824]: I1211 03:01:00.390912 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a20415c4-3d24-4424-b32f-9e88eb6305e3-combined-ca-bundle\") pod \"keystone-cron-29423701-hlhbx\" (UID: \"a20415c4-3d24-4424-b32f-9e88eb6305e3\") " pod="openstack/keystone-cron-29423701-hlhbx" Dec 11 03:01:00 crc kubenswrapper[4824]: I1211 03:01:00.399604 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a20415c4-3d24-4424-b32f-9e88eb6305e3-combined-ca-bundle\") pod \"keystone-cron-29423701-hlhbx\" (UID: \"a20415c4-3d24-4424-b32f-9e88eb6305e3\") " pod="openstack/keystone-cron-29423701-hlhbx" Dec 11 03:01:00 crc kubenswrapper[4824]: I1211 03:01:00.401047 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a20415c4-3d24-4424-b32f-9e88eb6305e3-fernet-keys\") pod \"keystone-cron-29423701-hlhbx\" (UID: \"a20415c4-3d24-4424-b32f-9e88eb6305e3\") " pod="openstack/keystone-cron-29423701-hlhbx" Dec 11 03:01:00 crc kubenswrapper[4824]: I1211 03:01:00.401695 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a20415c4-3d24-4424-b32f-9e88eb6305e3-config-data\") pod \"keystone-cron-29423701-hlhbx\" (UID: \"a20415c4-3d24-4424-b32f-9e88eb6305e3\") " pod="openstack/keystone-cron-29423701-hlhbx" Dec 11 03:01:00 crc kubenswrapper[4824]: I1211 03:01:00.414062 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-srj6q\" (UniqueName: \"kubernetes.io/projected/a20415c4-3d24-4424-b32f-9e88eb6305e3-kube-api-access-srj6q\") pod \"keystone-cron-29423701-hlhbx\" (UID: \"a20415c4-3d24-4424-b32f-9e88eb6305e3\") " pod="openstack/keystone-cron-29423701-hlhbx" Dec 11 03:01:00 crc kubenswrapper[4824]: I1211 03:01:00.496356 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29423701-hlhbx" Dec 11 03:01:01 crc kubenswrapper[4824]: I1211 03:01:01.024652 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29423701-hlhbx"] Dec 11 03:01:01 crc kubenswrapper[4824]: W1211 03:01:01.030932 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda20415c4_3d24_4424_b32f_9e88eb6305e3.slice/crio-22a69ea3ec7e74ffa68487a140efb6d1f3102ffdc2d7ad686d49ef678bf15bfd WatchSource:0}: Error finding container 22a69ea3ec7e74ffa68487a140efb6d1f3102ffdc2d7ad686d49ef678bf15bfd: Status 404 returned error can't find the container with id 22a69ea3ec7e74ffa68487a140efb6d1f3102ffdc2d7ad686d49ef678bf15bfd Dec 11 03:01:01 crc kubenswrapper[4824]: I1211 03:01:01.828826 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29423701-hlhbx" event={"ID":"a20415c4-3d24-4424-b32f-9e88eb6305e3","Type":"ContainerStarted","Data":"58c964d52acdb881e8daf8671198e2576f9a5681e0b3002a7ffb709af4c825a9"} Dec 11 03:01:01 crc kubenswrapper[4824]: I1211 03:01:01.829548 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29423701-hlhbx" event={"ID":"a20415c4-3d24-4424-b32f-9e88eb6305e3","Type":"ContainerStarted","Data":"22a69ea3ec7e74ffa68487a140efb6d1f3102ffdc2d7ad686d49ef678bf15bfd"} Dec 11 03:01:01 crc kubenswrapper[4824]: I1211 03:01:01.864381 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29423701-hlhbx" podStartSLOduration=1.864350127 podStartE2EDuration="1.864350127s" podCreationTimestamp="2025-12-11 03:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 03:01:01.853956256 +0000 UTC m=+3603.542993675" watchObservedRunningTime="2025-12-11 03:01:01.864350127 +0000 UTC m=+3603.553387536" Dec 11 03:01:03 crc kubenswrapper[4824]: I1211 03:01:03.859268 4824 generic.go:334] "Generic (PLEG): container finished" podID="a20415c4-3d24-4424-b32f-9e88eb6305e3" containerID="58c964d52acdb881e8daf8671198e2576f9a5681e0b3002a7ffb709af4c825a9" exitCode=0 Dec 11 03:01:03 crc kubenswrapper[4824]: I1211 03:01:03.859519 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29423701-hlhbx" event={"ID":"a20415c4-3d24-4424-b32f-9e88eb6305e3","Type":"ContainerDied","Data":"58c964d52acdb881e8daf8671198e2576f9a5681e0b3002a7ffb709af4c825a9"} Dec 11 03:01:04 crc kubenswrapper[4824]: I1211 03:01:04.633594 4824 scope.go:117] "RemoveContainer" containerID="8dcf42d380c145f29ce6b988ed66e3ef33605f29467da12e5fe7eac51562e700" Dec 11 03:01:04 crc kubenswrapper[4824]: E1211 03:01:04.634363 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 03:01:05 crc kubenswrapper[4824]: I1211 03:01:05.331596 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29423701-hlhbx" Dec 11 03:01:05 crc kubenswrapper[4824]: I1211 03:01:05.437969 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a20415c4-3d24-4424-b32f-9e88eb6305e3-config-data\") pod \"a20415c4-3d24-4424-b32f-9e88eb6305e3\" (UID: \"a20415c4-3d24-4424-b32f-9e88eb6305e3\") " Dec 11 03:01:05 crc kubenswrapper[4824]: I1211 03:01:05.438029 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a20415c4-3d24-4424-b32f-9e88eb6305e3-fernet-keys\") pod \"a20415c4-3d24-4424-b32f-9e88eb6305e3\" (UID: \"a20415c4-3d24-4424-b32f-9e88eb6305e3\") " Dec 11 03:01:05 crc kubenswrapper[4824]: I1211 03:01:05.438262 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-srj6q\" (UniqueName: \"kubernetes.io/projected/a20415c4-3d24-4424-b32f-9e88eb6305e3-kube-api-access-srj6q\") pod \"a20415c4-3d24-4424-b32f-9e88eb6305e3\" (UID: \"a20415c4-3d24-4424-b32f-9e88eb6305e3\") " Dec 11 03:01:05 crc kubenswrapper[4824]: I1211 03:01:05.438306 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a20415c4-3d24-4424-b32f-9e88eb6305e3-combined-ca-bundle\") pod \"a20415c4-3d24-4424-b32f-9e88eb6305e3\" (UID: \"a20415c4-3d24-4424-b32f-9e88eb6305e3\") " Dec 11 03:01:05 crc kubenswrapper[4824]: I1211 03:01:05.443941 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a20415c4-3d24-4424-b32f-9e88eb6305e3-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "a20415c4-3d24-4424-b32f-9e88eb6305e3" (UID: "a20415c4-3d24-4424-b32f-9e88eb6305e3"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 03:01:05 crc kubenswrapper[4824]: I1211 03:01:05.444874 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a20415c4-3d24-4424-b32f-9e88eb6305e3-kube-api-access-srj6q" (OuterVolumeSpecName: "kube-api-access-srj6q") pod "a20415c4-3d24-4424-b32f-9e88eb6305e3" (UID: "a20415c4-3d24-4424-b32f-9e88eb6305e3"). InnerVolumeSpecName "kube-api-access-srj6q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 03:01:05 crc kubenswrapper[4824]: I1211 03:01:05.474388 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a20415c4-3d24-4424-b32f-9e88eb6305e3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a20415c4-3d24-4424-b32f-9e88eb6305e3" (UID: "a20415c4-3d24-4424-b32f-9e88eb6305e3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 03:01:05 crc kubenswrapper[4824]: I1211 03:01:05.500042 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a20415c4-3d24-4424-b32f-9e88eb6305e3-config-data" (OuterVolumeSpecName: "config-data") pod "a20415c4-3d24-4424-b32f-9e88eb6305e3" (UID: "a20415c4-3d24-4424-b32f-9e88eb6305e3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 03:01:05 crc kubenswrapper[4824]: I1211 03:01:05.540154 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a20415c4-3d24-4424-b32f-9e88eb6305e3-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 03:01:05 crc kubenswrapper[4824]: I1211 03:01:05.540184 4824 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a20415c4-3d24-4424-b32f-9e88eb6305e3-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 11 03:01:05 crc kubenswrapper[4824]: I1211 03:01:05.540197 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-srj6q\" (UniqueName: \"kubernetes.io/projected/a20415c4-3d24-4424-b32f-9e88eb6305e3-kube-api-access-srj6q\") on node \"crc\" DevicePath \"\"" Dec 11 03:01:05 crc kubenswrapper[4824]: I1211 03:01:05.540208 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a20415c4-3d24-4424-b32f-9e88eb6305e3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 03:01:05 crc kubenswrapper[4824]: I1211 03:01:05.543303 4824 scope.go:117] "RemoveContainer" containerID="3e921b7d8cabc3f5640f6ce4cff6e9867224ae1ceb66ae93c4747189d2f3e643" Dec 11 03:01:05 crc kubenswrapper[4824]: I1211 03:01:05.886256 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29423701-hlhbx" event={"ID":"a20415c4-3d24-4424-b32f-9e88eb6305e3","Type":"ContainerDied","Data":"22a69ea3ec7e74ffa68487a140efb6d1f3102ffdc2d7ad686d49ef678bf15bfd"} Dec 11 03:01:05 crc kubenswrapper[4824]: I1211 03:01:05.886308 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="22a69ea3ec7e74ffa68487a140efb6d1f3102ffdc2d7ad686d49ef678bf15bfd" Dec 11 03:01:05 crc kubenswrapper[4824]: I1211 03:01:05.886360 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29423701-hlhbx" Dec 11 03:01:17 crc kubenswrapper[4824]: I1211 03:01:17.633073 4824 scope.go:117] "RemoveContainer" containerID="8dcf42d380c145f29ce6b988ed66e3ef33605f29467da12e5fe7eac51562e700" Dec 11 03:01:17 crc kubenswrapper[4824]: E1211 03:01:17.633985 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 03:01:28 crc kubenswrapper[4824]: I1211 03:01:28.643693 4824 scope.go:117] "RemoveContainer" containerID="8dcf42d380c145f29ce6b988ed66e3ef33605f29467da12e5fe7eac51562e700" Dec 11 03:01:28 crc kubenswrapper[4824]: E1211 03:01:28.644748 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 03:01:40 crc kubenswrapper[4824]: I1211 03:01:40.633088 4824 scope.go:117] "RemoveContainer" containerID="8dcf42d380c145f29ce6b988ed66e3ef33605f29467da12e5fe7eac51562e700" Dec 11 03:01:40 crc kubenswrapper[4824]: E1211 03:01:40.634261 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 03:01:54 crc kubenswrapper[4824]: I1211 03:01:54.633320 4824 scope.go:117] "RemoveContainer" containerID="8dcf42d380c145f29ce6b988ed66e3ef33605f29467da12e5fe7eac51562e700" Dec 11 03:01:54 crc kubenswrapper[4824]: E1211 03:01:54.636094 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 03:02:06 crc kubenswrapper[4824]: I1211 03:02:06.633903 4824 scope.go:117] "RemoveContainer" containerID="8dcf42d380c145f29ce6b988ed66e3ef33605f29467da12e5fe7eac51562e700" Dec 11 03:02:06 crc kubenswrapper[4824]: E1211 03:02:06.634911 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 03:02:20 crc kubenswrapper[4824]: I1211 03:02:20.633587 4824 scope.go:117] "RemoveContainer" containerID="8dcf42d380c145f29ce6b988ed66e3ef33605f29467da12e5fe7eac51562e700" Dec 11 03:02:20 crc kubenswrapper[4824]: E1211 03:02:20.639789 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 03:02:30 crc kubenswrapper[4824]: I1211 03:02:30.901315 4824 generic.go:334] "Generic (PLEG): container finished" podID="ab73a048-199c-464e-98ab-2e87063161b6" containerID="47d6cc75b297e87d539ca6fb5a3d045b63a034458b297206ee08a75db9a2f4d2" exitCode=0 Dec 11 03:02:30 crc kubenswrapper[4824]: I1211 03:02:30.901351 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"ab73a048-199c-464e-98ab-2e87063161b6","Type":"ContainerDied","Data":"47d6cc75b297e87d539ca6fb5a3d045b63a034458b297206ee08a75db9a2f4d2"} Dec 11 03:02:32 crc kubenswrapper[4824]: I1211 03:02:32.502459 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 11 03:02:32 crc kubenswrapper[4824]: I1211 03:02:32.652004 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ab73a048-199c-464e-98ab-2e87063161b6-config-data\") pod \"ab73a048-199c-464e-98ab-2e87063161b6\" (UID: \"ab73a048-199c-464e-98ab-2e87063161b6\") " Dec 11 03:02:32 crc kubenswrapper[4824]: I1211 03:02:32.652237 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ab73a048-199c-464e-98ab-2e87063161b6-ssh-key\") pod \"ab73a048-199c-464e-98ab-2e87063161b6\" (UID: \"ab73a048-199c-464e-98ab-2e87063161b6\") " Dec 11 03:02:32 crc kubenswrapper[4824]: I1211 03:02:32.652253 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/ab73a048-199c-464e-98ab-2e87063161b6-ca-certs\") pod \"ab73a048-199c-464e-98ab-2e87063161b6\" (UID: \"ab73a048-199c-464e-98ab-2e87063161b6\") " Dec 11 03:02:32 crc kubenswrapper[4824]: I1211 03:02:32.652319 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n8pl5\" (UniqueName: \"kubernetes.io/projected/ab73a048-199c-464e-98ab-2e87063161b6-kube-api-access-n8pl5\") pod \"ab73a048-199c-464e-98ab-2e87063161b6\" (UID: \"ab73a048-199c-464e-98ab-2e87063161b6\") " Dec 11 03:02:32 crc kubenswrapper[4824]: I1211 03:02:32.652352 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ab73a048-199c-464e-98ab-2e87063161b6-openstack-config\") pod \"ab73a048-199c-464e-98ab-2e87063161b6\" (UID: \"ab73a048-199c-464e-98ab-2e87063161b6\") " Dec 11 03:02:32 crc kubenswrapper[4824]: I1211 03:02:32.652371 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ab73a048-199c-464e-98ab-2e87063161b6-openstack-config-secret\") pod \"ab73a048-199c-464e-98ab-2e87063161b6\" (UID: \"ab73a048-199c-464e-98ab-2e87063161b6\") " Dec 11 03:02:32 crc kubenswrapper[4824]: I1211 03:02:32.652883 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/ab73a048-199c-464e-98ab-2e87063161b6-test-operator-ephemeral-workdir\") pod \"ab73a048-199c-464e-98ab-2e87063161b6\" (UID: \"ab73a048-199c-464e-98ab-2e87063161b6\") " Dec 11 03:02:32 crc kubenswrapper[4824]: I1211 03:02:32.653200 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ab73a048-199c-464e-98ab-2e87063161b6\" (UID: \"ab73a048-199c-464e-98ab-2e87063161b6\") " Dec 11 03:02:32 crc kubenswrapper[4824]: I1211 03:02:32.653246 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/ab73a048-199c-464e-98ab-2e87063161b6-test-operator-ephemeral-temporary\") pod \"ab73a048-199c-464e-98ab-2e87063161b6\" (UID: \"ab73a048-199c-464e-98ab-2e87063161b6\") " Dec 11 03:02:32 crc kubenswrapper[4824]: I1211 03:02:32.653755 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ab73a048-199c-464e-98ab-2e87063161b6-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "ab73a048-199c-464e-98ab-2e87063161b6" (UID: "ab73a048-199c-464e-98ab-2e87063161b6"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 03:02:32 crc kubenswrapper[4824]: I1211 03:02:32.655871 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ab73a048-199c-464e-98ab-2e87063161b6-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "ab73a048-199c-464e-98ab-2e87063161b6" (UID: "ab73a048-199c-464e-98ab-2e87063161b6"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 03:02:32 crc kubenswrapper[4824]: I1211 03:02:32.658175 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "test-operator-logs") pod "ab73a048-199c-464e-98ab-2e87063161b6" (UID: "ab73a048-199c-464e-98ab-2e87063161b6"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 11 03:02:32 crc kubenswrapper[4824]: I1211 03:02:32.658210 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab73a048-199c-464e-98ab-2e87063161b6-kube-api-access-n8pl5" (OuterVolumeSpecName: "kube-api-access-n8pl5") pod "ab73a048-199c-464e-98ab-2e87063161b6" (UID: "ab73a048-199c-464e-98ab-2e87063161b6"). InnerVolumeSpecName "kube-api-access-n8pl5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 03:02:32 crc kubenswrapper[4824]: I1211 03:02:32.667971 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab73a048-199c-464e-98ab-2e87063161b6-config-data" (OuterVolumeSpecName: "config-data") pod "ab73a048-199c-464e-98ab-2e87063161b6" (UID: "ab73a048-199c-464e-98ab-2e87063161b6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 03:02:32 crc kubenswrapper[4824]: I1211 03:02:32.684773 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab73a048-199c-464e-98ab-2e87063161b6-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "ab73a048-199c-464e-98ab-2e87063161b6" (UID: "ab73a048-199c-464e-98ab-2e87063161b6"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 03:02:32 crc kubenswrapper[4824]: I1211 03:02:32.685308 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab73a048-199c-464e-98ab-2e87063161b6-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "ab73a048-199c-464e-98ab-2e87063161b6" (UID: "ab73a048-199c-464e-98ab-2e87063161b6"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 03:02:32 crc kubenswrapper[4824]: I1211 03:02:32.687635 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab73a048-199c-464e-98ab-2e87063161b6-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ab73a048-199c-464e-98ab-2e87063161b6" (UID: "ab73a048-199c-464e-98ab-2e87063161b6"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 03:02:32 crc kubenswrapper[4824]: I1211 03:02:32.707638 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab73a048-199c-464e-98ab-2e87063161b6-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "ab73a048-199c-464e-98ab-2e87063161b6" (UID: "ab73a048-199c-464e-98ab-2e87063161b6"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 03:02:32 crc kubenswrapper[4824]: I1211 03:02:32.754989 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ab73a048-199c-464e-98ab-2e87063161b6-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 03:02:32 crc kubenswrapper[4824]: I1211 03:02:32.755019 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ab73a048-199c-464e-98ab-2e87063161b6-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 11 03:02:32 crc kubenswrapper[4824]: I1211 03:02:32.755027 4824 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/ab73a048-199c-464e-98ab-2e87063161b6-ca-certs\") on node \"crc\" DevicePath \"\"" Dec 11 03:02:32 crc kubenswrapper[4824]: I1211 03:02:32.755035 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n8pl5\" (UniqueName: \"kubernetes.io/projected/ab73a048-199c-464e-98ab-2e87063161b6-kube-api-access-n8pl5\") on node \"crc\" DevicePath \"\"" Dec 11 03:02:32 crc kubenswrapper[4824]: I1211 03:02:32.755045 4824 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ab73a048-199c-464e-98ab-2e87063161b6-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 11 03:02:32 crc kubenswrapper[4824]: I1211 03:02:32.755052 4824 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ab73a048-199c-464e-98ab-2e87063161b6-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 11 03:02:32 crc kubenswrapper[4824]: I1211 03:02:32.755061 4824 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/ab73a048-199c-464e-98ab-2e87063161b6-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Dec 11 03:02:32 crc kubenswrapper[4824]: I1211 03:02:32.755093 4824 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Dec 11 03:02:32 crc kubenswrapper[4824]: I1211 03:02:32.755181 4824 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/ab73a048-199c-464e-98ab-2e87063161b6-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Dec 11 03:02:32 crc kubenswrapper[4824]: I1211 03:02:32.774982 4824 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Dec 11 03:02:32 crc kubenswrapper[4824]: I1211 03:02:32.856327 4824 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Dec 11 03:02:32 crc kubenswrapper[4824]: I1211 03:02:32.927270 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"ab73a048-199c-464e-98ab-2e87063161b6","Type":"ContainerDied","Data":"5f19dd57346c6c5329a1a25379c65c995b28041e84abd07d0f7209e8e0744941"} Dec 11 03:02:32 crc kubenswrapper[4824]: I1211 03:02:32.927304 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5f19dd57346c6c5329a1a25379c65c995b28041e84abd07d0f7209e8e0744941" Dec 11 03:02:32 crc kubenswrapper[4824]: I1211 03:02:32.927357 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 11 03:02:35 crc kubenswrapper[4824]: I1211 03:02:35.633762 4824 scope.go:117] "RemoveContainer" containerID="8dcf42d380c145f29ce6b988ed66e3ef33605f29467da12e5fe7eac51562e700" Dec 11 03:02:35 crc kubenswrapper[4824]: E1211 03:02:35.634889 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 03:02:36 crc kubenswrapper[4824]: I1211 03:02:36.887227 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 11 03:02:36 crc kubenswrapper[4824]: E1211 03:02:36.887676 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab73a048-199c-464e-98ab-2e87063161b6" containerName="tempest-tests-tempest-tests-runner" Dec 11 03:02:36 crc kubenswrapper[4824]: I1211 03:02:36.887692 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab73a048-199c-464e-98ab-2e87063161b6" containerName="tempest-tests-tempest-tests-runner" Dec 11 03:02:36 crc kubenswrapper[4824]: E1211 03:02:36.887731 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a20415c4-3d24-4424-b32f-9e88eb6305e3" containerName="keystone-cron" Dec 11 03:02:36 crc kubenswrapper[4824]: I1211 03:02:36.887740 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="a20415c4-3d24-4424-b32f-9e88eb6305e3" containerName="keystone-cron" Dec 11 03:02:36 crc kubenswrapper[4824]: I1211 03:02:36.887966 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab73a048-199c-464e-98ab-2e87063161b6" containerName="tempest-tests-tempest-tests-runner" Dec 11 03:02:36 crc kubenswrapper[4824]: I1211 03:02:36.887980 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="a20415c4-3d24-4424-b32f-9e88eb6305e3" containerName="keystone-cron" Dec 11 03:02:36 crc kubenswrapper[4824]: I1211 03:02:36.888685 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 11 03:02:36 crc kubenswrapper[4824]: I1211 03:02:36.894183 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-nl4zv" Dec 11 03:02:36 crc kubenswrapper[4824]: I1211 03:02:36.907428 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 11 03:02:37 crc kubenswrapper[4824]: I1211 03:02:37.038809 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pdfvr\" (UniqueName: \"kubernetes.io/projected/1283daba-11e7-4232-81dd-c4316478f297-kube-api-access-pdfvr\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"1283daba-11e7-4232-81dd-c4316478f297\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 11 03:02:37 crc kubenswrapper[4824]: I1211 03:02:37.038893 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"1283daba-11e7-4232-81dd-c4316478f297\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 11 03:02:37 crc kubenswrapper[4824]: I1211 03:02:37.141084 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pdfvr\" (UniqueName: \"kubernetes.io/projected/1283daba-11e7-4232-81dd-c4316478f297-kube-api-access-pdfvr\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"1283daba-11e7-4232-81dd-c4316478f297\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 11 03:02:37 crc kubenswrapper[4824]: I1211 03:02:37.141183 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"1283daba-11e7-4232-81dd-c4316478f297\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 11 03:02:37 crc kubenswrapper[4824]: I1211 03:02:37.141596 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"1283daba-11e7-4232-81dd-c4316478f297\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 11 03:02:37 crc kubenswrapper[4824]: I1211 03:02:37.171084 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pdfvr\" (UniqueName: \"kubernetes.io/projected/1283daba-11e7-4232-81dd-c4316478f297-kube-api-access-pdfvr\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"1283daba-11e7-4232-81dd-c4316478f297\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 11 03:02:37 crc kubenswrapper[4824]: I1211 03:02:37.178259 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"1283daba-11e7-4232-81dd-c4316478f297\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 11 03:02:37 crc kubenswrapper[4824]: I1211 03:02:37.217331 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 11 03:02:37 crc kubenswrapper[4824]: I1211 03:02:37.690006 4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 11 03:02:37 crc kubenswrapper[4824]: I1211 03:02:37.690290 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 11 03:02:37 crc kubenswrapper[4824]: I1211 03:02:37.982840 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"1283daba-11e7-4232-81dd-c4316478f297","Type":"ContainerStarted","Data":"1677b6821acc0646e723ab5e11d5e2d1ed5493887e3a058ecc17b6d25ef7b183"} Dec 11 03:02:39 crc kubenswrapper[4824]: I1211 03:02:38.995271 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"1283daba-11e7-4232-81dd-c4316478f297","Type":"ContainerStarted","Data":"78a9fc8486dbfa4a8a17bbe25e747138677c02e999f3d3dbac9ca6d4c40b8ade"} Dec 11 03:02:39 crc kubenswrapper[4824]: I1211 03:02:39.026860 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=2.053906293 podStartE2EDuration="3.026828165s" podCreationTimestamp="2025-12-11 03:02:36 +0000 UTC" firstStartedPulling="2025-12-11 03:02:37.689541081 +0000 UTC m=+3699.378578460" lastFinishedPulling="2025-12-11 03:02:38.662462953 +0000 UTC m=+3700.351500332" observedRunningTime="2025-12-11 03:02:39.01253564 +0000 UTC m=+3700.701573069" watchObservedRunningTime="2025-12-11 03:02:39.026828165 +0000 UTC m=+3700.715865584" Dec 11 03:02:39 crc kubenswrapper[4824]: I1211 03:02:39.881031 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-4mk7q"] Dec 11 03:02:39 crc kubenswrapper[4824]: I1211 03:02:39.884238 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4mk7q" Dec 11 03:02:39 crc kubenswrapper[4824]: I1211 03:02:39.896722 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4mk7q"] Dec 11 03:02:40 crc kubenswrapper[4824]: I1211 03:02:40.001556 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5lhz2\" (UniqueName: \"kubernetes.io/projected/433def73-c3be-419e-9ad8-743aa1de6472-kube-api-access-5lhz2\") pod \"certified-operators-4mk7q\" (UID: \"433def73-c3be-419e-9ad8-743aa1de6472\") " pod="openshift-marketplace/certified-operators-4mk7q" Dec 11 03:02:40 crc kubenswrapper[4824]: I1211 03:02:40.001936 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/433def73-c3be-419e-9ad8-743aa1de6472-utilities\") pod \"certified-operators-4mk7q\" (UID: \"433def73-c3be-419e-9ad8-743aa1de6472\") " pod="openshift-marketplace/certified-operators-4mk7q" Dec 11 03:02:40 crc kubenswrapper[4824]: I1211 03:02:40.002060 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/433def73-c3be-419e-9ad8-743aa1de6472-catalog-content\") pod \"certified-operators-4mk7q\" (UID: \"433def73-c3be-419e-9ad8-743aa1de6472\") " pod="openshift-marketplace/certified-operators-4mk7q" Dec 11 03:02:40 crc kubenswrapper[4824]: I1211 03:02:40.103733 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/433def73-c3be-419e-9ad8-743aa1de6472-catalog-content\") pod \"certified-operators-4mk7q\" (UID: \"433def73-c3be-419e-9ad8-743aa1de6472\") " pod="openshift-marketplace/certified-operators-4mk7q" Dec 11 03:02:40 crc kubenswrapper[4824]: I1211 03:02:40.103857 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5lhz2\" (UniqueName: \"kubernetes.io/projected/433def73-c3be-419e-9ad8-743aa1de6472-kube-api-access-5lhz2\") pod \"certified-operators-4mk7q\" (UID: \"433def73-c3be-419e-9ad8-743aa1de6472\") " pod="openshift-marketplace/certified-operators-4mk7q" Dec 11 03:02:40 crc kubenswrapper[4824]: I1211 03:02:40.103948 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/433def73-c3be-419e-9ad8-743aa1de6472-utilities\") pod \"certified-operators-4mk7q\" (UID: \"433def73-c3be-419e-9ad8-743aa1de6472\") " pod="openshift-marketplace/certified-operators-4mk7q" Dec 11 03:02:40 crc kubenswrapper[4824]: I1211 03:02:40.104289 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/433def73-c3be-419e-9ad8-743aa1de6472-catalog-content\") pod \"certified-operators-4mk7q\" (UID: \"433def73-c3be-419e-9ad8-743aa1de6472\") " pod="openshift-marketplace/certified-operators-4mk7q" Dec 11 03:02:40 crc kubenswrapper[4824]: I1211 03:02:40.104413 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/433def73-c3be-419e-9ad8-743aa1de6472-utilities\") pod \"certified-operators-4mk7q\" (UID: \"433def73-c3be-419e-9ad8-743aa1de6472\") " pod="openshift-marketplace/certified-operators-4mk7q" Dec 11 03:02:40 crc kubenswrapper[4824]: I1211 03:02:40.124226 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5lhz2\" (UniqueName: \"kubernetes.io/projected/433def73-c3be-419e-9ad8-743aa1de6472-kube-api-access-5lhz2\") pod \"certified-operators-4mk7q\" (UID: \"433def73-c3be-419e-9ad8-743aa1de6472\") " pod="openshift-marketplace/certified-operators-4mk7q" Dec 11 03:02:40 crc kubenswrapper[4824]: I1211 03:02:40.205941 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4mk7q" Dec 11 03:02:40 crc kubenswrapper[4824]: I1211 03:02:40.736229 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4mk7q"] Dec 11 03:02:40 crc kubenswrapper[4824]: W1211 03:02:40.741151 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod433def73_c3be_419e_9ad8_743aa1de6472.slice/crio-f1f71aafe2df7c12eaa3c5d9199c23e92b6cc586743a2fa0f65583efb20aa706 WatchSource:0}: Error finding container f1f71aafe2df7c12eaa3c5d9199c23e92b6cc586743a2fa0f65583efb20aa706: Status 404 returned error can't find the container with id f1f71aafe2df7c12eaa3c5d9199c23e92b6cc586743a2fa0f65583efb20aa706 Dec 11 03:02:41 crc kubenswrapper[4824]: I1211 03:02:41.017549 4824 generic.go:334] "Generic (PLEG): container finished" podID="433def73-c3be-419e-9ad8-743aa1de6472" containerID="a02fb93fee8a09f200929a9d6e4abf149bd6ad80890782514dac6f1375545d1e" exitCode=0 Dec 11 03:02:41 crc kubenswrapper[4824]: I1211 03:02:41.017651 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4mk7q" event={"ID":"433def73-c3be-419e-9ad8-743aa1de6472","Type":"ContainerDied","Data":"a02fb93fee8a09f200929a9d6e4abf149bd6ad80890782514dac6f1375545d1e"} Dec 11 03:02:41 crc kubenswrapper[4824]: I1211 03:02:41.017893 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4mk7q" event={"ID":"433def73-c3be-419e-9ad8-743aa1de6472","Type":"ContainerStarted","Data":"f1f71aafe2df7c12eaa3c5d9199c23e92b6cc586743a2fa0f65583efb20aa706"} Dec 11 03:02:42 crc kubenswrapper[4824]: I1211 03:02:42.030476 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4mk7q" event={"ID":"433def73-c3be-419e-9ad8-743aa1de6472","Type":"ContainerStarted","Data":"43172e97b91c4fed87ffd0378bcb4c0818f2087219c390886565dcbf5aec2c2a"} Dec 11 03:02:43 crc kubenswrapper[4824]: I1211 03:02:43.047868 4824 generic.go:334] "Generic (PLEG): container finished" podID="433def73-c3be-419e-9ad8-743aa1de6472" containerID="43172e97b91c4fed87ffd0378bcb4c0818f2087219c390886565dcbf5aec2c2a" exitCode=0 Dec 11 03:02:43 crc kubenswrapper[4824]: I1211 03:02:43.047910 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4mk7q" event={"ID":"433def73-c3be-419e-9ad8-743aa1de6472","Type":"ContainerDied","Data":"43172e97b91c4fed87ffd0378bcb4c0818f2087219c390886565dcbf5aec2c2a"} Dec 11 03:02:45 crc kubenswrapper[4824]: I1211 03:02:45.073458 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4mk7q" event={"ID":"433def73-c3be-419e-9ad8-743aa1de6472","Type":"ContainerStarted","Data":"a75c7380d7f70ea3e38770e857fc6a323fcbaaa452f2d2cbfd691bd13dc1a7b6"} Dec 11 03:02:45 crc kubenswrapper[4824]: I1211 03:02:45.105532 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-4mk7q" podStartSLOduration=2.4504992039999998 podStartE2EDuration="6.105504503s" podCreationTimestamp="2025-12-11 03:02:39 +0000 UTC" firstStartedPulling="2025-12-11 03:02:41.018910901 +0000 UTC m=+3702.707948280" lastFinishedPulling="2025-12-11 03:02:44.67391619 +0000 UTC m=+3706.362953579" observedRunningTime="2025-12-11 03:02:45.098856134 +0000 UTC m=+3706.787893513" watchObservedRunningTime="2025-12-11 03:02:45.105504503 +0000 UTC m=+3706.794541922" Dec 11 03:02:50 crc kubenswrapper[4824]: I1211 03:02:50.206496 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-4mk7q" Dec 11 03:02:50 crc kubenswrapper[4824]: I1211 03:02:50.207099 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-4mk7q" Dec 11 03:02:50 crc kubenswrapper[4824]: I1211 03:02:50.258956 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-4mk7q" Dec 11 03:02:50 crc kubenswrapper[4824]: I1211 03:02:50.633234 4824 scope.go:117] "RemoveContainer" containerID="8dcf42d380c145f29ce6b988ed66e3ef33605f29467da12e5fe7eac51562e700" Dec 11 03:02:50 crc kubenswrapper[4824]: E1211 03:02:50.634041 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 03:02:51 crc kubenswrapper[4824]: I1211 03:02:51.226028 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-4mk7q" Dec 11 03:02:51 crc kubenswrapper[4824]: I1211 03:02:51.268521 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4mk7q"] Dec 11 03:02:53 crc kubenswrapper[4824]: I1211 03:02:53.203194 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-4mk7q" podUID="433def73-c3be-419e-9ad8-743aa1de6472" containerName="registry-server" containerID="cri-o://a75c7380d7f70ea3e38770e857fc6a323fcbaaa452f2d2cbfd691bd13dc1a7b6" gracePeriod=2 Dec 11 03:02:54 crc kubenswrapper[4824]: I1211 03:02:54.172589 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4mk7q" Dec 11 03:02:54 crc kubenswrapper[4824]: I1211 03:02:54.213558 4824 generic.go:334] "Generic (PLEG): container finished" podID="433def73-c3be-419e-9ad8-743aa1de6472" containerID="a75c7380d7f70ea3e38770e857fc6a323fcbaaa452f2d2cbfd691bd13dc1a7b6" exitCode=0 Dec 11 03:02:54 crc kubenswrapper[4824]: I1211 03:02:54.213594 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4mk7q" event={"ID":"433def73-c3be-419e-9ad8-743aa1de6472","Type":"ContainerDied","Data":"a75c7380d7f70ea3e38770e857fc6a323fcbaaa452f2d2cbfd691bd13dc1a7b6"} Dec 11 03:02:54 crc kubenswrapper[4824]: I1211 03:02:54.213611 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4mk7q" Dec 11 03:02:54 crc kubenswrapper[4824]: I1211 03:02:54.213637 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4mk7q" event={"ID":"433def73-c3be-419e-9ad8-743aa1de6472","Type":"ContainerDied","Data":"f1f71aafe2df7c12eaa3c5d9199c23e92b6cc586743a2fa0f65583efb20aa706"} Dec 11 03:02:54 crc kubenswrapper[4824]: I1211 03:02:54.213653 4824 scope.go:117] "RemoveContainer" containerID="a75c7380d7f70ea3e38770e857fc6a323fcbaaa452f2d2cbfd691bd13dc1a7b6" Dec 11 03:02:54 crc kubenswrapper[4824]: I1211 03:02:54.220837 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5lhz2\" (UniqueName: \"kubernetes.io/projected/433def73-c3be-419e-9ad8-743aa1de6472-kube-api-access-5lhz2\") pod \"433def73-c3be-419e-9ad8-743aa1de6472\" (UID: \"433def73-c3be-419e-9ad8-743aa1de6472\") " Dec 11 03:02:54 crc kubenswrapper[4824]: I1211 03:02:54.220881 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/433def73-c3be-419e-9ad8-743aa1de6472-catalog-content\") pod \"433def73-c3be-419e-9ad8-743aa1de6472\" (UID: \"433def73-c3be-419e-9ad8-743aa1de6472\") " Dec 11 03:02:54 crc kubenswrapper[4824]: I1211 03:02:54.220981 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/433def73-c3be-419e-9ad8-743aa1de6472-utilities\") pod \"433def73-c3be-419e-9ad8-743aa1de6472\" (UID: \"433def73-c3be-419e-9ad8-743aa1de6472\") " Dec 11 03:02:54 crc kubenswrapper[4824]: I1211 03:02:54.222046 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/433def73-c3be-419e-9ad8-743aa1de6472-utilities" (OuterVolumeSpecName: "utilities") pod "433def73-c3be-419e-9ad8-743aa1de6472" (UID: "433def73-c3be-419e-9ad8-743aa1de6472"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 03:02:54 crc kubenswrapper[4824]: I1211 03:02:54.226594 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/433def73-c3be-419e-9ad8-743aa1de6472-kube-api-access-5lhz2" (OuterVolumeSpecName: "kube-api-access-5lhz2") pod "433def73-c3be-419e-9ad8-743aa1de6472" (UID: "433def73-c3be-419e-9ad8-743aa1de6472"). InnerVolumeSpecName "kube-api-access-5lhz2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 03:02:54 crc kubenswrapper[4824]: I1211 03:02:54.241394 4824 scope.go:117] "RemoveContainer" containerID="43172e97b91c4fed87ffd0378bcb4c0818f2087219c390886565dcbf5aec2c2a" Dec 11 03:02:54 crc kubenswrapper[4824]: I1211 03:02:54.275257 4824 scope.go:117] "RemoveContainer" containerID="a02fb93fee8a09f200929a9d6e4abf149bd6ad80890782514dac6f1375545d1e" Dec 11 03:02:54 crc kubenswrapper[4824]: I1211 03:02:54.275480 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/433def73-c3be-419e-9ad8-743aa1de6472-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "433def73-c3be-419e-9ad8-743aa1de6472" (UID: "433def73-c3be-419e-9ad8-743aa1de6472"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 03:02:54 crc kubenswrapper[4824]: I1211 03:02:54.311250 4824 scope.go:117] "RemoveContainer" containerID="a75c7380d7f70ea3e38770e857fc6a323fcbaaa452f2d2cbfd691bd13dc1a7b6" Dec 11 03:02:54 crc kubenswrapper[4824]: E1211 03:02:54.311568 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a75c7380d7f70ea3e38770e857fc6a323fcbaaa452f2d2cbfd691bd13dc1a7b6\": container with ID starting with a75c7380d7f70ea3e38770e857fc6a323fcbaaa452f2d2cbfd691bd13dc1a7b6 not found: ID does not exist" containerID="a75c7380d7f70ea3e38770e857fc6a323fcbaaa452f2d2cbfd691bd13dc1a7b6" Dec 11 03:02:54 crc kubenswrapper[4824]: I1211 03:02:54.311601 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a75c7380d7f70ea3e38770e857fc6a323fcbaaa452f2d2cbfd691bd13dc1a7b6"} err="failed to get container status \"a75c7380d7f70ea3e38770e857fc6a323fcbaaa452f2d2cbfd691bd13dc1a7b6\": rpc error: code = NotFound desc = could not find container \"a75c7380d7f70ea3e38770e857fc6a323fcbaaa452f2d2cbfd691bd13dc1a7b6\": container with ID starting with a75c7380d7f70ea3e38770e857fc6a323fcbaaa452f2d2cbfd691bd13dc1a7b6 not found: ID does not exist" Dec 11 03:02:54 crc kubenswrapper[4824]: I1211 03:02:54.311624 4824 scope.go:117] "RemoveContainer" containerID="43172e97b91c4fed87ffd0378bcb4c0818f2087219c390886565dcbf5aec2c2a" Dec 11 03:02:54 crc kubenswrapper[4824]: E1211 03:02:54.311807 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"43172e97b91c4fed87ffd0378bcb4c0818f2087219c390886565dcbf5aec2c2a\": container with ID starting with 43172e97b91c4fed87ffd0378bcb4c0818f2087219c390886565dcbf5aec2c2a not found: ID does not exist" containerID="43172e97b91c4fed87ffd0378bcb4c0818f2087219c390886565dcbf5aec2c2a" Dec 11 03:02:54 crc kubenswrapper[4824]: I1211 03:02:54.311830 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"43172e97b91c4fed87ffd0378bcb4c0818f2087219c390886565dcbf5aec2c2a"} err="failed to get container status \"43172e97b91c4fed87ffd0378bcb4c0818f2087219c390886565dcbf5aec2c2a\": rpc error: code = NotFound desc = could not find container \"43172e97b91c4fed87ffd0378bcb4c0818f2087219c390886565dcbf5aec2c2a\": container with ID starting with 43172e97b91c4fed87ffd0378bcb4c0818f2087219c390886565dcbf5aec2c2a not found: ID does not exist" Dec 11 03:02:54 crc kubenswrapper[4824]: I1211 03:02:54.311844 4824 scope.go:117] "RemoveContainer" containerID="a02fb93fee8a09f200929a9d6e4abf149bd6ad80890782514dac6f1375545d1e" Dec 11 03:02:54 crc kubenswrapper[4824]: E1211 03:02:54.312295 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a02fb93fee8a09f200929a9d6e4abf149bd6ad80890782514dac6f1375545d1e\": container with ID starting with a02fb93fee8a09f200929a9d6e4abf149bd6ad80890782514dac6f1375545d1e not found: ID does not exist" containerID="a02fb93fee8a09f200929a9d6e4abf149bd6ad80890782514dac6f1375545d1e" Dec 11 03:02:54 crc kubenswrapper[4824]: I1211 03:02:54.312325 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a02fb93fee8a09f200929a9d6e4abf149bd6ad80890782514dac6f1375545d1e"} err="failed to get container status \"a02fb93fee8a09f200929a9d6e4abf149bd6ad80890782514dac6f1375545d1e\": rpc error: code = NotFound desc = could not find container \"a02fb93fee8a09f200929a9d6e4abf149bd6ad80890782514dac6f1375545d1e\": container with ID starting with a02fb93fee8a09f200929a9d6e4abf149bd6ad80890782514dac6f1375545d1e not found: ID does not exist" Dec 11 03:02:54 crc kubenswrapper[4824]: I1211 03:02:54.322849 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/433def73-c3be-419e-9ad8-743aa1de6472-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 03:02:54 crc kubenswrapper[4824]: I1211 03:02:54.322891 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5lhz2\" (UniqueName: \"kubernetes.io/projected/433def73-c3be-419e-9ad8-743aa1de6472-kube-api-access-5lhz2\") on node \"crc\" DevicePath \"\"" Dec 11 03:02:54 crc kubenswrapper[4824]: I1211 03:02:54.322907 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/433def73-c3be-419e-9ad8-743aa1de6472-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 03:02:54 crc kubenswrapper[4824]: I1211 03:02:54.554976 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4mk7q"] Dec 11 03:02:54 crc kubenswrapper[4824]: I1211 03:02:54.563952 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-4mk7q"] Dec 11 03:02:54 crc kubenswrapper[4824]: I1211 03:02:54.651873 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="433def73-c3be-419e-9ad8-743aa1de6472" path="/var/lib/kubelet/pods/433def73-c3be-419e-9ad8-743aa1de6472/volumes" Dec 11 03:03:02 crc kubenswrapper[4824]: I1211 03:03:02.204267 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-v65vx/must-gather-rwnp8"] Dec 11 03:03:02 crc kubenswrapper[4824]: E1211 03:03:02.205252 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="433def73-c3be-419e-9ad8-743aa1de6472" containerName="extract-content" Dec 11 03:03:02 crc kubenswrapper[4824]: I1211 03:03:02.205268 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="433def73-c3be-419e-9ad8-743aa1de6472" containerName="extract-content" Dec 11 03:03:02 crc kubenswrapper[4824]: E1211 03:03:02.205291 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="433def73-c3be-419e-9ad8-743aa1de6472" containerName="extract-utilities" Dec 11 03:03:02 crc kubenswrapper[4824]: I1211 03:03:02.205298 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="433def73-c3be-419e-9ad8-743aa1de6472" containerName="extract-utilities" Dec 11 03:03:02 crc kubenswrapper[4824]: E1211 03:03:02.205314 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="433def73-c3be-419e-9ad8-743aa1de6472" containerName="registry-server" Dec 11 03:03:02 crc kubenswrapper[4824]: I1211 03:03:02.205320 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="433def73-c3be-419e-9ad8-743aa1de6472" containerName="registry-server" Dec 11 03:03:02 crc kubenswrapper[4824]: I1211 03:03:02.205522 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="433def73-c3be-419e-9ad8-743aa1de6472" containerName="registry-server" Dec 11 03:03:02 crc kubenswrapper[4824]: I1211 03:03:02.206637 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-v65vx/must-gather-rwnp8" Dec 11 03:03:02 crc kubenswrapper[4824]: I1211 03:03:02.209635 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-v65vx"/"openshift-service-ca.crt" Dec 11 03:03:02 crc kubenswrapper[4824]: I1211 03:03:02.210334 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-v65vx"/"default-dockercfg-mwfwl" Dec 11 03:03:02 crc kubenswrapper[4824]: I1211 03:03:02.210455 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-v65vx"/"kube-root-ca.crt" Dec 11 03:03:02 crc kubenswrapper[4824]: I1211 03:03:02.212024 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-v65vx/must-gather-rwnp8"] Dec 11 03:03:02 crc kubenswrapper[4824]: I1211 03:03:02.303033 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7gkg\" (UniqueName: \"kubernetes.io/projected/bb60c05d-1bda-4726-a2f2-5967992fd930-kube-api-access-v7gkg\") pod \"must-gather-rwnp8\" (UID: \"bb60c05d-1bda-4726-a2f2-5967992fd930\") " pod="openshift-must-gather-v65vx/must-gather-rwnp8" Dec 11 03:03:02 crc kubenswrapper[4824]: I1211 03:03:02.303310 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/bb60c05d-1bda-4726-a2f2-5967992fd930-must-gather-output\") pod \"must-gather-rwnp8\" (UID: \"bb60c05d-1bda-4726-a2f2-5967992fd930\") " pod="openshift-must-gather-v65vx/must-gather-rwnp8" Dec 11 03:03:02 crc kubenswrapper[4824]: I1211 03:03:02.404684 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/bb60c05d-1bda-4726-a2f2-5967992fd930-must-gather-output\") pod \"must-gather-rwnp8\" (UID: \"bb60c05d-1bda-4726-a2f2-5967992fd930\") " pod="openshift-must-gather-v65vx/must-gather-rwnp8" Dec 11 03:03:02 crc kubenswrapper[4824]: I1211 03:03:02.404740 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7gkg\" (UniqueName: \"kubernetes.io/projected/bb60c05d-1bda-4726-a2f2-5967992fd930-kube-api-access-v7gkg\") pod \"must-gather-rwnp8\" (UID: \"bb60c05d-1bda-4726-a2f2-5967992fd930\") " pod="openshift-must-gather-v65vx/must-gather-rwnp8" Dec 11 03:03:02 crc kubenswrapper[4824]: I1211 03:03:02.405331 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/bb60c05d-1bda-4726-a2f2-5967992fd930-must-gather-output\") pod \"must-gather-rwnp8\" (UID: \"bb60c05d-1bda-4726-a2f2-5967992fd930\") " pod="openshift-must-gather-v65vx/must-gather-rwnp8" Dec 11 03:03:02 crc kubenswrapper[4824]: I1211 03:03:02.429007 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7gkg\" (UniqueName: \"kubernetes.io/projected/bb60c05d-1bda-4726-a2f2-5967992fd930-kube-api-access-v7gkg\") pod \"must-gather-rwnp8\" (UID: \"bb60c05d-1bda-4726-a2f2-5967992fd930\") " pod="openshift-must-gather-v65vx/must-gather-rwnp8" Dec 11 03:03:02 crc kubenswrapper[4824]: I1211 03:03:02.523502 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-v65vx/must-gather-rwnp8" Dec 11 03:03:02 crc kubenswrapper[4824]: I1211 03:03:02.983696 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-v65vx/must-gather-rwnp8"] Dec 11 03:03:03 crc kubenswrapper[4824]: I1211 03:03:03.319350 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-v65vx/must-gather-rwnp8" event={"ID":"bb60c05d-1bda-4726-a2f2-5967992fd930","Type":"ContainerStarted","Data":"21ddf97993abb650fa7f33d2331332cd9845cb2931fb66c24ecfc03d9faa07cc"} Dec 11 03:03:05 crc kubenswrapper[4824]: I1211 03:03:05.632635 4824 scope.go:117] "RemoveContainer" containerID="8dcf42d380c145f29ce6b988ed66e3ef33605f29467da12e5fe7eac51562e700" Dec 11 03:03:05 crc kubenswrapper[4824]: E1211 03:03:05.633636 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 03:03:12 crc kubenswrapper[4824]: I1211 03:03:12.414735 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-v65vx/must-gather-rwnp8" event={"ID":"bb60c05d-1bda-4726-a2f2-5967992fd930","Type":"ContainerStarted","Data":"b60d2ff1c3092c6e50b7344313070e8e1d93cdc8cb0b8d3024dc8cc48b2d7f15"} Dec 11 03:03:12 crc kubenswrapper[4824]: I1211 03:03:12.415628 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-v65vx/must-gather-rwnp8" event={"ID":"bb60c05d-1bda-4726-a2f2-5967992fd930","Type":"ContainerStarted","Data":"899e0d503155df37fa14e0dd1b38b28b8f72e830e29c0e83ba0b27f26c27dc3c"} Dec 11 03:03:12 crc kubenswrapper[4824]: I1211 03:03:12.453184 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-v65vx/must-gather-rwnp8" podStartSLOduration=2.0763553359999998 podStartE2EDuration="10.453156093s" podCreationTimestamp="2025-12-11 03:03:02 +0000 UTC" firstStartedPulling="2025-12-11 03:03:02.997014039 +0000 UTC m=+3724.686051418" lastFinishedPulling="2025-12-11 03:03:11.373814796 +0000 UTC m=+3733.062852175" observedRunningTime="2025-12-11 03:03:12.436041152 +0000 UTC m=+3734.125078541" watchObservedRunningTime="2025-12-11 03:03:12.453156093 +0000 UTC m=+3734.142193512" Dec 11 03:03:15 crc kubenswrapper[4824]: I1211 03:03:15.454094 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-v65vx/crc-debug-n6m7v"] Dec 11 03:03:15 crc kubenswrapper[4824]: I1211 03:03:15.457365 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-v65vx/crc-debug-n6m7v" Dec 11 03:03:15 crc kubenswrapper[4824]: I1211 03:03:15.583237 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3c336928-35ca-4502-92bc-3e9cf7ca7898-host\") pod \"crc-debug-n6m7v\" (UID: \"3c336928-35ca-4502-92bc-3e9cf7ca7898\") " pod="openshift-must-gather-v65vx/crc-debug-n6m7v" Dec 11 03:03:15 crc kubenswrapper[4824]: I1211 03:03:15.583329 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qlpg7\" (UniqueName: \"kubernetes.io/projected/3c336928-35ca-4502-92bc-3e9cf7ca7898-kube-api-access-qlpg7\") pod \"crc-debug-n6m7v\" (UID: \"3c336928-35ca-4502-92bc-3e9cf7ca7898\") " pod="openshift-must-gather-v65vx/crc-debug-n6m7v" Dec 11 03:03:15 crc kubenswrapper[4824]: I1211 03:03:15.684945 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3c336928-35ca-4502-92bc-3e9cf7ca7898-host\") pod \"crc-debug-n6m7v\" (UID: \"3c336928-35ca-4502-92bc-3e9cf7ca7898\") " pod="openshift-must-gather-v65vx/crc-debug-n6m7v" Dec 11 03:03:15 crc kubenswrapper[4824]: I1211 03:03:15.685024 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qlpg7\" (UniqueName: \"kubernetes.io/projected/3c336928-35ca-4502-92bc-3e9cf7ca7898-kube-api-access-qlpg7\") pod \"crc-debug-n6m7v\" (UID: \"3c336928-35ca-4502-92bc-3e9cf7ca7898\") " pod="openshift-must-gather-v65vx/crc-debug-n6m7v" Dec 11 03:03:15 crc kubenswrapper[4824]: I1211 03:03:15.685062 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3c336928-35ca-4502-92bc-3e9cf7ca7898-host\") pod \"crc-debug-n6m7v\" (UID: \"3c336928-35ca-4502-92bc-3e9cf7ca7898\") " pod="openshift-must-gather-v65vx/crc-debug-n6m7v" Dec 11 03:03:15 crc kubenswrapper[4824]: I1211 03:03:15.709037 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qlpg7\" (UniqueName: \"kubernetes.io/projected/3c336928-35ca-4502-92bc-3e9cf7ca7898-kube-api-access-qlpg7\") pod \"crc-debug-n6m7v\" (UID: \"3c336928-35ca-4502-92bc-3e9cf7ca7898\") " pod="openshift-must-gather-v65vx/crc-debug-n6m7v" Dec 11 03:03:15 crc kubenswrapper[4824]: I1211 03:03:15.785865 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-v65vx/crc-debug-n6m7v" Dec 11 03:03:15 crc kubenswrapper[4824]: W1211 03:03:15.842062 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3c336928_35ca_4502_92bc_3e9cf7ca7898.slice/crio-d8cc8202b8dbf764542f4126fd87e389d62d1d342d24a1b8737d616dd3edf76d WatchSource:0}: Error finding container d8cc8202b8dbf764542f4126fd87e389d62d1d342d24a1b8737d616dd3edf76d: Status 404 returned error can't find the container with id d8cc8202b8dbf764542f4126fd87e389d62d1d342d24a1b8737d616dd3edf76d Dec 11 03:03:16 crc kubenswrapper[4824]: I1211 03:03:16.459929 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-v65vx/crc-debug-n6m7v" event={"ID":"3c336928-35ca-4502-92bc-3e9cf7ca7898","Type":"ContainerStarted","Data":"d8cc8202b8dbf764542f4126fd87e389d62d1d342d24a1b8737d616dd3edf76d"} Dec 11 03:03:16 crc kubenswrapper[4824]: I1211 03:03:16.637369 4824 scope.go:117] "RemoveContainer" containerID="8dcf42d380c145f29ce6b988ed66e3ef33605f29467da12e5fe7eac51562e700" Dec 11 03:03:16 crc kubenswrapper[4824]: E1211 03:03:16.637595 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 03:03:27 crc kubenswrapper[4824]: I1211 03:03:27.568675 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-v65vx/crc-debug-n6m7v" event={"ID":"3c336928-35ca-4502-92bc-3e9cf7ca7898","Type":"ContainerStarted","Data":"63480fbd91893611b434d2373e3ceba76ce71e39d9a8eb73082c30bbd5348d9e"} Dec 11 03:03:27 crc kubenswrapper[4824]: I1211 03:03:27.585188 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-v65vx/crc-debug-n6m7v" podStartSLOduration=1.2427067649999999 podStartE2EDuration="12.585163896s" podCreationTimestamp="2025-12-11 03:03:15 +0000 UTC" firstStartedPulling="2025-12-11 03:03:15.844601314 +0000 UTC m=+3737.533638693" lastFinishedPulling="2025-12-11 03:03:27.187058445 +0000 UTC m=+3748.876095824" observedRunningTime="2025-12-11 03:03:27.581474417 +0000 UTC m=+3749.270511806" watchObservedRunningTime="2025-12-11 03:03:27.585163896 +0000 UTC m=+3749.274201285" Dec 11 03:03:30 crc kubenswrapper[4824]: I1211 03:03:30.633962 4824 scope.go:117] "RemoveContainer" containerID="8dcf42d380c145f29ce6b988ed66e3ef33605f29467da12e5fe7eac51562e700" Dec 11 03:03:30 crc kubenswrapper[4824]: E1211 03:03:30.634748 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 03:03:44 crc kubenswrapper[4824]: I1211 03:03:44.633482 4824 scope.go:117] "RemoveContainer" containerID="8dcf42d380c145f29ce6b988ed66e3ef33605f29467da12e5fe7eac51562e700" Dec 11 03:03:44 crc kubenswrapper[4824]: E1211 03:03:44.634192 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 03:03:55 crc kubenswrapper[4824]: I1211 03:03:55.635521 4824 scope.go:117] "RemoveContainer" containerID="8dcf42d380c145f29ce6b988ed66e3ef33605f29467da12e5fe7eac51562e700" Dec 11 03:03:55 crc kubenswrapper[4824]: E1211 03:03:55.637980 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 03:04:06 crc kubenswrapper[4824]: I1211 03:04:06.633487 4824 scope.go:117] "RemoveContainer" containerID="8dcf42d380c145f29ce6b988ed66e3ef33605f29467da12e5fe7eac51562e700" Dec 11 03:04:06 crc kubenswrapper[4824]: E1211 03:04:06.638140 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 03:04:17 crc kubenswrapper[4824]: I1211 03:04:17.033283 4824 generic.go:334] "Generic (PLEG): container finished" podID="3c336928-35ca-4502-92bc-3e9cf7ca7898" containerID="63480fbd91893611b434d2373e3ceba76ce71e39d9a8eb73082c30bbd5348d9e" exitCode=0 Dec 11 03:04:17 crc kubenswrapper[4824]: I1211 03:04:17.033405 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-v65vx/crc-debug-n6m7v" event={"ID":"3c336928-35ca-4502-92bc-3e9cf7ca7898","Type":"ContainerDied","Data":"63480fbd91893611b434d2373e3ceba76ce71e39d9a8eb73082c30bbd5348d9e"} Dec 11 03:04:17 crc kubenswrapper[4824]: I1211 03:04:17.633282 4824 scope.go:117] "RemoveContainer" containerID="8dcf42d380c145f29ce6b988ed66e3ef33605f29467da12e5fe7eac51562e700" Dec 11 03:04:17 crc kubenswrapper[4824]: E1211 03:04:17.634025 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 03:04:18 crc kubenswrapper[4824]: I1211 03:04:18.181041 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-v65vx/crc-debug-n6m7v" Dec 11 03:04:18 crc kubenswrapper[4824]: I1211 03:04:18.195066 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3c336928-35ca-4502-92bc-3e9cf7ca7898-host\") pod \"3c336928-35ca-4502-92bc-3e9cf7ca7898\" (UID: \"3c336928-35ca-4502-92bc-3e9cf7ca7898\") " Dec 11 03:04:18 crc kubenswrapper[4824]: I1211 03:04:18.195197 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qlpg7\" (UniqueName: \"kubernetes.io/projected/3c336928-35ca-4502-92bc-3e9cf7ca7898-kube-api-access-qlpg7\") pod \"3c336928-35ca-4502-92bc-3e9cf7ca7898\" (UID: \"3c336928-35ca-4502-92bc-3e9cf7ca7898\") " Dec 11 03:04:18 crc kubenswrapper[4824]: I1211 03:04:18.195492 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3c336928-35ca-4502-92bc-3e9cf7ca7898-host" (OuterVolumeSpecName: "host") pod "3c336928-35ca-4502-92bc-3e9cf7ca7898" (UID: "3c336928-35ca-4502-92bc-3e9cf7ca7898"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 03:04:18 crc kubenswrapper[4824]: I1211 03:04:18.195803 4824 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3c336928-35ca-4502-92bc-3e9cf7ca7898-host\") on node \"crc\" DevicePath \"\"" Dec 11 03:04:18 crc kubenswrapper[4824]: I1211 03:04:18.201874 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c336928-35ca-4502-92bc-3e9cf7ca7898-kube-api-access-qlpg7" (OuterVolumeSpecName: "kube-api-access-qlpg7") pod "3c336928-35ca-4502-92bc-3e9cf7ca7898" (UID: "3c336928-35ca-4502-92bc-3e9cf7ca7898"). InnerVolumeSpecName "kube-api-access-qlpg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 03:04:18 crc kubenswrapper[4824]: I1211 03:04:18.220422 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-v65vx/crc-debug-n6m7v"] Dec 11 03:04:18 crc kubenswrapper[4824]: I1211 03:04:18.229304 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-v65vx/crc-debug-n6m7v"] Dec 11 03:04:18 crc kubenswrapper[4824]: I1211 03:04:18.299203 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qlpg7\" (UniqueName: \"kubernetes.io/projected/3c336928-35ca-4502-92bc-3e9cf7ca7898-kube-api-access-qlpg7\") on node \"crc\" DevicePath \"\"" Dec 11 03:04:18 crc kubenswrapper[4824]: I1211 03:04:18.652082 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c336928-35ca-4502-92bc-3e9cf7ca7898" path="/var/lib/kubelet/pods/3c336928-35ca-4502-92bc-3e9cf7ca7898/volumes" Dec 11 03:04:19 crc kubenswrapper[4824]: I1211 03:04:19.058756 4824 scope.go:117] "RemoveContainer" containerID="63480fbd91893611b434d2373e3ceba76ce71e39d9a8eb73082c30bbd5348d9e" Dec 11 03:04:19 crc kubenswrapper[4824]: I1211 03:04:19.058953 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-v65vx/crc-debug-n6m7v" Dec 11 03:04:19 crc kubenswrapper[4824]: I1211 03:04:19.521710 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-v65vx/crc-debug-xznvz"] Dec 11 03:04:19 crc kubenswrapper[4824]: E1211 03:04:19.523248 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c336928-35ca-4502-92bc-3e9cf7ca7898" containerName="container-00" Dec 11 03:04:19 crc kubenswrapper[4824]: I1211 03:04:19.523275 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c336928-35ca-4502-92bc-3e9cf7ca7898" containerName="container-00" Dec 11 03:04:19 crc kubenswrapper[4824]: I1211 03:04:19.523602 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c336928-35ca-4502-92bc-3e9cf7ca7898" containerName="container-00" Dec 11 03:04:19 crc kubenswrapper[4824]: I1211 03:04:19.524403 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-v65vx/crc-debug-xznvz" Dec 11 03:04:19 crc kubenswrapper[4824]: I1211 03:04:19.530949 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3007f31b-deae-4518-982a-5591f4958ffb-host\") pod \"crc-debug-xznvz\" (UID: \"3007f31b-deae-4518-982a-5591f4958ffb\") " pod="openshift-must-gather-v65vx/crc-debug-xznvz" Dec 11 03:04:19 crc kubenswrapper[4824]: I1211 03:04:19.531026 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vd8jt\" (UniqueName: \"kubernetes.io/projected/3007f31b-deae-4518-982a-5591f4958ffb-kube-api-access-vd8jt\") pod \"crc-debug-xznvz\" (UID: \"3007f31b-deae-4518-982a-5591f4958ffb\") " pod="openshift-must-gather-v65vx/crc-debug-xznvz" Dec 11 03:04:19 crc kubenswrapper[4824]: I1211 03:04:19.633175 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3007f31b-deae-4518-982a-5591f4958ffb-host\") pod \"crc-debug-xznvz\" (UID: \"3007f31b-deae-4518-982a-5591f4958ffb\") " pod="openshift-must-gather-v65vx/crc-debug-xznvz" Dec 11 03:04:19 crc kubenswrapper[4824]: I1211 03:04:19.633292 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3007f31b-deae-4518-982a-5591f4958ffb-host\") pod \"crc-debug-xznvz\" (UID: \"3007f31b-deae-4518-982a-5591f4958ffb\") " pod="openshift-must-gather-v65vx/crc-debug-xznvz" Dec 11 03:04:19 crc kubenswrapper[4824]: I1211 03:04:19.633303 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vd8jt\" (UniqueName: \"kubernetes.io/projected/3007f31b-deae-4518-982a-5591f4958ffb-kube-api-access-vd8jt\") pod \"crc-debug-xznvz\" (UID: \"3007f31b-deae-4518-982a-5591f4958ffb\") " pod="openshift-must-gather-v65vx/crc-debug-xznvz" Dec 11 03:04:19 crc kubenswrapper[4824]: I1211 03:04:19.663989 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vd8jt\" (UniqueName: \"kubernetes.io/projected/3007f31b-deae-4518-982a-5591f4958ffb-kube-api-access-vd8jt\") pod \"crc-debug-xznvz\" (UID: \"3007f31b-deae-4518-982a-5591f4958ffb\") " pod="openshift-must-gather-v65vx/crc-debug-xznvz" Dec 11 03:04:19 crc kubenswrapper[4824]: I1211 03:04:19.865654 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-v65vx/crc-debug-xznvz" Dec 11 03:04:20 crc kubenswrapper[4824]: I1211 03:04:20.081840 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-v65vx/crc-debug-xznvz" event={"ID":"3007f31b-deae-4518-982a-5591f4958ffb","Type":"ContainerStarted","Data":"b4492461628c1e2adca3038971812d3c35706bbd55db0ac31c737664032ab964"} Dec 11 03:04:21 crc kubenswrapper[4824]: I1211 03:04:21.093512 4824 generic.go:334] "Generic (PLEG): container finished" podID="3007f31b-deae-4518-982a-5591f4958ffb" containerID="a5c479e5a888fc5560e6009a660b949b6dd5b4649e97c6e54d4e6a5cc9f51328" exitCode=0 Dec 11 03:04:21 crc kubenswrapper[4824]: I1211 03:04:21.093584 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-v65vx/crc-debug-xznvz" event={"ID":"3007f31b-deae-4518-982a-5591f4958ffb","Type":"ContainerDied","Data":"a5c479e5a888fc5560e6009a660b949b6dd5b4649e97c6e54d4e6a5cc9f51328"} Dec 11 03:04:21 crc kubenswrapper[4824]: I1211 03:04:21.633538 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-v65vx/crc-debug-xznvz"] Dec 11 03:04:21 crc kubenswrapper[4824]: I1211 03:04:21.640629 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-v65vx/crc-debug-xznvz"] Dec 11 03:04:22 crc kubenswrapper[4824]: I1211 03:04:22.219684 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-v65vx/crc-debug-xznvz" Dec 11 03:04:22 crc kubenswrapper[4824]: I1211 03:04:22.291041 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vd8jt\" (UniqueName: \"kubernetes.io/projected/3007f31b-deae-4518-982a-5591f4958ffb-kube-api-access-vd8jt\") pod \"3007f31b-deae-4518-982a-5591f4958ffb\" (UID: \"3007f31b-deae-4518-982a-5591f4958ffb\") " Dec 11 03:04:22 crc kubenswrapper[4824]: I1211 03:04:22.291084 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3007f31b-deae-4518-982a-5591f4958ffb-host\") pod \"3007f31b-deae-4518-982a-5591f4958ffb\" (UID: \"3007f31b-deae-4518-982a-5591f4958ffb\") " Dec 11 03:04:22 crc kubenswrapper[4824]: I1211 03:04:22.291594 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3007f31b-deae-4518-982a-5591f4958ffb-host" (OuterVolumeSpecName: "host") pod "3007f31b-deae-4518-982a-5591f4958ffb" (UID: "3007f31b-deae-4518-982a-5591f4958ffb"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 03:04:22 crc kubenswrapper[4824]: I1211 03:04:22.303613 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3007f31b-deae-4518-982a-5591f4958ffb-kube-api-access-vd8jt" (OuterVolumeSpecName: "kube-api-access-vd8jt") pod "3007f31b-deae-4518-982a-5591f4958ffb" (UID: "3007f31b-deae-4518-982a-5591f4958ffb"). InnerVolumeSpecName "kube-api-access-vd8jt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 03:04:22 crc kubenswrapper[4824]: I1211 03:04:22.392519 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vd8jt\" (UniqueName: \"kubernetes.io/projected/3007f31b-deae-4518-982a-5591f4958ffb-kube-api-access-vd8jt\") on node \"crc\" DevicePath \"\"" Dec 11 03:04:22 crc kubenswrapper[4824]: I1211 03:04:22.392550 4824 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3007f31b-deae-4518-982a-5591f4958ffb-host\") on node \"crc\" DevicePath \"\"" Dec 11 03:04:22 crc kubenswrapper[4824]: I1211 03:04:22.643991 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3007f31b-deae-4518-982a-5591f4958ffb" path="/var/lib/kubelet/pods/3007f31b-deae-4518-982a-5591f4958ffb/volumes" Dec 11 03:04:22 crc kubenswrapper[4824]: I1211 03:04:22.801395 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-v65vx/crc-debug-qpfxr"] Dec 11 03:04:22 crc kubenswrapper[4824]: E1211 03:04:22.801732 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3007f31b-deae-4518-982a-5591f4958ffb" containerName="container-00" Dec 11 03:04:22 crc kubenswrapper[4824]: I1211 03:04:22.801752 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="3007f31b-deae-4518-982a-5591f4958ffb" containerName="container-00" Dec 11 03:04:22 crc kubenswrapper[4824]: I1211 03:04:22.801956 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="3007f31b-deae-4518-982a-5591f4958ffb" containerName="container-00" Dec 11 03:04:22 crc kubenswrapper[4824]: I1211 03:04:22.802589 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-v65vx/crc-debug-qpfxr" Dec 11 03:04:22 crc kubenswrapper[4824]: I1211 03:04:22.900204 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bfc6dc48-ab52-4b16-83a4-798b7d75c41c-host\") pod \"crc-debug-qpfxr\" (UID: \"bfc6dc48-ab52-4b16-83a4-798b7d75c41c\") " pod="openshift-must-gather-v65vx/crc-debug-qpfxr" Dec 11 03:04:22 crc kubenswrapper[4824]: I1211 03:04:22.900535 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hh58f\" (UniqueName: \"kubernetes.io/projected/bfc6dc48-ab52-4b16-83a4-798b7d75c41c-kube-api-access-hh58f\") pod \"crc-debug-qpfxr\" (UID: \"bfc6dc48-ab52-4b16-83a4-798b7d75c41c\") " pod="openshift-must-gather-v65vx/crc-debug-qpfxr" Dec 11 03:04:23 crc kubenswrapper[4824]: I1211 03:04:23.001285 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bfc6dc48-ab52-4b16-83a4-798b7d75c41c-host\") pod \"crc-debug-qpfxr\" (UID: \"bfc6dc48-ab52-4b16-83a4-798b7d75c41c\") " pod="openshift-must-gather-v65vx/crc-debug-qpfxr" Dec 11 03:04:23 crc kubenswrapper[4824]: I1211 03:04:23.001355 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hh58f\" (UniqueName: \"kubernetes.io/projected/bfc6dc48-ab52-4b16-83a4-798b7d75c41c-kube-api-access-hh58f\") pod \"crc-debug-qpfxr\" (UID: \"bfc6dc48-ab52-4b16-83a4-798b7d75c41c\") " pod="openshift-must-gather-v65vx/crc-debug-qpfxr" Dec 11 03:04:23 crc kubenswrapper[4824]: I1211 03:04:23.001478 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bfc6dc48-ab52-4b16-83a4-798b7d75c41c-host\") pod \"crc-debug-qpfxr\" (UID: \"bfc6dc48-ab52-4b16-83a4-798b7d75c41c\") " pod="openshift-must-gather-v65vx/crc-debug-qpfxr" Dec 11 03:04:23 crc kubenswrapper[4824]: I1211 03:04:23.040027 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hh58f\" (UniqueName: \"kubernetes.io/projected/bfc6dc48-ab52-4b16-83a4-798b7d75c41c-kube-api-access-hh58f\") pod \"crc-debug-qpfxr\" (UID: \"bfc6dc48-ab52-4b16-83a4-798b7d75c41c\") " pod="openshift-must-gather-v65vx/crc-debug-qpfxr" Dec 11 03:04:23 crc kubenswrapper[4824]: I1211 03:04:23.120256 4824 scope.go:117] "RemoveContainer" containerID="a5c479e5a888fc5560e6009a660b949b6dd5b4649e97c6e54d4e6a5cc9f51328" Dec 11 03:04:23 crc kubenswrapper[4824]: I1211 03:04:23.120350 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-v65vx/crc-debug-xznvz" Dec 11 03:04:23 crc kubenswrapper[4824]: I1211 03:04:23.120795 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-v65vx/crc-debug-qpfxr" Dec 11 03:04:23 crc kubenswrapper[4824]: W1211 03:04:23.169875 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbfc6dc48_ab52_4b16_83a4_798b7d75c41c.slice/crio-3bab99af13d540f07082aaba7b36251b356fe4c5d2a86ae30abfe536fcb3de41 WatchSource:0}: Error finding container 3bab99af13d540f07082aaba7b36251b356fe4c5d2a86ae30abfe536fcb3de41: Status 404 returned error can't find the container with id 3bab99af13d540f07082aaba7b36251b356fe4c5d2a86ae30abfe536fcb3de41 Dec 11 03:04:24 crc kubenswrapper[4824]: I1211 03:04:24.135524 4824 generic.go:334] "Generic (PLEG): container finished" podID="bfc6dc48-ab52-4b16-83a4-798b7d75c41c" containerID="2ec7519a44465dd971e277167c768aa35c80a33a99525239df9686ac59fdff02" exitCode=0 Dec 11 03:04:24 crc kubenswrapper[4824]: I1211 03:04:24.135576 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-v65vx/crc-debug-qpfxr" event={"ID":"bfc6dc48-ab52-4b16-83a4-798b7d75c41c","Type":"ContainerDied","Data":"2ec7519a44465dd971e277167c768aa35c80a33a99525239df9686ac59fdff02"} Dec 11 03:04:24 crc kubenswrapper[4824]: I1211 03:04:24.135607 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-v65vx/crc-debug-qpfxr" event={"ID":"bfc6dc48-ab52-4b16-83a4-798b7d75c41c","Type":"ContainerStarted","Data":"3bab99af13d540f07082aaba7b36251b356fe4c5d2a86ae30abfe536fcb3de41"} Dec 11 03:04:24 crc kubenswrapper[4824]: I1211 03:04:24.193191 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-v65vx/crc-debug-qpfxr"] Dec 11 03:04:24 crc kubenswrapper[4824]: I1211 03:04:24.203494 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-v65vx/crc-debug-qpfxr"] Dec 11 03:04:25 crc kubenswrapper[4824]: I1211 03:04:25.260296 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-v65vx/crc-debug-qpfxr" Dec 11 03:04:25 crc kubenswrapper[4824]: I1211 03:04:25.450092 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hh58f\" (UniqueName: \"kubernetes.io/projected/bfc6dc48-ab52-4b16-83a4-798b7d75c41c-kube-api-access-hh58f\") pod \"bfc6dc48-ab52-4b16-83a4-798b7d75c41c\" (UID: \"bfc6dc48-ab52-4b16-83a4-798b7d75c41c\") " Dec 11 03:04:25 crc kubenswrapper[4824]: I1211 03:04:25.450324 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bfc6dc48-ab52-4b16-83a4-798b7d75c41c-host\") pod \"bfc6dc48-ab52-4b16-83a4-798b7d75c41c\" (UID: \"bfc6dc48-ab52-4b16-83a4-798b7d75c41c\") " Dec 11 03:04:25 crc kubenswrapper[4824]: I1211 03:04:25.450384 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bfc6dc48-ab52-4b16-83a4-798b7d75c41c-host" (OuterVolumeSpecName: "host") pod "bfc6dc48-ab52-4b16-83a4-798b7d75c41c" (UID: "bfc6dc48-ab52-4b16-83a4-798b7d75c41c"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 03:04:25 crc kubenswrapper[4824]: I1211 03:04:25.450941 4824 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bfc6dc48-ab52-4b16-83a4-798b7d75c41c-host\") on node \"crc\" DevicePath \"\"" Dec 11 03:04:25 crc kubenswrapper[4824]: I1211 03:04:25.454910 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bfc6dc48-ab52-4b16-83a4-798b7d75c41c-kube-api-access-hh58f" (OuterVolumeSpecName: "kube-api-access-hh58f") pod "bfc6dc48-ab52-4b16-83a4-798b7d75c41c" (UID: "bfc6dc48-ab52-4b16-83a4-798b7d75c41c"). InnerVolumeSpecName "kube-api-access-hh58f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 03:04:25 crc kubenswrapper[4824]: I1211 03:04:25.552971 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hh58f\" (UniqueName: \"kubernetes.io/projected/bfc6dc48-ab52-4b16-83a4-798b7d75c41c-kube-api-access-hh58f\") on node \"crc\" DevicePath \"\"" Dec 11 03:04:26 crc kubenswrapper[4824]: I1211 03:04:26.155545 4824 scope.go:117] "RemoveContainer" containerID="2ec7519a44465dd971e277167c768aa35c80a33a99525239df9686ac59fdff02" Dec 11 03:04:26 crc kubenswrapper[4824]: I1211 03:04:26.155670 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-v65vx/crc-debug-qpfxr" Dec 11 03:04:26 crc kubenswrapper[4824]: I1211 03:04:26.642289 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bfc6dc48-ab52-4b16-83a4-798b7d75c41c" path="/var/lib/kubelet/pods/bfc6dc48-ab52-4b16-83a4-798b7d75c41c/volumes" Dec 11 03:04:30 crc kubenswrapper[4824]: I1211 03:04:30.633200 4824 scope.go:117] "RemoveContainer" containerID="8dcf42d380c145f29ce6b988ed66e3ef33605f29467da12e5fe7eac51562e700" Dec 11 03:04:30 crc kubenswrapper[4824]: E1211 03:04:30.634159 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 03:04:40 crc kubenswrapper[4824]: I1211 03:04:40.539043 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-c88ccd6d6-bg6lk_bd669057-4834-42ed-83ea-f7a454c0a013/barbican-api/0.log" Dec 11 03:04:40 crc kubenswrapper[4824]: I1211 03:04:40.685054 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-c88ccd6d6-bg6lk_bd669057-4834-42ed-83ea-f7a454c0a013/barbican-api-log/0.log" Dec 11 03:04:40 crc kubenswrapper[4824]: I1211 03:04:40.736771 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-664bd5db46-xt8tr_cd747b97-a287-4df6-8b35-f30121c8c0b9/barbican-keystone-listener/0.log" Dec 11 03:04:40 crc kubenswrapper[4824]: I1211 03:04:40.770769 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-664bd5db46-xt8tr_cd747b97-a287-4df6-8b35-f30121c8c0b9/barbican-keystone-listener-log/0.log" Dec 11 03:04:40 crc kubenswrapper[4824]: I1211 03:04:40.901344 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-6967b7c86c-47gl7_33482b7a-6aaa-48c4-9a46-c3de05cacebe/barbican-worker/0.log" Dec 11 03:04:40 crc kubenswrapper[4824]: I1211 03:04:40.917291 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-6967b7c86c-47gl7_33482b7a-6aaa-48c4-9a46-c3de05cacebe/barbican-worker-log/0.log" Dec 11 03:04:41 crc kubenswrapper[4824]: I1211 03:04:41.143820 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-bsknw_1bf6cf36-d7eb-42be-aeee-c6824de8b6ba/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 11 03:04:41 crc kubenswrapper[4824]: I1211 03:04:41.179953 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_c65e7e08-e393-4cd6-b17e-3d9e81783e39/ceilometer-central-agent/0.log" Dec 11 03:04:41 crc kubenswrapper[4824]: I1211 03:04:41.243391 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_c65e7e08-e393-4cd6-b17e-3d9e81783e39/ceilometer-notification-agent/0.log" Dec 11 03:04:41 crc kubenswrapper[4824]: I1211 03:04:41.329848 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_c65e7e08-e393-4cd6-b17e-3d9e81783e39/proxy-httpd/0.log" Dec 11 03:04:41 crc kubenswrapper[4824]: I1211 03:04:41.350791 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_c65e7e08-e393-4cd6-b17e-3d9e81783e39/sg-core/0.log" Dec 11 03:04:41 crc kubenswrapper[4824]: I1211 03:04:41.466699 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_717f2bc0-db62-483f-83e6-4d2f247eb2d6/cinder-api/0.log" Dec 11 03:04:41 crc kubenswrapper[4824]: I1211 03:04:41.502766 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_717f2bc0-db62-483f-83e6-4d2f247eb2d6/cinder-api-log/0.log" Dec 11 03:04:41 crc kubenswrapper[4824]: I1211 03:04:41.595384 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_0dde14f9-85a8-4731-94ef-8199ea6d8a4b/cinder-scheduler/0.log" Dec 11 03:04:41 crc kubenswrapper[4824]: I1211 03:04:41.633213 4824 scope.go:117] "RemoveContainer" containerID="8dcf42d380c145f29ce6b988ed66e3ef33605f29467da12e5fe7eac51562e700" Dec 11 03:04:41 crc kubenswrapper[4824]: E1211 03:04:41.633679 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 03:04:41 crc kubenswrapper[4824]: I1211 03:04:41.654758 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_0dde14f9-85a8-4731-94ef-8199ea6d8a4b/probe/0.log" Dec 11 03:04:41 crc kubenswrapper[4824]: I1211 03:04:41.743843 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-v4fz9_aeee6ae2-66ac-45e6-9a98-753649777abe/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 11 03:04:41 crc kubenswrapper[4824]: I1211 03:04:41.835999 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-4qkdg_6b2de157-739b-48d1-8bcb-585e4e6f475a/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 11 03:04:41 crc kubenswrapper[4824]: I1211 03:04:41.947222 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-mptxc_71c2c849-fc4f-44ac-92d4-75848ea56874/init/0.log" Dec 11 03:04:42 crc kubenswrapper[4824]: I1211 03:04:42.140473 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-mptxc_71c2c849-fc4f-44ac-92d4-75848ea56874/init/0.log" Dec 11 03:04:42 crc kubenswrapper[4824]: I1211 03:04:42.149964 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-mptxc_71c2c849-fc4f-44ac-92d4-75848ea56874/dnsmasq-dns/0.log" Dec 11 03:04:42 crc kubenswrapper[4824]: I1211 03:04:42.171186 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-47r5p_6c6c5ca9-ed78-4f51-a88b-12ec3df9efa2/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Dec 11 03:04:42 crc kubenswrapper[4824]: I1211 03:04:42.343839 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_ac4513fb-74e7-4bf2-8766-ef96066f6c13/glance-httpd/0.log" Dec 11 03:04:42 crc kubenswrapper[4824]: I1211 03:04:42.345737 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_ac4513fb-74e7-4bf2-8766-ef96066f6c13/glance-log/0.log" Dec 11 03:04:42 crc kubenswrapper[4824]: I1211 03:04:42.534851 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_e6776fbd-a091-4715-9942-4dafac0630e2/glance-httpd/0.log" Dec 11 03:04:42 crc kubenswrapper[4824]: I1211 03:04:42.536246 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_e6776fbd-a091-4715-9942-4dafac0630e2/glance-log/0.log" Dec 11 03:04:42 crc kubenswrapper[4824]: I1211 03:04:42.768262 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-74cc6c88d8-zvlk7_1da5eb01-d709-42ad-b5f1-e7ea6310cc9e/horizon/0.log" Dec 11 03:04:42 crc kubenswrapper[4824]: I1211 03:04:42.867201 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-lw68c_20ff7897-84b7-4c8c-b9f9-fc7721488ea7/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Dec 11 03:04:43 crc kubenswrapper[4824]: I1211 03:04:43.022560 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-9k74z_a786c754-b2c4-4f19-b082-e8d64d47ac44/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 11 03:04:43 crc kubenswrapper[4824]: I1211 03:04:43.066716 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-74cc6c88d8-zvlk7_1da5eb01-d709-42ad-b5f1-e7ea6310cc9e/horizon-log/0.log" Dec 11 03:04:43 crc kubenswrapper[4824]: I1211 03:04:43.305941 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29423701-hlhbx_a20415c4-3d24-4424-b32f-9e88eb6305e3/keystone-cron/0.log" Dec 11 03:04:43 crc kubenswrapper[4824]: I1211 03:04:43.340428 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-6c99457c66-dcp4p_55bc7333-3e36-4d58-9451-fe018b085c0b/keystone-api/0.log" Dec 11 03:04:43 crc kubenswrapper[4824]: I1211 03:04:43.554382 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_1ed2f6d5-a61f-4c5b-9b73-1089d13cf9b6/kube-state-metrics/0.log" Dec 11 03:04:43 crc kubenswrapper[4824]: I1211 03:04:43.557481 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-rs2vt_bc0c56ce-23bd-4883-b3dd-879d58c540b5/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Dec 11 03:04:43 crc kubenswrapper[4824]: I1211 03:04:43.909630 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5b766ddd67-lflrw_cde933a7-1f3a-4567-b6fd-7644a02cb745/neutron-api/0.log" Dec 11 03:04:43 crc kubenswrapper[4824]: I1211 03:04:43.923369 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5b766ddd67-lflrw_cde933a7-1f3a-4567-b6fd-7644a02cb745/neutron-httpd/0.log" Dec 11 03:04:44 crc kubenswrapper[4824]: I1211 03:04:44.097910 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-8rbcc_8736a409-d9d8-437e-965a-630f1ee6ae85/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Dec 11 03:04:44 crc kubenswrapper[4824]: I1211 03:04:44.613288 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_a36b8e48-3fc4-4132-a824-a0078424150f/nova-api-log/0.log" Dec 11 03:04:44 crc kubenswrapper[4824]: I1211 03:04:44.643571 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_1443d344-6dc0-4aab-96fd-522ae53f31cc/nova-cell0-conductor-conductor/0.log" Dec 11 03:04:44 crc kubenswrapper[4824]: I1211 03:04:44.905609 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_fddad47b-d680-4be3-8bfa-a7708daef973/nova-cell1-conductor-conductor/0.log" Dec 11 03:04:44 crc kubenswrapper[4824]: I1211 03:04:44.950191 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_c246a287-b22f-405b-8a24-c50765f5b153/nova-cell1-novncproxy-novncproxy/0.log" Dec 11 03:04:44 crc kubenswrapper[4824]: I1211 03:04:44.975976 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_a36b8e48-3fc4-4132-a824-a0078424150f/nova-api-api/0.log" Dec 11 03:04:45 crc kubenswrapper[4824]: I1211 03:04:45.166054 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-7x98s_f844f569-89fd-410c-b2e1-5f9883d9127a/nova-edpm-deployment-openstack-edpm-ipam/0.log" Dec 11 03:04:45 crc kubenswrapper[4824]: I1211 03:04:45.313254 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_e0e1ab61-fe6b-4b2e-aa01-2bf77e119ce4/nova-metadata-log/0.log" Dec 11 03:04:45 crc kubenswrapper[4824]: I1211 03:04:45.577920 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_bc3713cb-287a-4350-a147-e13a5151ac71/nova-scheduler-scheduler/0.log" Dec 11 03:04:45 crc kubenswrapper[4824]: I1211 03:04:45.671550 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_f35efb32-a031-4c75-9358-3f5143335131/mysql-bootstrap/0.log" Dec 11 03:04:45 crc kubenswrapper[4824]: I1211 03:04:45.832790 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_f35efb32-a031-4c75-9358-3f5143335131/mysql-bootstrap/0.log" Dec 11 03:04:45 crc kubenswrapper[4824]: I1211 03:04:45.850911 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_f35efb32-a031-4c75-9358-3f5143335131/galera/0.log" Dec 11 03:04:46 crc kubenswrapper[4824]: I1211 03:04:46.019695 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_6e5994f4-2c9e-4e8c-b7d3-84a15af846cd/mysql-bootstrap/0.log" Dec 11 03:04:46 crc kubenswrapper[4824]: I1211 03:04:46.201288 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_6e5994f4-2c9e-4e8c-b7d3-84a15af846cd/mysql-bootstrap/0.log" Dec 11 03:04:46 crc kubenswrapper[4824]: I1211 03:04:46.208006 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_6e5994f4-2c9e-4e8c-b7d3-84a15af846cd/galera/0.log" Dec 11 03:04:46 crc kubenswrapper[4824]: I1211 03:04:46.406506 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_a66f0fef-8984-4c67-b797-6762b404cd36/openstackclient/0.log" Dec 11 03:04:46 crc kubenswrapper[4824]: I1211 03:04:46.476205 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-779x4_148c69f8-3121-4a55-b3a6-be44b816f643/ovn-controller/0.log" Dec 11 03:04:46 crc kubenswrapper[4824]: I1211 03:04:46.516564 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_e0e1ab61-fe6b-4b2e-aa01-2bf77e119ce4/nova-metadata-metadata/0.log" Dec 11 03:04:46 crc kubenswrapper[4824]: I1211 03:04:46.623829 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-9fj44_d4fa84ff-af7c-4d69-9dfd-e4c1f20b0ce9/openstack-network-exporter/0.log" Dec 11 03:04:46 crc kubenswrapper[4824]: I1211 03:04:46.770439 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-xzkzd_17cc80fc-a51f-41e4-be0f-f593fc23476e/ovsdb-server-init/0.log" Dec 11 03:04:46 crc kubenswrapper[4824]: I1211 03:04:46.899883 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-xzkzd_17cc80fc-a51f-41e4-be0f-f593fc23476e/ovs-vswitchd/0.log" Dec 11 03:04:46 crc kubenswrapper[4824]: I1211 03:04:46.911364 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-xzkzd_17cc80fc-a51f-41e4-be0f-f593fc23476e/ovsdb-server-init/0.log" Dec 11 03:04:46 crc kubenswrapper[4824]: I1211 03:04:46.970436 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-xzkzd_17cc80fc-a51f-41e4-be0f-f593fc23476e/ovsdb-server/0.log" Dec 11 03:04:47 crc kubenswrapper[4824]: I1211 03:04:47.143716 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-grms5_c8ca447b-581a-45ba-af85-a1a4ff83687f/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Dec 11 03:04:47 crc kubenswrapper[4824]: I1211 03:04:47.177687 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_964729d8-30e0-4d0b-ae8e-6f9bfc2536d0/openstack-network-exporter/0.log" Dec 11 03:04:47 crc kubenswrapper[4824]: I1211 03:04:47.231303 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_964729d8-30e0-4d0b-ae8e-6f9bfc2536d0/ovn-northd/0.log" Dec 11 03:04:47 crc kubenswrapper[4824]: I1211 03:04:47.316380 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_068df61c-d193-4911-b077-fd80d25fa4de/openstack-network-exporter/0.log" Dec 11 03:04:47 crc kubenswrapper[4824]: I1211 03:04:47.456729 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_068df61c-d193-4911-b077-fd80d25fa4de/ovsdbserver-nb/0.log" Dec 11 03:04:47 crc kubenswrapper[4824]: I1211 03:04:47.543871 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_402ac70d-73e7-4697-ac46-a92bbbaf1aac/openstack-network-exporter/0.log" Dec 11 03:04:47 crc kubenswrapper[4824]: I1211 03:04:47.597706 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_402ac70d-73e7-4697-ac46-a92bbbaf1aac/ovsdbserver-sb/0.log" Dec 11 03:04:47 crc kubenswrapper[4824]: I1211 03:04:47.723158 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-78b68944cb-4p8xq_d4663f55-6344-4154-9df6-96a17ca59bdb/placement-api/0.log" Dec 11 03:04:47 crc kubenswrapper[4824]: I1211 03:04:47.816701 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-78b68944cb-4p8xq_d4663f55-6344-4154-9df6-96a17ca59bdb/placement-log/0.log" Dec 11 03:04:47 crc kubenswrapper[4824]: I1211 03:04:47.877293 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_89eddaec-a9d6-4b73-b654-2f176d45d162/setup-container/0.log" Dec 11 03:04:48 crc kubenswrapper[4824]: I1211 03:04:48.060985 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_8799d853-613d-4376-a372-7dedd968f264/setup-container/0.log" Dec 11 03:04:48 crc kubenswrapper[4824]: I1211 03:04:48.074175 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_89eddaec-a9d6-4b73-b654-2f176d45d162/setup-container/0.log" Dec 11 03:04:48 crc kubenswrapper[4824]: I1211 03:04:48.114824 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_89eddaec-a9d6-4b73-b654-2f176d45d162/rabbitmq/0.log" Dec 11 03:04:48 crc kubenswrapper[4824]: I1211 03:04:48.301754 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_8799d853-613d-4376-a372-7dedd968f264/setup-container/0.log" Dec 11 03:04:48 crc kubenswrapper[4824]: I1211 03:04:48.321734 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_8799d853-613d-4376-a372-7dedd968f264/rabbitmq/0.log" Dec 11 03:04:48 crc kubenswrapper[4824]: I1211 03:04:48.339998 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-wx57q_a24e9b38-d375-4b6b-a641-12aa6bf0fe18/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 11 03:04:48 crc kubenswrapper[4824]: I1211 03:04:48.663674 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-4292s_17d30a7d-e050-463d-ba60-bc8ae69cb21e/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Dec 11 03:04:48 crc kubenswrapper[4824]: I1211 03:04:48.694125 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-nl76s_09a89381-9670-4aff-965b-bd6d0d3e5ae7/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 11 03:04:48 crc kubenswrapper[4824]: I1211 03:04:48.840599 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-gk2qc_ab605768-f69e-4079-892a-a5b78e8dfc7d/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 11 03:04:48 crc kubenswrapper[4824]: I1211 03:04:48.919540 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-pkkbb_bf0de509-4023-4c82-b39a-f42d78d535a7/ssh-known-hosts-edpm-deployment/0.log" Dec 11 03:04:49 crc kubenswrapper[4824]: I1211 03:04:49.168889 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-6458cc7487-d2555_ead63438-3159-4a5e-889f-5f7c8a8d6a30/proxy-httpd/0.log" Dec 11 03:04:49 crc kubenswrapper[4824]: I1211 03:04:49.170053 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-6458cc7487-d2555_ead63438-3159-4a5e-889f-5f7c8a8d6a30/proxy-server/0.log" Dec 11 03:04:49 crc kubenswrapper[4824]: I1211 03:04:49.301693 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-zbgp5_5719adff-b313-426e-afd1-69986bdc81bc/swift-ring-rebalance/0.log" Dec 11 03:04:49 crc kubenswrapper[4824]: I1211 03:04:49.369971 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7913c5de-48cb-4e9a-8b9b-ee976476bab2/account-auditor/0.log" Dec 11 03:04:49 crc kubenswrapper[4824]: I1211 03:04:49.433415 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7913c5de-48cb-4e9a-8b9b-ee976476bab2/account-reaper/0.log" Dec 11 03:04:49 crc kubenswrapper[4824]: I1211 03:04:49.501181 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7913c5de-48cb-4e9a-8b9b-ee976476bab2/account-replicator/0.log" Dec 11 03:04:49 crc kubenswrapper[4824]: I1211 03:04:49.547034 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7913c5de-48cb-4e9a-8b9b-ee976476bab2/account-server/0.log" Dec 11 03:04:49 crc kubenswrapper[4824]: I1211 03:04:49.638890 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7913c5de-48cb-4e9a-8b9b-ee976476bab2/container-auditor/0.log" Dec 11 03:04:49 crc kubenswrapper[4824]: I1211 03:04:49.710641 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7913c5de-48cb-4e9a-8b9b-ee976476bab2/container-server/0.log" Dec 11 03:04:49 crc kubenswrapper[4824]: I1211 03:04:49.756872 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7913c5de-48cb-4e9a-8b9b-ee976476bab2/container-replicator/0.log" Dec 11 03:04:49 crc kubenswrapper[4824]: I1211 03:04:49.803228 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7913c5de-48cb-4e9a-8b9b-ee976476bab2/container-updater/0.log" Dec 11 03:04:49 crc kubenswrapper[4824]: I1211 03:04:49.867515 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7913c5de-48cb-4e9a-8b9b-ee976476bab2/object-auditor/0.log" Dec 11 03:04:49 crc kubenswrapper[4824]: I1211 03:04:49.941152 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7913c5de-48cb-4e9a-8b9b-ee976476bab2/object-expirer/0.log" Dec 11 03:04:50 crc kubenswrapper[4824]: I1211 03:04:50.013038 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7913c5de-48cb-4e9a-8b9b-ee976476bab2/object-replicator/0.log" Dec 11 03:04:50 crc kubenswrapper[4824]: I1211 03:04:50.037905 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7913c5de-48cb-4e9a-8b9b-ee976476bab2/object-server/0.log" Dec 11 03:04:50 crc kubenswrapper[4824]: I1211 03:04:50.126684 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7913c5de-48cb-4e9a-8b9b-ee976476bab2/object-updater/0.log" Dec 11 03:04:50 crc kubenswrapper[4824]: I1211 03:04:50.140809 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7913c5de-48cb-4e9a-8b9b-ee976476bab2/rsync/0.log" Dec 11 03:04:50 crc kubenswrapper[4824]: I1211 03:04:50.163450 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7913c5de-48cb-4e9a-8b9b-ee976476bab2/swift-recon-cron/0.log" Dec 11 03:04:50 crc kubenswrapper[4824]: I1211 03:04:50.345317 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_ab73a048-199c-464e-98ab-2e87063161b6/tempest-tests-tempest-tests-runner/0.log" Dec 11 03:04:50 crc kubenswrapper[4824]: I1211 03:04:50.408753 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-czhq8_f7ed3232-8b86-4804-85e1-0b87715ed1f2/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Dec 11 03:04:50 crc kubenswrapper[4824]: I1211 03:04:50.496911 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_1283daba-11e7-4232-81dd-c4316478f297/test-operator-logs-container/0.log" Dec 11 03:04:50 crc kubenswrapper[4824]: I1211 03:04:50.628140 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-t4pfx_f694423f-b8e7-49bd-9032-4b29a64fda4e/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 11 03:04:54 crc kubenswrapper[4824]: I1211 03:04:54.632464 4824 scope.go:117] "RemoveContainer" containerID="8dcf42d380c145f29ce6b988ed66e3ef33605f29467da12e5fe7eac51562e700" Dec 11 03:04:54 crc kubenswrapper[4824]: E1211 03:04:54.633302 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 03:04:57 crc kubenswrapper[4824]: I1211 03:04:57.765888 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_709f3d23-a475-443a-8787-7f018c486b56/memcached/0.log" Dec 11 03:05:07 crc kubenswrapper[4824]: I1211 03:05:07.637683 4824 scope.go:117] "RemoveContainer" containerID="8dcf42d380c145f29ce6b988ed66e3ef33605f29467da12e5fe7eac51562e700" Dec 11 03:05:07 crc kubenswrapper[4824]: E1211 03:05:07.638550 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 03:05:15 crc kubenswrapper[4824]: I1211 03:05:15.469288 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05vd4cz_487cf0fc-4368-4f0e-878a-d3cd3d1e54f4/util/0.log" Dec 11 03:05:15 crc kubenswrapper[4824]: I1211 03:05:15.590411 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05vd4cz_487cf0fc-4368-4f0e-878a-d3cd3d1e54f4/util/0.log" Dec 11 03:05:15 crc kubenswrapper[4824]: I1211 03:05:15.597762 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05vd4cz_487cf0fc-4368-4f0e-878a-d3cd3d1e54f4/pull/0.log" Dec 11 03:05:15 crc kubenswrapper[4824]: I1211 03:05:15.626529 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05vd4cz_487cf0fc-4368-4f0e-878a-d3cd3d1e54f4/pull/0.log" Dec 11 03:05:15 crc kubenswrapper[4824]: I1211 03:05:15.831629 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05vd4cz_487cf0fc-4368-4f0e-878a-d3cd3d1e54f4/pull/0.log" Dec 11 03:05:15 crc kubenswrapper[4824]: I1211 03:05:15.874020 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05vd4cz_487cf0fc-4368-4f0e-878a-d3cd3d1e54f4/util/0.log" Dec 11 03:05:15 crc kubenswrapper[4824]: I1211 03:05:15.885503 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05vd4cz_487cf0fc-4368-4f0e-878a-d3cd3d1e54f4/extract/0.log" Dec 11 03:05:16 crc kubenswrapper[4824]: I1211 03:05:16.029481 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-r8tnj_df6864cf-9559-4398-bb3c-0f58f4c45563/kube-rbac-proxy/0.log" Dec 11 03:05:16 crc kubenswrapper[4824]: I1211 03:05:16.092448 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-r8tnj_df6864cf-9559-4398-bb3c-0f58f4c45563/manager/0.log" Dec 11 03:05:16 crc kubenswrapper[4824]: I1211 03:05:16.127463 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6c677c69b-8hvxh_0b7b142b-04e2-4bb3-98ff-ed3b4da63746/kube-rbac-proxy/0.log" Dec 11 03:05:16 crc kubenswrapper[4824]: I1211 03:05:16.240851 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6c677c69b-8hvxh_0b7b142b-04e2-4bb3-98ff-ed3b4da63746/manager/0.log" Dec 11 03:05:16 crc kubenswrapper[4824]: I1211 03:05:16.265486 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-697fb699cf-4b2cb_6743e907-9570-4dde-bb99-83e67202b224/kube-rbac-proxy/0.log" Dec 11 03:05:16 crc kubenswrapper[4824]: I1211 03:05:16.355255 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-697fb699cf-4b2cb_6743e907-9570-4dde-bb99-83e67202b224/manager/0.log" Dec 11 03:05:16 crc kubenswrapper[4824]: I1211 03:05:16.521590 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5697bb5779-4cjld_0eeea3b6-c514-4547-a2e0-52870ef572c6/kube-rbac-proxy/0.log" Dec 11 03:05:16 crc kubenswrapper[4824]: I1211 03:05:16.636458 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5697bb5779-4cjld_0eeea3b6-c514-4547-a2e0-52870ef572c6/manager/0.log" Dec 11 03:05:16 crc kubenswrapper[4824]: I1211 03:05:16.675892 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-l9jrd_db4453b9-ead1-45fc-91c5-d90541fa9c78/kube-rbac-proxy/0.log" Dec 11 03:05:16 crc kubenswrapper[4824]: I1211 03:05:16.726972 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-l9jrd_db4453b9-ead1-45fc-91c5-d90541fa9c78/manager/0.log" Dec 11 03:05:16 crc kubenswrapper[4824]: I1211 03:05:16.826652 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-5qp6v_1fc0b6e3-15c9-4f14-90f8-3f39a04ebc2c/kube-rbac-proxy/0.log" Dec 11 03:05:16 crc kubenswrapper[4824]: I1211 03:05:16.883780 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-5qp6v_1fc0b6e3-15c9-4f14-90f8-3f39a04ebc2c/manager/0.log" Dec 11 03:05:17 crc kubenswrapper[4824]: I1211 03:05:17.032230 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-78d48bff9d-j68mf_f219b9bf-34b1-48c4-96c8-20580b6ec9a4/kube-rbac-proxy/0.log" Dec 11 03:05:17 crc kubenswrapper[4824]: I1211 03:05:17.221085 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-967d97867-5tpj5_aea47aaf-ee88-45a0-87b7-375b9f656828/kube-rbac-proxy/0.log" Dec 11 03:05:17 crc kubenswrapper[4824]: I1211 03:05:17.241077 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-78d48bff9d-j68mf_f219b9bf-34b1-48c4-96c8-20580b6ec9a4/manager/0.log" Dec 11 03:05:17 crc kubenswrapper[4824]: I1211 03:05:17.260318 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-967d97867-5tpj5_aea47aaf-ee88-45a0-87b7-375b9f656828/manager/0.log" Dec 11 03:05:17 crc kubenswrapper[4824]: I1211 03:05:17.409457 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-4np4x_cec1f29f-a1db-495f-b59c-e9308a9b53a2/kube-rbac-proxy/0.log" Dec 11 03:05:17 crc kubenswrapper[4824]: I1211 03:05:17.531407 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-4np4x_cec1f29f-a1db-495f-b59c-e9308a9b53a2/manager/0.log" Dec 11 03:05:17 crc kubenswrapper[4824]: I1211 03:05:17.623762 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5b5fd79c9c-w24mx_46bf5c63-42b4-45bd-ade3-446a4b2e4f6e/kube-rbac-proxy/0.log" Dec 11 03:05:17 crc kubenswrapper[4824]: I1211 03:05:17.624564 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5b5fd79c9c-w24mx_46bf5c63-42b4-45bd-ade3-446a4b2e4f6e/manager/0.log" Dec 11 03:05:17 crc kubenswrapper[4824]: I1211 03:05:17.751692 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-79c8c4686c-hpjng_fd3368a6-77e2-4c8f-b535-1d602b794d7d/kube-rbac-proxy/0.log" Dec 11 03:05:17 crc kubenswrapper[4824]: I1211 03:05:17.851606 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-79c8c4686c-hpjng_fd3368a6-77e2-4c8f-b535-1d602b794d7d/manager/0.log" Dec 11 03:05:17 crc kubenswrapper[4824]: I1211 03:05:17.922342 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-lmszx_dca85063-ba7a-45ac-ab88-5498a44834ba/kube-rbac-proxy/0.log" Dec 11 03:05:17 crc kubenswrapper[4824]: I1211 03:05:17.994393 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-lmszx_dca85063-ba7a-45ac-ab88-5498a44834ba/manager/0.log" Dec 11 03:05:18 crc kubenswrapper[4824]: I1211 03:05:18.025803 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-zkz72_b206f71d-32cd-4b7d-ae38-a5125ab481dd/kube-rbac-proxy/0.log" Dec 11 03:05:18 crc kubenswrapper[4824]: I1211 03:05:18.205012 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-zkz72_b206f71d-32cd-4b7d-ae38-a5125ab481dd/manager/0.log" Dec 11 03:05:18 crc kubenswrapper[4824]: I1211 03:05:18.248585 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-qqpf2_f1183bb7-356d-43b7-8ee7-55ab60dcbe54/kube-rbac-proxy/0.log" Dec 11 03:05:18 crc kubenswrapper[4824]: I1211 03:05:18.281334 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-qqpf2_f1183bb7-356d-43b7-8ee7-55ab60dcbe54/manager/0.log" Dec 11 03:05:18 crc kubenswrapper[4824]: I1211 03:05:18.404253 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-84b575879f4k8rk_cb370fee-74a4-4f43-98a9-3df8c6e61335/kube-rbac-proxy/0.log" Dec 11 03:05:18 crc kubenswrapper[4824]: I1211 03:05:18.482223 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-84b575879f4k8rk_cb370fee-74a4-4f43-98a9-3df8c6e61335/manager/0.log" Dec 11 03:05:18 crc kubenswrapper[4824]: I1211 03:05:18.904778 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-7795b79658-tzqvv_52218f50-e21e-4566-b323-9b5ece7a5e35/operator/0.log" Dec 11 03:05:19 crc kubenswrapper[4824]: I1211 03:05:19.028671 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-nb526_ab26e0f0-f195-481f-84bd-0a7bca8d0432/registry-server/0.log" Dec 11 03:05:19 crc kubenswrapper[4824]: I1211 03:05:19.157142 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-r2bsg_4d26d90a-05a8-4960-b2ca-12fa220383b7/kube-rbac-proxy/0.log" Dec 11 03:05:19 crc kubenswrapper[4824]: I1211 03:05:19.189669 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-r2bsg_4d26d90a-05a8-4960-b2ca-12fa220383b7/manager/0.log" Dec 11 03:05:19 crc kubenswrapper[4824]: I1211 03:05:19.284989 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-xx9gc_10325d55-7efd-49e8-95e9-46d68c7cd7ae/kube-rbac-proxy/0.log" Dec 11 03:05:19 crc kubenswrapper[4824]: I1211 03:05:19.407926 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-xx9gc_10325d55-7efd-49e8-95e9-46d68c7cd7ae/manager/0.log" Dec 11 03:05:19 crc kubenswrapper[4824]: I1211 03:05:19.466434 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-wsjbh_34bc4184-9cfb-4278-a9b9-80bf2f1ae612/operator/0.log" Dec 11 03:05:19 crc kubenswrapper[4824]: I1211 03:05:19.636016 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-9d58d64bc-fmvm7_8047af9c-3a6d-4b2e-a088-df967e42671c/kube-rbac-proxy/0.log" Dec 11 03:05:19 crc kubenswrapper[4824]: I1211 03:05:19.694628 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-9d58d64bc-fmvm7_8047af9c-3a6d-4b2e-a088-df967e42671c/manager/0.log" Dec 11 03:05:19 crc kubenswrapper[4824]: I1211 03:05:19.777059 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-58d5ff84df-qhhzn_aff71eda-47bc-4b47-89df-fe9bf8357fba/kube-rbac-proxy/0.log" Dec 11 03:05:19 crc kubenswrapper[4824]: I1211 03:05:19.788895 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-56cf947496-nhjjr_7180c14a-0c3a-4267-b122-5c30c5685d60/manager/0.log" Dec 11 03:05:19 crc kubenswrapper[4824]: I1211 03:05:19.902051 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-58d5ff84df-qhhzn_aff71eda-47bc-4b47-89df-fe9bf8357fba/manager/0.log" Dec 11 03:05:19 crc kubenswrapper[4824]: I1211 03:05:19.915432 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-4zfbk_cf19e8a0-8f29-48a4-b715-f499644fc896/kube-rbac-proxy/0.log" Dec 11 03:05:20 crc kubenswrapper[4824]: I1211 03:05:20.012619 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-4zfbk_cf19e8a0-8f29-48a4-b715-f499644fc896/manager/0.log" Dec 11 03:05:20 crc kubenswrapper[4824]: I1211 03:05:20.075499 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-75944c9b7-h6bgf_9793f7fe-2e1b-46c0-985e-1f77733f925f/manager/0.log" Dec 11 03:05:20 crc kubenswrapper[4824]: I1211 03:05:20.079815 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-75944c9b7-h6bgf_9793f7fe-2e1b-46c0-985e-1f77733f925f/kube-rbac-proxy/0.log" Dec 11 03:05:21 crc kubenswrapper[4824]: I1211 03:05:21.633207 4824 scope.go:117] "RemoveContainer" containerID="8dcf42d380c145f29ce6b988ed66e3ef33605f29467da12e5fe7eac51562e700" Dec 11 03:05:21 crc kubenswrapper[4824]: E1211 03:05:21.633784 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 03:05:35 crc kubenswrapper[4824]: I1211 03:05:35.632510 4824 scope.go:117] "RemoveContainer" containerID="8dcf42d380c145f29ce6b988ed66e3ef33605f29467da12e5fe7eac51562e700" Dec 11 03:05:35 crc kubenswrapper[4824]: E1211 03:05:35.633399 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 03:05:40 crc kubenswrapper[4824]: I1211 03:05:40.113124 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-gp8dt_c70805eb-e095-4462-a202-e01860559471/control-plane-machine-set-operator/0.log" Dec 11 03:05:40 crc kubenswrapper[4824]: I1211 03:05:40.278507 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-d89dq_7704cded-5ba1-4474-8a31-c0ba947b7679/kube-rbac-proxy/0.log" Dec 11 03:05:40 crc kubenswrapper[4824]: I1211 03:05:40.295947 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-d89dq_7704cded-5ba1-4474-8a31-c0ba947b7679/machine-api-operator/0.log" Dec 11 03:05:47 crc kubenswrapper[4824]: I1211 03:05:47.633838 4824 scope.go:117] "RemoveContainer" containerID="8dcf42d380c145f29ce6b988ed66e3ef33605f29467da12e5fe7eac51562e700" Dec 11 03:05:48 crc kubenswrapper[4824]: I1211 03:05:48.431052 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" event={"ID":"44b156e4-64a4-4d45-aa5e-9b10a862faed","Type":"ContainerStarted","Data":"a0748aabb594d1c62d2ac36a888d8e889bc711466d2b3fd7120852453140e334"} Dec 11 03:05:53 crc kubenswrapper[4824]: I1211 03:05:53.936979 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-7jv9b_81bcc53d-3664-4987-8724-765fda6c8c09/cert-manager-controller/0.log" Dec 11 03:05:54 crc kubenswrapper[4824]: I1211 03:05:54.101510 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-bstmj_919ec639-d2fa-4447-86a4-a88b39d05ec3/cert-manager-cainjector/0.log" Dec 11 03:05:54 crc kubenswrapper[4824]: I1211 03:05:54.173142 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-hxtz9_b9bca511-59e3-4650-86c8-a82c0175d51a/cert-manager-webhook/0.log" Dec 11 03:06:09 crc kubenswrapper[4824]: I1211 03:06:09.145152 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-6hcp5_ffe586ab-061d-47f6-a3e8-220bf9ffebe2/nmstate-console-plugin/0.log" Dec 11 03:06:09 crc kubenswrapper[4824]: I1211 03:06:09.343352 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-b4h4m_22b4f565-3eff-4432-8171-56fa15cf7fcb/nmstate-handler/0.log" Dec 11 03:06:09 crc kubenswrapper[4824]: I1211 03:06:09.436986 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-qpjg9_731b5801-36ca-4f1d-8a34-4ad8554a0c3b/kube-rbac-proxy/0.log" Dec 11 03:06:09 crc kubenswrapper[4824]: I1211 03:06:09.454183 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-qpjg9_731b5801-36ca-4f1d-8a34-4ad8554a0c3b/nmstate-metrics/0.log" Dec 11 03:06:09 crc kubenswrapper[4824]: I1211 03:06:09.650481 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-fd644_4ad81e5d-ce6e-4e3c-ae63-7a87ce2a4def/nmstate-operator/0.log" Dec 11 03:06:09 crc kubenswrapper[4824]: I1211 03:06:09.651397 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-8v6lh_2b946a34-6de7-4afe-8876-18b15e09ea98/nmstate-webhook/0.log" Dec 11 03:06:25 crc kubenswrapper[4824]: I1211 03:06:25.555351 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-r49c8_7fbb3013-54ad-421f-8022-26b0ffad44f6/kube-rbac-proxy/0.log" Dec 11 03:06:25 crc kubenswrapper[4824]: I1211 03:06:25.609784 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-r49c8_7fbb3013-54ad-421f-8022-26b0ffad44f6/controller/0.log" Dec 11 03:06:25 crc kubenswrapper[4824]: I1211 03:06:25.721752 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-s7hz7_57d0325f-b062-4148-8b2a-c21f42de39a3/cp-frr-files/0.log" Dec 11 03:06:25 crc kubenswrapper[4824]: I1211 03:06:25.937913 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-s7hz7_57d0325f-b062-4148-8b2a-c21f42de39a3/cp-frr-files/0.log" Dec 11 03:06:25 crc kubenswrapper[4824]: I1211 03:06:25.952516 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-s7hz7_57d0325f-b062-4148-8b2a-c21f42de39a3/cp-reloader/0.log" Dec 11 03:06:25 crc kubenswrapper[4824]: I1211 03:06:25.970730 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-s7hz7_57d0325f-b062-4148-8b2a-c21f42de39a3/cp-reloader/0.log" Dec 11 03:06:26 crc kubenswrapper[4824]: I1211 03:06:26.005553 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-s7hz7_57d0325f-b062-4148-8b2a-c21f42de39a3/cp-metrics/0.log" Dec 11 03:06:26 crc kubenswrapper[4824]: I1211 03:06:26.211238 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-s7hz7_57d0325f-b062-4148-8b2a-c21f42de39a3/cp-metrics/0.log" Dec 11 03:06:26 crc kubenswrapper[4824]: I1211 03:06:26.211380 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-s7hz7_57d0325f-b062-4148-8b2a-c21f42de39a3/cp-metrics/0.log" Dec 11 03:06:26 crc kubenswrapper[4824]: I1211 03:06:26.212909 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-s7hz7_57d0325f-b062-4148-8b2a-c21f42de39a3/cp-frr-files/0.log" Dec 11 03:06:26 crc kubenswrapper[4824]: I1211 03:06:26.240285 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-s7hz7_57d0325f-b062-4148-8b2a-c21f42de39a3/cp-reloader/0.log" Dec 11 03:06:26 crc kubenswrapper[4824]: I1211 03:06:26.388779 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-s7hz7_57d0325f-b062-4148-8b2a-c21f42de39a3/cp-frr-files/0.log" Dec 11 03:06:26 crc kubenswrapper[4824]: I1211 03:06:26.434542 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-s7hz7_57d0325f-b062-4148-8b2a-c21f42de39a3/cp-reloader/0.log" Dec 11 03:06:26 crc kubenswrapper[4824]: I1211 03:06:26.438836 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-s7hz7_57d0325f-b062-4148-8b2a-c21f42de39a3/cp-metrics/0.log" Dec 11 03:06:26 crc kubenswrapper[4824]: I1211 03:06:26.471785 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-s7hz7_57d0325f-b062-4148-8b2a-c21f42de39a3/controller/0.log" Dec 11 03:06:26 crc kubenswrapper[4824]: I1211 03:06:26.598171 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-s7hz7_57d0325f-b062-4148-8b2a-c21f42de39a3/frr-metrics/0.log" Dec 11 03:06:26 crc kubenswrapper[4824]: I1211 03:06:26.641014 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-s7hz7_57d0325f-b062-4148-8b2a-c21f42de39a3/kube-rbac-proxy/0.log" Dec 11 03:06:26 crc kubenswrapper[4824]: I1211 03:06:26.673069 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-s7hz7_57d0325f-b062-4148-8b2a-c21f42de39a3/kube-rbac-proxy-frr/0.log" Dec 11 03:06:26 crc kubenswrapper[4824]: I1211 03:06:26.810025 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-s7hz7_57d0325f-b062-4148-8b2a-c21f42de39a3/reloader/0.log" Dec 11 03:06:26 crc kubenswrapper[4824]: I1211 03:06:26.919477 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-mnh62_053662c4-f168-42b3-a9c9-dc21ca9908ea/frr-k8s-webhook-server/0.log" Dec 11 03:06:27 crc kubenswrapper[4824]: I1211 03:06:27.098319 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-7d6477f866-c4dqb_47581f0c-dffe-42e8-8211-bb40625aadf1/manager/0.log" Dec 11 03:06:27 crc kubenswrapper[4824]: I1211 03:06:27.266835 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-6cb87f7d4c-jt2ns_0c1d110c-0bd5-410f-af71-7af964dfb17d/webhook-server/0.log" Dec 11 03:06:27 crc kubenswrapper[4824]: I1211 03:06:27.372279 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-d99pv_6e59f040-b574-48e8-9a0d-05072fc1a3f6/kube-rbac-proxy/0.log" Dec 11 03:06:27 crc kubenswrapper[4824]: I1211 03:06:27.814129 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-s7hz7_57d0325f-b062-4148-8b2a-c21f42de39a3/frr/0.log" Dec 11 03:06:27 crc kubenswrapper[4824]: I1211 03:06:27.841370 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-d99pv_6e59f040-b574-48e8-9a0d-05072fc1a3f6/speaker/0.log" Dec 11 03:06:41 crc kubenswrapper[4824]: I1211 03:06:41.945182 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f97jkp_bb38965f-b3bb-4d2e-8430-8c29298730e4/util/0.log" Dec 11 03:06:42 crc kubenswrapper[4824]: I1211 03:06:42.096419 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f97jkp_bb38965f-b3bb-4d2e-8430-8c29298730e4/util/0.log" Dec 11 03:06:42 crc kubenswrapper[4824]: I1211 03:06:42.136267 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f97jkp_bb38965f-b3bb-4d2e-8430-8c29298730e4/pull/0.log" Dec 11 03:06:42 crc kubenswrapper[4824]: I1211 03:06:42.161090 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f97jkp_bb38965f-b3bb-4d2e-8430-8c29298730e4/pull/0.log" Dec 11 03:06:42 crc kubenswrapper[4824]: I1211 03:06:42.315035 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f97jkp_bb38965f-b3bb-4d2e-8430-8c29298730e4/pull/0.log" Dec 11 03:06:42 crc kubenswrapper[4824]: I1211 03:06:42.350251 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f97jkp_bb38965f-b3bb-4d2e-8430-8c29298730e4/extract/0.log" Dec 11 03:06:42 crc kubenswrapper[4824]: I1211 03:06:42.361716 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f97jkp_bb38965f-b3bb-4d2e-8430-8c29298730e4/util/0.log" Dec 11 03:06:42 crc kubenswrapper[4824]: I1211 03:06:42.494971 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83x97sf_55c9ce15-2335-433b-a7ac-cfdf1a3ef599/util/0.log" Dec 11 03:06:42 crc kubenswrapper[4824]: I1211 03:06:42.686037 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83x97sf_55c9ce15-2335-433b-a7ac-cfdf1a3ef599/pull/0.log" Dec 11 03:06:42 crc kubenswrapper[4824]: I1211 03:06:42.696802 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83x97sf_55c9ce15-2335-433b-a7ac-cfdf1a3ef599/util/0.log" Dec 11 03:06:42 crc kubenswrapper[4824]: I1211 03:06:42.717855 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83x97sf_55c9ce15-2335-433b-a7ac-cfdf1a3ef599/pull/0.log" Dec 11 03:06:42 crc kubenswrapper[4824]: I1211 03:06:42.949991 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83x97sf_55c9ce15-2335-433b-a7ac-cfdf1a3ef599/util/0.log" Dec 11 03:06:42 crc kubenswrapper[4824]: I1211 03:06:42.958720 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83x97sf_55c9ce15-2335-433b-a7ac-cfdf1a3ef599/pull/0.log" Dec 11 03:06:42 crc kubenswrapper[4824]: I1211 03:06:42.991239 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83x97sf_55c9ce15-2335-433b-a7ac-cfdf1a3ef599/extract/0.log" Dec 11 03:06:43 crc kubenswrapper[4824]: I1211 03:06:43.120271 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-nqwwb_a61fd179-f3bd-4a79-ae61-d338e7994602/extract-utilities/0.log" Dec 11 03:06:43 crc kubenswrapper[4824]: I1211 03:06:43.286815 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-nqwwb_a61fd179-f3bd-4a79-ae61-d338e7994602/extract-utilities/0.log" Dec 11 03:06:43 crc kubenswrapper[4824]: I1211 03:06:43.311969 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-nqwwb_a61fd179-f3bd-4a79-ae61-d338e7994602/extract-content/0.log" Dec 11 03:06:43 crc kubenswrapper[4824]: I1211 03:06:43.326129 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-nqwwb_a61fd179-f3bd-4a79-ae61-d338e7994602/extract-content/0.log" Dec 11 03:06:43 crc kubenswrapper[4824]: I1211 03:06:43.492342 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-nqwwb_a61fd179-f3bd-4a79-ae61-d338e7994602/extract-utilities/0.log" Dec 11 03:06:43 crc kubenswrapper[4824]: I1211 03:06:43.520145 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-nqwwb_a61fd179-f3bd-4a79-ae61-d338e7994602/extract-content/0.log" Dec 11 03:06:43 crc kubenswrapper[4824]: I1211 03:06:43.713777 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4qn4g_d5e0df10-f888-47f3-87a9-889529a89473/extract-utilities/0.log" Dec 11 03:06:43 crc kubenswrapper[4824]: I1211 03:06:43.898091 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4qn4g_d5e0df10-f888-47f3-87a9-889529a89473/extract-utilities/0.log" Dec 11 03:06:43 crc kubenswrapper[4824]: I1211 03:06:43.944163 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-nqwwb_a61fd179-f3bd-4a79-ae61-d338e7994602/registry-server/0.log" Dec 11 03:06:43 crc kubenswrapper[4824]: I1211 03:06:43.966518 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4qn4g_d5e0df10-f888-47f3-87a9-889529a89473/extract-content/0.log" Dec 11 03:06:43 crc kubenswrapper[4824]: I1211 03:06:43.993026 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4qn4g_d5e0df10-f888-47f3-87a9-889529a89473/extract-content/0.log" Dec 11 03:06:44 crc kubenswrapper[4824]: I1211 03:06:44.150170 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4qn4g_d5e0df10-f888-47f3-87a9-889529a89473/extract-utilities/0.log" Dec 11 03:06:44 crc kubenswrapper[4824]: I1211 03:06:44.158055 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4qn4g_d5e0df10-f888-47f3-87a9-889529a89473/extract-content/0.log" Dec 11 03:06:44 crc kubenswrapper[4824]: I1211 03:06:44.431220 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-xj7qj_19528072-20f2-4741-bfd9-b7c44430b834/marketplace-operator/0.log" Dec 11 03:06:44 crc kubenswrapper[4824]: I1211 03:06:44.455711 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-k4msm_38e53ad7-42f0-4b59-a721-b5e38b6e1904/extract-utilities/0.log" Dec 11 03:06:44 crc kubenswrapper[4824]: I1211 03:06:44.642086 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4qn4g_d5e0df10-f888-47f3-87a9-889529a89473/registry-server/0.log" Dec 11 03:06:44 crc kubenswrapper[4824]: I1211 03:06:44.642830 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-k4msm_38e53ad7-42f0-4b59-a721-b5e38b6e1904/extract-content/0.log" Dec 11 03:06:44 crc kubenswrapper[4824]: I1211 03:06:44.691818 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-k4msm_38e53ad7-42f0-4b59-a721-b5e38b6e1904/extract-content/0.log" Dec 11 03:06:44 crc kubenswrapper[4824]: I1211 03:06:44.717762 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-k4msm_38e53ad7-42f0-4b59-a721-b5e38b6e1904/extract-utilities/0.log" Dec 11 03:06:44 crc kubenswrapper[4824]: I1211 03:06:44.978393 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-k4msm_38e53ad7-42f0-4b59-a721-b5e38b6e1904/extract-content/0.log" Dec 11 03:06:44 crc kubenswrapper[4824]: I1211 03:06:44.998015 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-k4msm_38e53ad7-42f0-4b59-a721-b5e38b6e1904/extract-utilities/0.log" Dec 11 03:06:45 crc kubenswrapper[4824]: I1211 03:06:45.116047 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-k4msm_38e53ad7-42f0-4b59-a721-b5e38b6e1904/registry-server/0.log" Dec 11 03:06:45 crc kubenswrapper[4824]: I1211 03:06:45.216188 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-wjmvj_351ca498-5ccd-4d80-a99f-3d56f7c5dbf2/extract-utilities/0.log" Dec 11 03:06:45 crc kubenswrapper[4824]: I1211 03:06:45.420786 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-wjmvj_351ca498-5ccd-4d80-a99f-3d56f7c5dbf2/extract-utilities/0.log" Dec 11 03:06:45 crc kubenswrapper[4824]: I1211 03:06:45.460747 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-wjmvj_351ca498-5ccd-4d80-a99f-3d56f7c5dbf2/extract-content/0.log" Dec 11 03:06:45 crc kubenswrapper[4824]: I1211 03:06:45.463632 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-wjmvj_351ca498-5ccd-4d80-a99f-3d56f7c5dbf2/extract-content/0.log" Dec 11 03:06:45 crc kubenswrapper[4824]: I1211 03:06:45.670175 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-wjmvj_351ca498-5ccd-4d80-a99f-3d56f7c5dbf2/extract-content/0.log" Dec 11 03:06:45 crc kubenswrapper[4824]: I1211 03:06:45.707470 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-wjmvj_351ca498-5ccd-4d80-a99f-3d56f7c5dbf2/extract-utilities/0.log" Dec 11 03:06:45 crc kubenswrapper[4824]: I1211 03:06:45.764539 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-wjmvj_351ca498-5ccd-4d80-a99f-3d56f7c5dbf2/registry-server/0.log" Dec 11 03:07:08 crc kubenswrapper[4824]: I1211 03:07:08.318164 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-jx665"] Dec 11 03:07:08 crc kubenswrapper[4824]: E1211 03:07:08.319305 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfc6dc48-ab52-4b16-83a4-798b7d75c41c" containerName="container-00" Dec 11 03:07:08 crc kubenswrapper[4824]: I1211 03:07:08.319322 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfc6dc48-ab52-4b16-83a4-798b7d75c41c" containerName="container-00" Dec 11 03:07:08 crc kubenswrapper[4824]: I1211 03:07:08.319588 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="bfc6dc48-ab52-4b16-83a4-798b7d75c41c" containerName="container-00" Dec 11 03:07:08 crc kubenswrapper[4824]: I1211 03:07:08.321326 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jx665" Dec 11 03:07:08 crc kubenswrapper[4824]: I1211 03:07:08.337439 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jx665"] Dec 11 03:07:08 crc kubenswrapper[4824]: I1211 03:07:08.480055 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-77jxp\" (UniqueName: \"kubernetes.io/projected/8e73a7e9-d1e9-4a39-9389-f10a95e3973d-kube-api-access-77jxp\") pod \"redhat-marketplace-jx665\" (UID: \"8e73a7e9-d1e9-4a39-9389-f10a95e3973d\") " pod="openshift-marketplace/redhat-marketplace-jx665" Dec 11 03:07:08 crc kubenswrapper[4824]: I1211 03:07:08.480207 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e73a7e9-d1e9-4a39-9389-f10a95e3973d-utilities\") pod \"redhat-marketplace-jx665\" (UID: \"8e73a7e9-d1e9-4a39-9389-f10a95e3973d\") " pod="openshift-marketplace/redhat-marketplace-jx665" Dec 11 03:07:08 crc kubenswrapper[4824]: I1211 03:07:08.480243 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e73a7e9-d1e9-4a39-9389-f10a95e3973d-catalog-content\") pod \"redhat-marketplace-jx665\" (UID: \"8e73a7e9-d1e9-4a39-9389-f10a95e3973d\") " pod="openshift-marketplace/redhat-marketplace-jx665" Dec 11 03:07:08 crc kubenswrapper[4824]: I1211 03:07:08.582394 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e73a7e9-d1e9-4a39-9389-f10a95e3973d-utilities\") pod \"redhat-marketplace-jx665\" (UID: \"8e73a7e9-d1e9-4a39-9389-f10a95e3973d\") " pod="openshift-marketplace/redhat-marketplace-jx665" Dec 11 03:07:08 crc kubenswrapper[4824]: I1211 03:07:08.582447 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e73a7e9-d1e9-4a39-9389-f10a95e3973d-catalog-content\") pod \"redhat-marketplace-jx665\" (UID: \"8e73a7e9-d1e9-4a39-9389-f10a95e3973d\") " pod="openshift-marketplace/redhat-marketplace-jx665" Dec 11 03:07:08 crc kubenswrapper[4824]: I1211 03:07:08.582490 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-77jxp\" (UniqueName: \"kubernetes.io/projected/8e73a7e9-d1e9-4a39-9389-f10a95e3973d-kube-api-access-77jxp\") pod \"redhat-marketplace-jx665\" (UID: \"8e73a7e9-d1e9-4a39-9389-f10a95e3973d\") " pod="openshift-marketplace/redhat-marketplace-jx665" Dec 11 03:07:08 crc kubenswrapper[4824]: I1211 03:07:08.583062 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e73a7e9-d1e9-4a39-9389-f10a95e3973d-utilities\") pod \"redhat-marketplace-jx665\" (UID: \"8e73a7e9-d1e9-4a39-9389-f10a95e3973d\") " pod="openshift-marketplace/redhat-marketplace-jx665" Dec 11 03:07:08 crc kubenswrapper[4824]: I1211 03:07:08.583097 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e73a7e9-d1e9-4a39-9389-f10a95e3973d-catalog-content\") pod \"redhat-marketplace-jx665\" (UID: \"8e73a7e9-d1e9-4a39-9389-f10a95e3973d\") " pod="openshift-marketplace/redhat-marketplace-jx665" Dec 11 03:07:08 crc kubenswrapper[4824]: I1211 03:07:08.602079 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-77jxp\" (UniqueName: \"kubernetes.io/projected/8e73a7e9-d1e9-4a39-9389-f10a95e3973d-kube-api-access-77jxp\") pod \"redhat-marketplace-jx665\" (UID: \"8e73a7e9-d1e9-4a39-9389-f10a95e3973d\") " pod="openshift-marketplace/redhat-marketplace-jx665" Dec 11 03:07:08 crc kubenswrapper[4824]: I1211 03:07:08.641549 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jx665" Dec 11 03:07:09 crc kubenswrapper[4824]: I1211 03:07:09.166241 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jx665"] Dec 11 03:07:09 crc kubenswrapper[4824]: I1211 03:07:09.244645 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jx665" event={"ID":"8e73a7e9-d1e9-4a39-9389-f10a95e3973d","Type":"ContainerStarted","Data":"34558eb998a2ca535b9fe6e0e69a18c5d271be1e3939b8fa25cb7589915cc849"} Dec 11 03:07:10 crc kubenswrapper[4824]: I1211 03:07:10.257367 4824 generic.go:334] "Generic (PLEG): container finished" podID="8e73a7e9-d1e9-4a39-9389-f10a95e3973d" containerID="be665c0c35210abc40adcc7a376ce549b18ad532b46af9e5a9146bcf9026924f" exitCode=0 Dec 11 03:07:10 crc kubenswrapper[4824]: I1211 03:07:10.257607 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jx665" event={"ID":"8e73a7e9-d1e9-4a39-9389-f10a95e3973d","Type":"ContainerDied","Data":"be665c0c35210abc40adcc7a376ce549b18ad532b46af9e5a9146bcf9026924f"} Dec 11 03:07:11 crc kubenswrapper[4824]: I1211 03:07:11.269525 4824 generic.go:334] "Generic (PLEG): container finished" podID="8e73a7e9-d1e9-4a39-9389-f10a95e3973d" containerID="b3dc4d263454514c521e5361f8aab554d69ecfb81b340da21d8cd4879a20dc61" exitCode=0 Dec 11 03:07:11 crc kubenswrapper[4824]: I1211 03:07:11.269710 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jx665" event={"ID":"8e73a7e9-d1e9-4a39-9389-f10a95e3973d","Type":"ContainerDied","Data":"b3dc4d263454514c521e5361f8aab554d69ecfb81b340da21d8cd4879a20dc61"} Dec 11 03:07:12 crc kubenswrapper[4824]: I1211 03:07:12.294158 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jx665" event={"ID":"8e73a7e9-d1e9-4a39-9389-f10a95e3973d","Type":"ContainerStarted","Data":"e55bcc8e5bd9fdef4c0baf287e0ba6684f5442a91db95b281bf2ea1d8342cb28"} Dec 11 03:07:12 crc kubenswrapper[4824]: I1211 03:07:12.332946 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-jx665" podStartSLOduration=2.7876299209999997 podStartE2EDuration="4.332924447s" podCreationTimestamp="2025-12-11 03:07:08 +0000 UTC" firstStartedPulling="2025-12-11 03:07:10.260736142 +0000 UTC m=+3971.949773531" lastFinishedPulling="2025-12-11 03:07:11.806030688 +0000 UTC m=+3973.495068057" observedRunningTime="2025-12-11 03:07:12.322757653 +0000 UTC m=+3974.011795032" watchObservedRunningTime="2025-12-11 03:07:12.332924447 +0000 UTC m=+3974.021961826" Dec 11 03:07:18 crc kubenswrapper[4824]: I1211 03:07:18.648328 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-jx665" Dec 11 03:07:18 crc kubenswrapper[4824]: I1211 03:07:18.648927 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-jx665" Dec 11 03:07:18 crc kubenswrapper[4824]: I1211 03:07:18.699344 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-jx665" Dec 11 03:07:19 crc kubenswrapper[4824]: I1211 03:07:19.446796 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-jx665" Dec 11 03:07:19 crc kubenswrapper[4824]: I1211 03:07:19.513968 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jx665"] Dec 11 03:07:21 crc kubenswrapper[4824]: I1211 03:07:21.378424 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-jx665" podUID="8e73a7e9-d1e9-4a39-9389-f10a95e3973d" containerName="registry-server" containerID="cri-o://e55bcc8e5bd9fdef4c0baf287e0ba6684f5442a91db95b281bf2ea1d8342cb28" gracePeriod=2 Dec 11 03:07:22 crc kubenswrapper[4824]: I1211 03:07:22.056654 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jx665" Dec 11 03:07:22 crc kubenswrapper[4824]: I1211 03:07:22.131525 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e73a7e9-d1e9-4a39-9389-f10a95e3973d-catalog-content\") pod \"8e73a7e9-d1e9-4a39-9389-f10a95e3973d\" (UID: \"8e73a7e9-d1e9-4a39-9389-f10a95e3973d\") " Dec 11 03:07:22 crc kubenswrapper[4824]: I1211 03:07:22.132092 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-77jxp\" (UniqueName: \"kubernetes.io/projected/8e73a7e9-d1e9-4a39-9389-f10a95e3973d-kube-api-access-77jxp\") pod \"8e73a7e9-d1e9-4a39-9389-f10a95e3973d\" (UID: \"8e73a7e9-d1e9-4a39-9389-f10a95e3973d\") " Dec 11 03:07:22 crc kubenswrapper[4824]: I1211 03:07:22.132242 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e73a7e9-d1e9-4a39-9389-f10a95e3973d-utilities\") pod \"8e73a7e9-d1e9-4a39-9389-f10a95e3973d\" (UID: \"8e73a7e9-d1e9-4a39-9389-f10a95e3973d\") " Dec 11 03:07:22 crc kubenswrapper[4824]: I1211 03:07:22.133026 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e73a7e9-d1e9-4a39-9389-f10a95e3973d-utilities" (OuterVolumeSpecName: "utilities") pod "8e73a7e9-d1e9-4a39-9389-f10a95e3973d" (UID: "8e73a7e9-d1e9-4a39-9389-f10a95e3973d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 03:07:22 crc kubenswrapper[4824]: I1211 03:07:22.137678 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e73a7e9-d1e9-4a39-9389-f10a95e3973d-kube-api-access-77jxp" (OuterVolumeSpecName: "kube-api-access-77jxp") pod "8e73a7e9-d1e9-4a39-9389-f10a95e3973d" (UID: "8e73a7e9-d1e9-4a39-9389-f10a95e3973d"). InnerVolumeSpecName "kube-api-access-77jxp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 03:07:22 crc kubenswrapper[4824]: I1211 03:07:22.150707 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e73a7e9-d1e9-4a39-9389-f10a95e3973d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8e73a7e9-d1e9-4a39-9389-f10a95e3973d" (UID: "8e73a7e9-d1e9-4a39-9389-f10a95e3973d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 03:07:22 crc kubenswrapper[4824]: I1211 03:07:22.234834 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-77jxp\" (UniqueName: \"kubernetes.io/projected/8e73a7e9-d1e9-4a39-9389-f10a95e3973d-kube-api-access-77jxp\") on node \"crc\" DevicePath \"\"" Dec 11 03:07:22 crc kubenswrapper[4824]: I1211 03:07:22.234862 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e73a7e9-d1e9-4a39-9389-f10a95e3973d-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 03:07:22 crc kubenswrapper[4824]: I1211 03:07:22.234872 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e73a7e9-d1e9-4a39-9389-f10a95e3973d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 03:07:22 crc kubenswrapper[4824]: I1211 03:07:22.393016 4824 generic.go:334] "Generic (PLEG): container finished" podID="8e73a7e9-d1e9-4a39-9389-f10a95e3973d" containerID="e55bcc8e5bd9fdef4c0baf287e0ba6684f5442a91db95b281bf2ea1d8342cb28" exitCode=0 Dec 11 03:07:22 crc kubenswrapper[4824]: I1211 03:07:22.393075 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jx665" event={"ID":"8e73a7e9-d1e9-4a39-9389-f10a95e3973d","Type":"ContainerDied","Data":"e55bcc8e5bd9fdef4c0baf287e0ba6684f5442a91db95b281bf2ea1d8342cb28"} Dec 11 03:07:22 crc kubenswrapper[4824]: I1211 03:07:22.393122 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jx665" Dec 11 03:07:22 crc kubenswrapper[4824]: I1211 03:07:22.393163 4824 scope.go:117] "RemoveContainer" containerID="e55bcc8e5bd9fdef4c0baf287e0ba6684f5442a91db95b281bf2ea1d8342cb28" Dec 11 03:07:22 crc kubenswrapper[4824]: I1211 03:07:22.393129 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jx665" event={"ID":"8e73a7e9-d1e9-4a39-9389-f10a95e3973d","Type":"ContainerDied","Data":"34558eb998a2ca535b9fe6e0e69a18c5d271be1e3939b8fa25cb7589915cc849"} Dec 11 03:07:22 crc kubenswrapper[4824]: I1211 03:07:22.435700 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jx665"] Dec 11 03:07:22 crc kubenswrapper[4824]: I1211 03:07:22.438053 4824 scope.go:117] "RemoveContainer" containerID="b3dc4d263454514c521e5361f8aab554d69ecfb81b340da21d8cd4879a20dc61" Dec 11 03:07:22 crc kubenswrapper[4824]: I1211 03:07:22.448874 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-jx665"] Dec 11 03:07:22 crc kubenswrapper[4824]: I1211 03:07:22.466083 4824 scope.go:117] "RemoveContainer" containerID="be665c0c35210abc40adcc7a376ce549b18ad532b46af9e5a9146bcf9026924f" Dec 11 03:07:22 crc kubenswrapper[4824]: I1211 03:07:22.525784 4824 scope.go:117] "RemoveContainer" containerID="e55bcc8e5bd9fdef4c0baf287e0ba6684f5442a91db95b281bf2ea1d8342cb28" Dec 11 03:07:22 crc kubenswrapper[4824]: E1211 03:07:22.526458 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e55bcc8e5bd9fdef4c0baf287e0ba6684f5442a91db95b281bf2ea1d8342cb28\": container with ID starting with e55bcc8e5bd9fdef4c0baf287e0ba6684f5442a91db95b281bf2ea1d8342cb28 not found: ID does not exist" containerID="e55bcc8e5bd9fdef4c0baf287e0ba6684f5442a91db95b281bf2ea1d8342cb28" Dec 11 03:07:22 crc kubenswrapper[4824]: I1211 03:07:22.526576 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e55bcc8e5bd9fdef4c0baf287e0ba6684f5442a91db95b281bf2ea1d8342cb28"} err="failed to get container status \"e55bcc8e5bd9fdef4c0baf287e0ba6684f5442a91db95b281bf2ea1d8342cb28\": rpc error: code = NotFound desc = could not find container \"e55bcc8e5bd9fdef4c0baf287e0ba6684f5442a91db95b281bf2ea1d8342cb28\": container with ID starting with e55bcc8e5bd9fdef4c0baf287e0ba6684f5442a91db95b281bf2ea1d8342cb28 not found: ID does not exist" Dec 11 03:07:22 crc kubenswrapper[4824]: I1211 03:07:22.526657 4824 scope.go:117] "RemoveContainer" containerID="b3dc4d263454514c521e5361f8aab554d69ecfb81b340da21d8cd4879a20dc61" Dec 11 03:07:22 crc kubenswrapper[4824]: E1211 03:07:22.527105 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b3dc4d263454514c521e5361f8aab554d69ecfb81b340da21d8cd4879a20dc61\": container with ID starting with b3dc4d263454514c521e5361f8aab554d69ecfb81b340da21d8cd4879a20dc61 not found: ID does not exist" containerID="b3dc4d263454514c521e5361f8aab554d69ecfb81b340da21d8cd4879a20dc61" Dec 11 03:07:22 crc kubenswrapper[4824]: I1211 03:07:22.527202 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3dc4d263454514c521e5361f8aab554d69ecfb81b340da21d8cd4879a20dc61"} err="failed to get container status \"b3dc4d263454514c521e5361f8aab554d69ecfb81b340da21d8cd4879a20dc61\": rpc error: code = NotFound desc = could not find container \"b3dc4d263454514c521e5361f8aab554d69ecfb81b340da21d8cd4879a20dc61\": container with ID starting with b3dc4d263454514c521e5361f8aab554d69ecfb81b340da21d8cd4879a20dc61 not found: ID does not exist" Dec 11 03:07:22 crc kubenswrapper[4824]: I1211 03:07:22.527270 4824 scope.go:117] "RemoveContainer" containerID="be665c0c35210abc40adcc7a376ce549b18ad532b46af9e5a9146bcf9026924f" Dec 11 03:07:22 crc kubenswrapper[4824]: E1211 03:07:22.527707 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"be665c0c35210abc40adcc7a376ce549b18ad532b46af9e5a9146bcf9026924f\": container with ID starting with be665c0c35210abc40adcc7a376ce549b18ad532b46af9e5a9146bcf9026924f not found: ID does not exist" containerID="be665c0c35210abc40adcc7a376ce549b18ad532b46af9e5a9146bcf9026924f" Dec 11 03:07:22 crc kubenswrapper[4824]: I1211 03:07:22.527800 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be665c0c35210abc40adcc7a376ce549b18ad532b46af9e5a9146bcf9026924f"} err="failed to get container status \"be665c0c35210abc40adcc7a376ce549b18ad532b46af9e5a9146bcf9026924f\": rpc error: code = NotFound desc = could not find container \"be665c0c35210abc40adcc7a376ce549b18ad532b46af9e5a9146bcf9026924f\": container with ID starting with be665c0c35210abc40adcc7a376ce549b18ad532b46af9e5a9146bcf9026924f not found: ID does not exist" Dec 11 03:07:22 crc kubenswrapper[4824]: I1211 03:07:22.645058 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e73a7e9-d1e9-4a39-9389-f10a95e3973d" path="/var/lib/kubelet/pods/8e73a7e9-d1e9-4a39-9389-f10a95e3973d/volumes" Dec 11 03:07:34 crc kubenswrapper[4824]: I1211 03:07:34.747725 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-lb22k"] Dec 11 03:07:34 crc kubenswrapper[4824]: E1211 03:07:34.748938 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e73a7e9-d1e9-4a39-9389-f10a95e3973d" containerName="registry-server" Dec 11 03:07:34 crc kubenswrapper[4824]: I1211 03:07:34.748960 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e73a7e9-d1e9-4a39-9389-f10a95e3973d" containerName="registry-server" Dec 11 03:07:34 crc kubenswrapper[4824]: E1211 03:07:34.748990 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e73a7e9-d1e9-4a39-9389-f10a95e3973d" containerName="extract-utilities" Dec 11 03:07:34 crc kubenswrapper[4824]: I1211 03:07:34.749001 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e73a7e9-d1e9-4a39-9389-f10a95e3973d" containerName="extract-utilities" Dec 11 03:07:34 crc kubenswrapper[4824]: E1211 03:07:34.749065 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e73a7e9-d1e9-4a39-9389-f10a95e3973d" containerName="extract-content" Dec 11 03:07:34 crc kubenswrapper[4824]: I1211 03:07:34.749077 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e73a7e9-d1e9-4a39-9389-f10a95e3973d" containerName="extract-content" Dec 11 03:07:34 crc kubenswrapper[4824]: I1211 03:07:34.749411 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e73a7e9-d1e9-4a39-9389-f10a95e3973d" containerName="registry-server" Dec 11 03:07:34 crc kubenswrapper[4824]: I1211 03:07:34.751584 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lb22k" Dec 11 03:07:34 crc kubenswrapper[4824]: I1211 03:07:34.778757 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lb22k"] Dec 11 03:07:34 crc kubenswrapper[4824]: I1211 03:07:34.919533 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6de8b06d-f30a-49e8-93ff-d1fb9b419014-utilities\") pod \"community-operators-lb22k\" (UID: \"6de8b06d-f30a-49e8-93ff-d1fb9b419014\") " pod="openshift-marketplace/community-operators-lb22k" Dec 11 03:07:34 crc kubenswrapper[4824]: I1211 03:07:34.920602 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6de8b06d-f30a-49e8-93ff-d1fb9b419014-catalog-content\") pod \"community-operators-lb22k\" (UID: \"6de8b06d-f30a-49e8-93ff-d1fb9b419014\") " pod="openshift-marketplace/community-operators-lb22k" Dec 11 03:07:34 crc kubenswrapper[4824]: I1211 03:07:34.920854 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ghj9p\" (UniqueName: \"kubernetes.io/projected/6de8b06d-f30a-49e8-93ff-d1fb9b419014-kube-api-access-ghj9p\") pod \"community-operators-lb22k\" (UID: \"6de8b06d-f30a-49e8-93ff-d1fb9b419014\") " pod="openshift-marketplace/community-operators-lb22k" Dec 11 03:07:35 crc kubenswrapper[4824]: I1211 03:07:35.023172 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6de8b06d-f30a-49e8-93ff-d1fb9b419014-utilities\") pod \"community-operators-lb22k\" (UID: \"6de8b06d-f30a-49e8-93ff-d1fb9b419014\") " pod="openshift-marketplace/community-operators-lb22k" Dec 11 03:07:35 crc kubenswrapper[4824]: I1211 03:07:35.023309 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6de8b06d-f30a-49e8-93ff-d1fb9b419014-catalog-content\") pod \"community-operators-lb22k\" (UID: \"6de8b06d-f30a-49e8-93ff-d1fb9b419014\") " pod="openshift-marketplace/community-operators-lb22k" Dec 11 03:07:35 crc kubenswrapper[4824]: I1211 03:07:35.023343 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ghj9p\" (UniqueName: \"kubernetes.io/projected/6de8b06d-f30a-49e8-93ff-d1fb9b419014-kube-api-access-ghj9p\") pod \"community-operators-lb22k\" (UID: \"6de8b06d-f30a-49e8-93ff-d1fb9b419014\") " pod="openshift-marketplace/community-operators-lb22k" Dec 11 03:07:35 crc kubenswrapper[4824]: I1211 03:07:35.024061 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6de8b06d-f30a-49e8-93ff-d1fb9b419014-utilities\") pod \"community-operators-lb22k\" (UID: \"6de8b06d-f30a-49e8-93ff-d1fb9b419014\") " pod="openshift-marketplace/community-operators-lb22k" Dec 11 03:07:35 crc kubenswrapper[4824]: I1211 03:07:35.024296 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6de8b06d-f30a-49e8-93ff-d1fb9b419014-catalog-content\") pod \"community-operators-lb22k\" (UID: \"6de8b06d-f30a-49e8-93ff-d1fb9b419014\") " pod="openshift-marketplace/community-operators-lb22k" Dec 11 03:07:35 crc kubenswrapper[4824]: I1211 03:07:35.042864 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ghj9p\" (UniqueName: \"kubernetes.io/projected/6de8b06d-f30a-49e8-93ff-d1fb9b419014-kube-api-access-ghj9p\") pod \"community-operators-lb22k\" (UID: \"6de8b06d-f30a-49e8-93ff-d1fb9b419014\") " pod="openshift-marketplace/community-operators-lb22k" Dec 11 03:07:35 crc kubenswrapper[4824]: I1211 03:07:35.092030 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lb22k" Dec 11 03:07:35 crc kubenswrapper[4824]: I1211 03:07:35.623206 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lb22k"] Dec 11 03:07:36 crc kubenswrapper[4824]: I1211 03:07:36.575613 4824 generic.go:334] "Generic (PLEG): container finished" podID="6de8b06d-f30a-49e8-93ff-d1fb9b419014" containerID="20768074b6233713d1cfaa63359c40865573a337e47c6b1b274e6aa72afd6c6f" exitCode=0 Dec 11 03:07:36 crc kubenswrapper[4824]: I1211 03:07:36.575734 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lb22k" event={"ID":"6de8b06d-f30a-49e8-93ff-d1fb9b419014","Type":"ContainerDied","Data":"20768074b6233713d1cfaa63359c40865573a337e47c6b1b274e6aa72afd6c6f"} Dec 11 03:07:36 crc kubenswrapper[4824]: I1211 03:07:36.576196 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lb22k" event={"ID":"6de8b06d-f30a-49e8-93ff-d1fb9b419014","Type":"ContainerStarted","Data":"64a746046f0a65da3a628027109d1143671c48978eef16e9621f0b2c281c05b3"} Dec 11 03:07:37 crc kubenswrapper[4824]: I1211 03:07:37.590038 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lb22k" event={"ID":"6de8b06d-f30a-49e8-93ff-d1fb9b419014","Type":"ContainerStarted","Data":"2fca370cb45f449f91acf3a4f0d92f7ad72800a5102cbeffd57eb84daa0e9020"} Dec 11 03:07:38 crc kubenswrapper[4824]: I1211 03:07:38.608341 4824 generic.go:334] "Generic (PLEG): container finished" podID="6de8b06d-f30a-49e8-93ff-d1fb9b419014" containerID="2fca370cb45f449f91acf3a4f0d92f7ad72800a5102cbeffd57eb84daa0e9020" exitCode=0 Dec 11 03:07:38 crc kubenswrapper[4824]: I1211 03:07:38.608741 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lb22k" event={"ID":"6de8b06d-f30a-49e8-93ff-d1fb9b419014","Type":"ContainerDied","Data":"2fca370cb45f449f91acf3a4f0d92f7ad72800a5102cbeffd57eb84daa0e9020"} Dec 11 03:07:38 crc kubenswrapper[4824]: I1211 03:07:38.611359 4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 11 03:07:39 crc kubenswrapper[4824]: I1211 03:07:39.621826 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lb22k" event={"ID":"6de8b06d-f30a-49e8-93ff-d1fb9b419014","Type":"ContainerStarted","Data":"f781567cf4f23cdd084449181a383e991186989bba409172879c03e58b60efff"} Dec 11 03:07:39 crc kubenswrapper[4824]: I1211 03:07:39.659295 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-lb22k" podStartSLOduration=3.119660201 podStartE2EDuration="5.659272783s" podCreationTimestamp="2025-12-11 03:07:34 +0000 UTC" firstStartedPulling="2025-12-11 03:07:36.577880091 +0000 UTC m=+3998.266917510" lastFinishedPulling="2025-12-11 03:07:39.117492703 +0000 UTC m=+4000.806530092" observedRunningTime="2025-12-11 03:07:39.652333526 +0000 UTC m=+4001.341370945" watchObservedRunningTime="2025-12-11 03:07:39.659272783 +0000 UTC m=+4001.348310172" Dec 11 03:07:45 crc kubenswrapper[4824]: I1211 03:07:45.093295 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-lb22k" Dec 11 03:07:45 crc kubenswrapper[4824]: I1211 03:07:45.094565 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-lb22k" Dec 11 03:07:45 crc kubenswrapper[4824]: I1211 03:07:45.161184 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-lb22k" Dec 11 03:07:45 crc kubenswrapper[4824]: I1211 03:07:45.795708 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-lb22k" Dec 11 03:07:45 crc kubenswrapper[4824]: I1211 03:07:45.873427 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lb22k"] Dec 11 03:07:47 crc kubenswrapper[4824]: I1211 03:07:47.731080 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-lb22k" podUID="6de8b06d-f30a-49e8-93ff-d1fb9b419014" containerName="registry-server" containerID="cri-o://f781567cf4f23cdd084449181a383e991186989bba409172879c03e58b60efff" gracePeriod=2 Dec 11 03:07:48 crc kubenswrapper[4824]: I1211 03:07:48.388766 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lb22k" Dec 11 03:07:48 crc kubenswrapper[4824]: I1211 03:07:48.549897 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6de8b06d-f30a-49e8-93ff-d1fb9b419014-utilities\") pod \"6de8b06d-f30a-49e8-93ff-d1fb9b419014\" (UID: \"6de8b06d-f30a-49e8-93ff-d1fb9b419014\") " Dec 11 03:07:48 crc kubenswrapper[4824]: I1211 03:07:48.550069 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ghj9p\" (UniqueName: \"kubernetes.io/projected/6de8b06d-f30a-49e8-93ff-d1fb9b419014-kube-api-access-ghj9p\") pod \"6de8b06d-f30a-49e8-93ff-d1fb9b419014\" (UID: \"6de8b06d-f30a-49e8-93ff-d1fb9b419014\") " Dec 11 03:07:48 crc kubenswrapper[4824]: I1211 03:07:48.550146 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6de8b06d-f30a-49e8-93ff-d1fb9b419014-catalog-content\") pod \"6de8b06d-f30a-49e8-93ff-d1fb9b419014\" (UID: \"6de8b06d-f30a-49e8-93ff-d1fb9b419014\") " Dec 11 03:07:48 crc kubenswrapper[4824]: I1211 03:07:48.552936 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6de8b06d-f30a-49e8-93ff-d1fb9b419014-utilities" (OuterVolumeSpecName: "utilities") pod "6de8b06d-f30a-49e8-93ff-d1fb9b419014" (UID: "6de8b06d-f30a-49e8-93ff-d1fb9b419014"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 03:07:48 crc kubenswrapper[4824]: I1211 03:07:48.564254 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6de8b06d-f30a-49e8-93ff-d1fb9b419014-kube-api-access-ghj9p" (OuterVolumeSpecName: "kube-api-access-ghj9p") pod "6de8b06d-f30a-49e8-93ff-d1fb9b419014" (UID: "6de8b06d-f30a-49e8-93ff-d1fb9b419014"). InnerVolumeSpecName "kube-api-access-ghj9p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 03:07:48 crc kubenswrapper[4824]: I1211 03:07:48.643841 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6de8b06d-f30a-49e8-93ff-d1fb9b419014-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6de8b06d-f30a-49e8-93ff-d1fb9b419014" (UID: "6de8b06d-f30a-49e8-93ff-d1fb9b419014"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 03:07:48 crc kubenswrapper[4824]: I1211 03:07:48.653188 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6de8b06d-f30a-49e8-93ff-d1fb9b419014-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 03:07:48 crc kubenswrapper[4824]: I1211 03:07:48.653221 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ghj9p\" (UniqueName: \"kubernetes.io/projected/6de8b06d-f30a-49e8-93ff-d1fb9b419014-kube-api-access-ghj9p\") on node \"crc\" DevicePath \"\"" Dec 11 03:07:48 crc kubenswrapper[4824]: I1211 03:07:48.653237 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6de8b06d-f30a-49e8-93ff-d1fb9b419014-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 03:07:48 crc kubenswrapper[4824]: I1211 03:07:48.742498 4824 generic.go:334] "Generic (PLEG): container finished" podID="6de8b06d-f30a-49e8-93ff-d1fb9b419014" containerID="f781567cf4f23cdd084449181a383e991186989bba409172879c03e58b60efff" exitCode=0 Dec 11 03:07:48 crc kubenswrapper[4824]: I1211 03:07:48.742560 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lb22k" event={"ID":"6de8b06d-f30a-49e8-93ff-d1fb9b419014","Type":"ContainerDied","Data":"f781567cf4f23cdd084449181a383e991186989bba409172879c03e58b60efff"} Dec 11 03:07:48 crc kubenswrapper[4824]: I1211 03:07:48.742602 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lb22k" Dec 11 03:07:48 crc kubenswrapper[4824]: I1211 03:07:48.742637 4824 scope.go:117] "RemoveContainer" containerID="f781567cf4f23cdd084449181a383e991186989bba409172879c03e58b60efff" Dec 11 03:07:48 crc kubenswrapper[4824]: I1211 03:07:48.742621 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lb22k" event={"ID":"6de8b06d-f30a-49e8-93ff-d1fb9b419014","Type":"ContainerDied","Data":"64a746046f0a65da3a628027109d1143671c48978eef16e9621f0b2c281c05b3"} Dec 11 03:07:48 crc kubenswrapper[4824]: I1211 03:07:48.774734 4824 scope.go:117] "RemoveContainer" containerID="2fca370cb45f449f91acf3a4f0d92f7ad72800a5102cbeffd57eb84daa0e9020" Dec 11 03:07:48 crc kubenswrapper[4824]: I1211 03:07:48.777764 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lb22k"] Dec 11 03:07:48 crc kubenswrapper[4824]: I1211 03:07:48.785712 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-lb22k"] Dec 11 03:07:48 crc kubenswrapper[4824]: I1211 03:07:48.810212 4824 scope.go:117] "RemoveContainer" containerID="20768074b6233713d1cfaa63359c40865573a337e47c6b1b274e6aa72afd6c6f" Dec 11 03:07:48 crc kubenswrapper[4824]: I1211 03:07:48.868679 4824 scope.go:117] "RemoveContainer" containerID="f781567cf4f23cdd084449181a383e991186989bba409172879c03e58b60efff" Dec 11 03:07:48 crc kubenswrapper[4824]: E1211 03:07:48.870070 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f781567cf4f23cdd084449181a383e991186989bba409172879c03e58b60efff\": container with ID starting with f781567cf4f23cdd084449181a383e991186989bba409172879c03e58b60efff not found: ID does not exist" containerID="f781567cf4f23cdd084449181a383e991186989bba409172879c03e58b60efff" Dec 11 03:07:48 crc kubenswrapper[4824]: I1211 03:07:48.870119 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f781567cf4f23cdd084449181a383e991186989bba409172879c03e58b60efff"} err="failed to get container status \"f781567cf4f23cdd084449181a383e991186989bba409172879c03e58b60efff\": rpc error: code = NotFound desc = could not find container \"f781567cf4f23cdd084449181a383e991186989bba409172879c03e58b60efff\": container with ID starting with f781567cf4f23cdd084449181a383e991186989bba409172879c03e58b60efff not found: ID does not exist" Dec 11 03:07:48 crc kubenswrapper[4824]: I1211 03:07:48.870140 4824 scope.go:117] "RemoveContainer" containerID="2fca370cb45f449f91acf3a4f0d92f7ad72800a5102cbeffd57eb84daa0e9020" Dec 11 03:07:48 crc kubenswrapper[4824]: E1211 03:07:48.870836 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2fca370cb45f449f91acf3a4f0d92f7ad72800a5102cbeffd57eb84daa0e9020\": container with ID starting with 2fca370cb45f449f91acf3a4f0d92f7ad72800a5102cbeffd57eb84daa0e9020 not found: ID does not exist" containerID="2fca370cb45f449f91acf3a4f0d92f7ad72800a5102cbeffd57eb84daa0e9020" Dec 11 03:07:48 crc kubenswrapper[4824]: I1211 03:07:48.870861 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2fca370cb45f449f91acf3a4f0d92f7ad72800a5102cbeffd57eb84daa0e9020"} err="failed to get container status \"2fca370cb45f449f91acf3a4f0d92f7ad72800a5102cbeffd57eb84daa0e9020\": rpc error: code = NotFound desc = could not find container \"2fca370cb45f449f91acf3a4f0d92f7ad72800a5102cbeffd57eb84daa0e9020\": container with ID starting with 2fca370cb45f449f91acf3a4f0d92f7ad72800a5102cbeffd57eb84daa0e9020 not found: ID does not exist" Dec 11 03:07:48 crc kubenswrapper[4824]: I1211 03:07:48.870875 4824 scope.go:117] "RemoveContainer" containerID="20768074b6233713d1cfaa63359c40865573a337e47c6b1b274e6aa72afd6c6f" Dec 11 03:07:48 crc kubenswrapper[4824]: E1211 03:07:48.871131 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"20768074b6233713d1cfaa63359c40865573a337e47c6b1b274e6aa72afd6c6f\": container with ID starting with 20768074b6233713d1cfaa63359c40865573a337e47c6b1b274e6aa72afd6c6f not found: ID does not exist" containerID="20768074b6233713d1cfaa63359c40865573a337e47c6b1b274e6aa72afd6c6f" Dec 11 03:07:48 crc kubenswrapper[4824]: I1211 03:07:48.871161 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20768074b6233713d1cfaa63359c40865573a337e47c6b1b274e6aa72afd6c6f"} err="failed to get container status \"20768074b6233713d1cfaa63359c40865573a337e47c6b1b274e6aa72afd6c6f\": rpc error: code = NotFound desc = could not find container \"20768074b6233713d1cfaa63359c40865573a337e47c6b1b274e6aa72afd6c6f\": container with ID starting with 20768074b6233713d1cfaa63359c40865573a337e47c6b1b274e6aa72afd6c6f not found: ID does not exist" Dec 11 03:07:50 crc kubenswrapper[4824]: I1211 03:07:50.670388 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6de8b06d-f30a-49e8-93ff-d1fb9b419014" path="/var/lib/kubelet/pods/6de8b06d-f30a-49e8-93ff-d1fb9b419014/volumes" Dec 11 03:08:07 crc kubenswrapper[4824]: I1211 03:08:07.251636 4824 patch_prober.go:28] interesting pod/machine-config-daemon-gx6xt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 03:08:07 crc kubenswrapper[4824]: I1211 03:08:07.252280 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 03:08:23 crc kubenswrapper[4824]: I1211 03:08:23.127783 4824 generic.go:334] "Generic (PLEG): container finished" podID="bb60c05d-1bda-4726-a2f2-5967992fd930" containerID="899e0d503155df37fa14e0dd1b38b28b8f72e830e29c0e83ba0b27f26c27dc3c" exitCode=0 Dec 11 03:08:23 crc kubenswrapper[4824]: I1211 03:08:23.127926 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-v65vx/must-gather-rwnp8" event={"ID":"bb60c05d-1bda-4726-a2f2-5967992fd930","Type":"ContainerDied","Data":"899e0d503155df37fa14e0dd1b38b28b8f72e830e29c0e83ba0b27f26c27dc3c"} Dec 11 03:08:23 crc kubenswrapper[4824]: I1211 03:08:23.128847 4824 scope.go:117] "RemoveContainer" containerID="899e0d503155df37fa14e0dd1b38b28b8f72e830e29c0e83ba0b27f26c27dc3c" Dec 11 03:08:24 crc kubenswrapper[4824]: I1211 03:08:24.121960 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-v65vx_must-gather-rwnp8_bb60c05d-1bda-4726-a2f2-5967992fd930/gather/0.log" Dec 11 03:08:30 crc kubenswrapper[4824]: I1211 03:08:30.795971 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-9fzhp"] Dec 11 03:08:30 crc kubenswrapper[4824]: E1211 03:08:30.797312 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6de8b06d-f30a-49e8-93ff-d1fb9b419014" containerName="registry-server" Dec 11 03:08:30 crc kubenswrapper[4824]: I1211 03:08:30.797333 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="6de8b06d-f30a-49e8-93ff-d1fb9b419014" containerName="registry-server" Dec 11 03:08:30 crc kubenswrapper[4824]: E1211 03:08:30.797368 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6de8b06d-f30a-49e8-93ff-d1fb9b419014" containerName="extract-utilities" Dec 11 03:08:30 crc kubenswrapper[4824]: I1211 03:08:30.797377 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="6de8b06d-f30a-49e8-93ff-d1fb9b419014" containerName="extract-utilities" Dec 11 03:08:30 crc kubenswrapper[4824]: E1211 03:08:30.797388 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6de8b06d-f30a-49e8-93ff-d1fb9b419014" containerName="extract-content" Dec 11 03:08:30 crc kubenswrapper[4824]: I1211 03:08:30.797395 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="6de8b06d-f30a-49e8-93ff-d1fb9b419014" containerName="extract-content" Dec 11 03:08:30 crc kubenswrapper[4824]: I1211 03:08:30.797651 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="6de8b06d-f30a-49e8-93ff-d1fb9b419014" containerName="registry-server" Dec 11 03:08:30 crc kubenswrapper[4824]: I1211 03:08:30.799586 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9fzhp" Dec 11 03:08:30 crc kubenswrapper[4824]: I1211 03:08:30.828314 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9fzhp"] Dec 11 03:08:30 crc kubenswrapper[4824]: I1211 03:08:30.902662 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/224f20e3-2b4a-4e0c-bbad-906eb92a8ce9-utilities\") pod \"redhat-operators-9fzhp\" (UID: \"224f20e3-2b4a-4e0c-bbad-906eb92a8ce9\") " pod="openshift-marketplace/redhat-operators-9fzhp" Dec 11 03:08:30 crc kubenswrapper[4824]: I1211 03:08:30.903453 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sswrn\" (UniqueName: \"kubernetes.io/projected/224f20e3-2b4a-4e0c-bbad-906eb92a8ce9-kube-api-access-sswrn\") pod \"redhat-operators-9fzhp\" (UID: \"224f20e3-2b4a-4e0c-bbad-906eb92a8ce9\") " pod="openshift-marketplace/redhat-operators-9fzhp" Dec 11 03:08:30 crc kubenswrapper[4824]: I1211 03:08:30.903557 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/224f20e3-2b4a-4e0c-bbad-906eb92a8ce9-catalog-content\") pod \"redhat-operators-9fzhp\" (UID: \"224f20e3-2b4a-4e0c-bbad-906eb92a8ce9\") " pod="openshift-marketplace/redhat-operators-9fzhp" Dec 11 03:08:31 crc kubenswrapper[4824]: I1211 03:08:31.005118 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/224f20e3-2b4a-4e0c-bbad-906eb92a8ce9-utilities\") pod \"redhat-operators-9fzhp\" (UID: \"224f20e3-2b4a-4e0c-bbad-906eb92a8ce9\") " pod="openshift-marketplace/redhat-operators-9fzhp" Dec 11 03:08:31 crc kubenswrapper[4824]: I1211 03:08:31.005169 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sswrn\" (UniqueName: \"kubernetes.io/projected/224f20e3-2b4a-4e0c-bbad-906eb92a8ce9-kube-api-access-sswrn\") pod \"redhat-operators-9fzhp\" (UID: \"224f20e3-2b4a-4e0c-bbad-906eb92a8ce9\") " pod="openshift-marketplace/redhat-operators-9fzhp" Dec 11 03:08:31 crc kubenswrapper[4824]: I1211 03:08:31.005242 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/224f20e3-2b4a-4e0c-bbad-906eb92a8ce9-catalog-content\") pod \"redhat-operators-9fzhp\" (UID: \"224f20e3-2b4a-4e0c-bbad-906eb92a8ce9\") " pod="openshift-marketplace/redhat-operators-9fzhp" Dec 11 03:08:31 crc kubenswrapper[4824]: I1211 03:08:31.005676 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/224f20e3-2b4a-4e0c-bbad-906eb92a8ce9-utilities\") pod \"redhat-operators-9fzhp\" (UID: \"224f20e3-2b4a-4e0c-bbad-906eb92a8ce9\") " pod="openshift-marketplace/redhat-operators-9fzhp" Dec 11 03:08:31 crc kubenswrapper[4824]: I1211 03:08:31.005695 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/224f20e3-2b4a-4e0c-bbad-906eb92a8ce9-catalog-content\") pod \"redhat-operators-9fzhp\" (UID: \"224f20e3-2b4a-4e0c-bbad-906eb92a8ce9\") " pod="openshift-marketplace/redhat-operators-9fzhp" Dec 11 03:08:31 crc kubenswrapper[4824]: I1211 03:08:31.034170 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sswrn\" (UniqueName: \"kubernetes.io/projected/224f20e3-2b4a-4e0c-bbad-906eb92a8ce9-kube-api-access-sswrn\") pod \"redhat-operators-9fzhp\" (UID: \"224f20e3-2b4a-4e0c-bbad-906eb92a8ce9\") " pod="openshift-marketplace/redhat-operators-9fzhp" Dec 11 03:08:31 crc kubenswrapper[4824]: I1211 03:08:31.179609 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9fzhp" Dec 11 03:08:31 crc kubenswrapper[4824]: I1211 03:08:31.624521 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9fzhp"] Dec 11 03:08:32 crc kubenswrapper[4824]: I1211 03:08:32.243279 4824 generic.go:334] "Generic (PLEG): container finished" podID="224f20e3-2b4a-4e0c-bbad-906eb92a8ce9" containerID="ef2065943f50ae1455d74fc848b9363c71e387dcad3074ac790760d28db0bacf" exitCode=0 Dec 11 03:08:32 crc kubenswrapper[4824]: I1211 03:08:32.243369 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9fzhp" event={"ID":"224f20e3-2b4a-4e0c-bbad-906eb92a8ce9","Type":"ContainerDied","Data":"ef2065943f50ae1455d74fc848b9363c71e387dcad3074ac790760d28db0bacf"} Dec 11 03:08:32 crc kubenswrapper[4824]: I1211 03:08:32.243614 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9fzhp" event={"ID":"224f20e3-2b4a-4e0c-bbad-906eb92a8ce9","Type":"ContainerStarted","Data":"655a0201cac6fd8ca971582d02951ba5990cf461687e35372cf6de6d4ea266b8"} Dec 11 03:08:32 crc kubenswrapper[4824]: I1211 03:08:32.717712 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-v65vx/must-gather-rwnp8"] Dec 11 03:08:32 crc kubenswrapper[4824]: I1211 03:08:32.718204 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-v65vx/must-gather-rwnp8" podUID="bb60c05d-1bda-4726-a2f2-5967992fd930" containerName="copy" containerID="cri-o://b60d2ff1c3092c6e50b7344313070e8e1d93cdc8cb0b8d3024dc8cc48b2d7f15" gracePeriod=2 Dec 11 03:08:32 crc kubenswrapper[4824]: I1211 03:08:32.728286 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-v65vx/must-gather-rwnp8"] Dec 11 03:08:33 crc kubenswrapper[4824]: I1211 03:08:33.256643 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-v65vx_must-gather-rwnp8_bb60c05d-1bda-4726-a2f2-5967992fd930/copy/0.log" Dec 11 03:08:33 crc kubenswrapper[4824]: I1211 03:08:33.258166 4824 generic.go:334] "Generic (PLEG): container finished" podID="bb60c05d-1bda-4726-a2f2-5967992fd930" containerID="b60d2ff1c3092c6e50b7344313070e8e1d93cdc8cb0b8d3024dc8cc48b2d7f15" exitCode=143 Dec 11 03:08:33 crc kubenswrapper[4824]: I1211 03:08:33.515066 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-v65vx_must-gather-rwnp8_bb60c05d-1bda-4726-a2f2-5967992fd930/copy/0.log" Dec 11 03:08:33 crc kubenswrapper[4824]: I1211 03:08:33.515768 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-v65vx/must-gather-rwnp8" Dec 11 03:08:33 crc kubenswrapper[4824]: I1211 03:08:33.661222 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/bb60c05d-1bda-4726-a2f2-5967992fd930-must-gather-output\") pod \"bb60c05d-1bda-4726-a2f2-5967992fd930\" (UID: \"bb60c05d-1bda-4726-a2f2-5967992fd930\") " Dec 11 03:08:33 crc kubenswrapper[4824]: I1211 03:08:33.661507 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v7gkg\" (UniqueName: \"kubernetes.io/projected/bb60c05d-1bda-4726-a2f2-5967992fd930-kube-api-access-v7gkg\") pod \"bb60c05d-1bda-4726-a2f2-5967992fd930\" (UID: \"bb60c05d-1bda-4726-a2f2-5967992fd930\") " Dec 11 03:08:33 crc kubenswrapper[4824]: I1211 03:08:33.666994 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb60c05d-1bda-4726-a2f2-5967992fd930-kube-api-access-v7gkg" (OuterVolumeSpecName: "kube-api-access-v7gkg") pod "bb60c05d-1bda-4726-a2f2-5967992fd930" (UID: "bb60c05d-1bda-4726-a2f2-5967992fd930"). InnerVolumeSpecName "kube-api-access-v7gkg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 03:08:33 crc kubenswrapper[4824]: I1211 03:08:33.764279 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v7gkg\" (UniqueName: \"kubernetes.io/projected/bb60c05d-1bda-4726-a2f2-5967992fd930-kube-api-access-v7gkg\") on node \"crc\" DevicePath \"\"" Dec 11 03:08:33 crc kubenswrapper[4824]: I1211 03:08:33.800366 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb60c05d-1bda-4726-a2f2-5967992fd930-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "bb60c05d-1bda-4726-a2f2-5967992fd930" (UID: "bb60c05d-1bda-4726-a2f2-5967992fd930"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 03:08:33 crc kubenswrapper[4824]: I1211 03:08:33.865674 4824 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/bb60c05d-1bda-4726-a2f2-5967992fd930-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 11 03:08:34 crc kubenswrapper[4824]: I1211 03:08:34.271241 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-v65vx_must-gather-rwnp8_bb60c05d-1bda-4726-a2f2-5967992fd930/copy/0.log" Dec 11 03:08:34 crc kubenswrapper[4824]: I1211 03:08:34.272188 4824 scope.go:117] "RemoveContainer" containerID="b60d2ff1c3092c6e50b7344313070e8e1d93cdc8cb0b8d3024dc8cc48b2d7f15" Dec 11 03:08:34 crc kubenswrapper[4824]: I1211 03:08:34.272271 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-v65vx/must-gather-rwnp8" Dec 11 03:08:34 crc kubenswrapper[4824]: I1211 03:08:34.329599 4824 scope.go:117] "RemoveContainer" containerID="899e0d503155df37fa14e0dd1b38b28b8f72e830e29c0e83ba0b27f26c27dc3c" Dec 11 03:08:34 crc kubenswrapper[4824]: I1211 03:08:34.646843 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb60c05d-1bda-4726-a2f2-5967992fd930" path="/var/lib/kubelet/pods/bb60c05d-1bda-4726-a2f2-5967992fd930/volumes" Dec 11 03:08:35 crc kubenswrapper[4824]: I1211 03:08:35.286653 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9fzhp" event={"ID":"224f20e3-2b4a-4e0c-bbad-906eb92a8ce9","Type":"ContainerStarted","Data":"50ef91e3ba567345cc80dedc87f7c4d1097377a5ed1f6bbe3f280efeb186dfba"} Dec 11 03:08:37 crc kubenswrapper[4824]: I1211 03:08:37.251053 4824 patch_prober.go:28] interesting pod/machine-config-daemon-gx6xt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 03:08:37 crc kubenswrapper[4824]: I1211 03:08:37.251586 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 03:08:37 crc kubenswrapper[4824]: I1211 03:08:37.315670 4824 generic.go:334] "Generic (PLEG): container finished" podID="224f20e3-2b4a-4e0c-bbad-906eb92a8ce9" containerID="50ef91e3ba567345cc80dedc87f7c4d1097377a5ed1f6bbe3f280efeb186dfba" exitCode=0 Dec 11 03:08:37 crc kubenswrapper[4824]: I1211 03:08:37.315677 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9fzhp" event={"ID":"224f20e3-2b4a-4e0c-bbad-906eb92a8ce9","Type":"ContainerDied","Data":"50ef91e3ba567345cc80dedc87f7c4d1097377a5ed1f6bbe3f280efeb186dfba"} Dec 11 03:08:38 crc kubenswrapper[4824]: I1211 03:08:38.331024 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9fzhp" event={"ID":"224f20e3-2b4a-4e0c-bbad-906eb92a8ce9","Type":"ContainerStarted","Data":"36482560dfa3b8e79bf67c09dd45660cd9a6279b592ee8b15b1192a284bd344f"} Dec 11 03:08:38 crc kubenswrapper[4824]: I1211 03:08:38.366134 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-9fzhp" podStartSLOduration=2.7335170140000002 podStartE2EDuration="8.366105065s" podCreationTimestamp="2025-12-11 03:08:30 +0000 UTC" firstStartedPulling="2025-12-11 03:08:32.244896762 +0000 UTC m=+4053.933934131" lastFinishedPulling="2025-12-11 03:08:37.877484783 +0000 UTC m=+4059.566522182" observedRunningTime="2025-12-11 03:08:38.358672786 +0000 UTC m=+4060.047710176" watchObservedRunningTime="2025-12-11 03:08:38.366105065 +0000 UTC m=+4060.055142474" Dec 11 03:08:41 crc kubenswrapper[4824]: I1211 03:08:41.179989 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-9fzhp" Dec 11 03:08:41 crc kubenswrapper[4824]: I1211 03:08:41.181202 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-9fzhp" Dec 11 03:08:42 crc kubenswrapper[4824]: I1211 03:08:42.240312 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-9fzhp" podUID="224f20e3-2b4a-4e0c-bbad-906eb92a8ce9" containerName="registry-server" probeResult="failure" output=< Dec 11 03:08:42 crc kubenswrapper[4824]: timeout: failed to connect service ":50051" within 1s Dec 11 03:08:42 crc kubenswrapper[4824]: > Dec 11 03:08:51 crc kubenswrapper[4824]: I1211 03:08:51.240769 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-9fzhp" Dec 11 03:08:51 crc kubenswrapper[4824]: I1211 03:08:51.286608 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-9fzhp" Dec 11 03:08:51 crc kubenswrapper[4824]: I1211 03:08:51.481074 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9fzhp"] Dec 11 03:08:52 crc kubenswrapper[4824]: I1211 03:08:52.489078 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-9fzhp" podUID="224f20e3-2b4a-4e0c-bbad-906eb92a8ce9" containerName="registry-server" containerID="cri-o://36482560dfa3b8e79bf67c09dd45660cd9a6279b592ee8b15b1192a284bd344f" gracePeriod=2 Dec 11 03:08:53 crc kubenswrapper[4824]: I1211 03:08:53.001324 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9fzhp" Dec 11 03:08:53 crc kubenswrapper[4824]: I1211 03:08:53.200057 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sswrn\" (UniqueName: \"kubernetes.io/projected/224f20e3-2b4a-4e0c-bbad-906eb92a8ce9-kube-api-access-sswrn\") pod \"224f20e3-2b4a-4e0c-bbad-906eb92a8ce9\" (UID: \"224f20e3-2b4a-4e0c-bbad-906eb92a8ce9\") " Dec 11 03:08:53 crc kubenswrapper[4824]: I1211 03:08:53.200247 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/224f20e3-2b4a-4e0c-bbad-906eb92a8ce9-utilities\") pod \"224f20e3-2b4a-4e0c-bbad-906eb92a8ce9\" (UID: \"224f20e3-2b4a-4e0c-bbad-906eb92a8ce9\") " Dec 11 03:08:53 crc kubenswrapper[4824]: I1211 03:08:53.200345 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/224f20e3-2b4a-4e0c-bbad-906eb92a8ce9-catalog-content\") pod \"224f20e3-2b4a-4e0c-bbad-906eb92a8ce9\" (UID: \"224f20e3-2b4a-4e0c-bbad-906eb92a8ce9\") " Dec 11 03:08:53 crc kubenswrapper[4824]: I1211 03:08:53.201820 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/224f20e3-2b4a-4e0c-bbad-906eb92a8ce9-utilities" (OuterVolumeSpecName: "utilities") pod "224f20e3-2b4a-4e0c-bbad-906eb92a8ce9" (UID: "224f20e3-2b4a-4e0c-bbad-906eb92a8ce9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 03:08:53 crc kubenswrapper[4824]: I1211 03:08:53.205571 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/224f20e3-2b4a-4e0c-bbad-906eb92a8ce9-kube-api-access-sswrn" (OuterVolumeSpecName: "kube-api-access-sswrn") pod "224f20e3-2b4a-4e0c-bbad-906eb92a8ce9" (UID: "224f20e3-2b4a-4e0c-bbad-906eb92a8ce9"). InnerVolumeSpecName "kube-api-access-sswrn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 03:08:53 crc kubenswrapper[4824]: I1211 03:08:53.303181 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sswrn\" (UniqueName: \"kubernetes.io/projected/224f20e3-2b4a-4e0c-bbad-906eb92a8ce9-kube-api-access-sswrn\") on node \"crc\" DevicePath \"\"" Dec 11 03:08:53 crc kubenswrapper[4824]: I1211 03:08:53.303227 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/224f20e3-2b4a-4e0c-bbad-906eb92a8ce9-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 03:08:53 crc kubenswrapper[4824]: I1211 03:08:53.344820 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/224f20e3-2b4a-4e0c-bbad-906eb92a8ce9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "224f20e3-2b4a-4e0c-bbad-906eb92a8ce9" (UID: "224f20e3-2b4a-4e0c-bbad-906eb92a8ce9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 03:08:53 crc kubenswrapper[4824]: I1211 03:08:53.404631 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/224f20e3-2b4a-4e0c-bbad-906eb92a8ce9-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 03:08:53 crc kubenswrapper[4824]: I1211 03:08:53.506322 4824 generic.go:334] "Generic (PLEG): container finished" podID="224f20e3-2b4a-4e0c-bbad-906eb92a8ce9" containerID="36482560dfa3b8e79bf67c09dd45660cd9a6279b592ee8b15b1192a284bd344f" exitCode=0 Dec 11 03:08:53 crc kubenswrapper[4824]: I1211 03:08:53.506387 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9fzhp" event={"ID":"224f20e3-2b4a-4e0c-bbad-906eb92a8ce9","Type":"ContainerDied","Data":"36482560dfa3b8e79bf67c09dd45660cd9a6279b592ee8b15b1192a284bd344f"} Dec 11 03:08:53 crc kubenswrapper[4824]: I1211 03:08:53.506429 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9fzhp" event={"ID":"224f20e3-2b4a-4e0c-bbad-906eb92a8ce9","Type":"ContainerDied","Data":"655a0201cac6fd8ca971582d02951ba5990cf461687e35372cf6de6d4ea266b8"} Dec 11 03:08:53 crc kubenswrapper[4824]: I1211 03:08:53.506426 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9fzhp" Dec 11 03:08:53 crc kubenswrapper[4824]: I1211 03:08:53.506451 4824 scope.go:117] "RemoveContainer" containerID="36482560dfa3b8e79bf67c09dd45660cd9a6279b592ee8b15b1192a284bd344f" Dec 11 03:08:53 crc kubenswrapper[4824]: I1211 03:08:53.548973 4824 scope.go:117] "RemoveContainer" containerID="50ef91e3ba567345cc80dedc87f7c4d1097377a5ed1f6bbe3f280efeb186dfba" Dec 11 03:08:53 crc kubenswrapper[4824]: I1211 03:08:53.570101 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9fzhp"] Dec 11 03:08:53 crc kubenswrapper[4824]: I1211 03:08:53.584824 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-9fzhp"] Dec 11 03:08:53 crc kubenswrapper[4824]: I1211 03:08:53.596159 4824 scope.go:117] "RemoveContainer" containerID="ef2065943f50ae1455d74fc848b9363c71e387dcad3074ac790760d28db0bacf" Dec 11 03:08:53 crc kubenswrapper[4824]: I1211 03:08:53.645678 4824 scope.go:117] "RemoveContainer" containerID="36482560dfa3b8e79bf67c09dd45660cd9a6279b592ee8b15b1192a284bd344f" Dec 11 03:08:53 crc kubenswrapper[4824]: E1211 03:08:53.646285 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"36482560dfa3b8e79bf67c09dd45660cd9a6279b592ee8b15b1192a284bd344f\": container with ID starting with 36482560dfa3b8e79bf67c09dd45660cd9a6279b592ee8b15b1192a284bd344f not found: ID does not exist" containerID="36482560dfa3b8e79bf67c09dd45660cd9a6279b592ee8b15b1192a284bd344f" Dec 11 03:08:53 crc kubenswrapper[4824]: I1211 03:08:53.646325 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"36482560dfa3b8e79bf67c09dd45660cd9a6279b592ee8b15b1192a284bd344f"} err="failed to get container status \"36482560dfa3b8e79bf67c09dd45660cd9a6279b592ee8b15b1192a284bd344f\": rpc error: code = NotFound desc = could not find container \"36482560dfa3b8e79bf67c09dd45660cd9a6279b592ee8b15b1192a284bd344f\": container with ID starting with 36482560dfa3b8e79bf67c09dd45660cd9a6279b592ee8b15b1192a284bd344f not found: ID does not exist" Dec 11 03:08:53 crc kubenswrapper[4824]: I1211 03:08:53.646350 4824 scope.go:117] "RemoveContainer" containerID="50ef91e3ba567345cc80dedc87f7c4d1097377a5ed1f6bbe3f280efeb186dfba" Dec 11 03:08:53 crc kubenswrapper[4824]: E1211 03:08:53.646667 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"50ef91e3ba567345cc80dedc87f7c4d1097377a5ed1f6bbe3f280efeb186dfba\": container with ID starting with 50ef91e3ba567345cc80dedc87f7c4d1097377a5ed1f6bbe3f280efeb186dfba not found: ID does not exist" containerID="50ef91e3ba567345cc80dedc87f7c4d1097377a5ed1f6bbe3f280efeb186dfba" Dec 11 03:08:53 crc kubenswrapper[4824]: I1211 03:08:53.646700 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50ef91e3ba567345cc80dedc87f7c4d1097377a5ed1f6bbe3f280efeb186dfba"} err="failed to get container status \"50ef91e3ba567345cc80dedc87f7c4d1097377a5ed1f6bbe3f280efeb186dfba\": rpc error: code = NotFound desc = could not find container \"50ef91e3ba567345cc80dedc87f7c4d1097377a5ed1f6bbe3f280efeb186dfba\": container with ID starting with 50ef91e3ba567345cc80dedc87f7c4d1097377a5ed1f6bbe3f280efeb186dfba not found: ID does not exist" Dec 11 03:08:53 crc kubenswrapper[4824]: I1211 03:08:53.646718 4824 scope.go:117] "RemoveContainer" containerID="ef2065943f50ae1455d74fc848b9363c71e387dcad3074ac790760d28db0bacf" Dec 11 03:08:53 crc kubenswrapper[4824]: E1211 03:08:53.647049 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef2065943f50ae1455d74fc848b9363c71e387dcad3074ac790760d28db0bacf\": container with ID starting with ef2065943f50ae1455d74fc848b9363c71e387dcad3074ac790760d28db0bacf not found: ID does not exist" containerID="ef2065943f50ae1455d74fc848b9363c71e387dcad3074ac790760d28db0bacf" Dec 11 03:08:53 crc kubenswrapper[4824]: I1211 03:08:53.647076 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef2065943f50ae1455d74fc848b9363c71e387dcad3074ac790760d28db0bacf"} err="failed to get container status \"ef2065943f50ae1455d74fc848b9363c71e387dcad3074ac790760d28db0bacf\": rpc error: code = NotFound desc = could not find container \"ef2065943f50ae1455d74fc848b9363c71e387dcad3074ac790760d28db0bacf\": container with ID starting with ef2065943f50ae1455d74fc848b9363c71e387dcad3074ac790760d28db0bacf not found: ID does not exist" Dec 11 03:08:54 crc kubenswrapper[4824]: I1211 03:08:54.684771 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="224f20e3-2b4a-4e0c-bbad-906eb92a8ce9" path="/var/lib/kubelet/pods/224f20e3-2b4a-4e0c-bbad-906eb92a8ce9/volumes" Dec 11 03:09:07 crc kubenswrapper[4824]: I1211 03:09:07.251587 4824 patch_prober.go:28] interesting pod/machine-config-daemon-gx6xt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 03:09:07 crc kubenswrapper[4824]: I1211 03:09:07.252352 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 03:09:07 crc kubenswrapper[4824]: I1211 03:09:07.252422 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" Dec 11 03:09:07 crc kubenswrapper[4824]: I1211 03:09:07.253449 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a0748aabb594d1c62d2ac36a888d8e889bc711466d2b3fd7120852453140e334"} pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 11 03:09:07 crc kubenswrapper[4824]: I1211 03:09:07.253550 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" containerName="machine-config-daemon" containerID="cri-o://a0748aabb594d1c62d2ac36a888d8e889bc711466d2b3fd7120852453140e334" gracePeriod=600 Dec 11 03:09:07 crc kubenswrapper[4824]: I1211 03:09:07.818343 4824 generic.go:334] "Generic (PLEG): container finished" podID="44b156e4-64a4-4d45-aa5e-9b10a862faed" containerID="a0748aabb594d1c62d2ac36a888d8e889bc711466d2b3fd7120852453140e334" exitCode=0 Dec 11 03:09:07 crc kubenswrapper[4824]: I1211 03:09:07.818810 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" event={"ID":"44b156e4-64a4-4d45-aa5e-9b10a862faed","Type":"ContainerDied","Data":"a0748aabb594d1c62d2ac36a888d8e889bc711466d2b3fd7120852453140e334"} Dec 11 03:09:07 crc kubenswrapper[4824]: I1211 03:09:07.818852 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" event={"ID":"44b156e4-64a4-4d45-aa5e-9b10a862faed","Type":"ContainerStarted","Data":"fdad360b2681cc88a198f2fc151daab36d27e39aa44b3ec03ace588804769464"} Dec 11 03:09:07 crc kubenswrapper[4824]: I1211 03:09:07.818880 4824 scope.go:117] "RemoveContainer" containerID="8dcf42d380c145f29ce6b988ed66e3ef33605f29467da12e5fe7eac51562e700" Dec 11 03:11:07 crc kubenswrapper[4824]: I1211 03:11:07.250936 4824 patch_prober.go:28] interesting pod/machine-config-daemon-gx6xt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 03:11:07 crc kubenswrapper[4824]: I1211 03:11:07.251941 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 03:11:21 crc kubenswrapper[4824]: I1211 03:11:21.823682 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-2pxpn/must-gather-8jk9m"] Dec 11 03:11:21 crc kubenswrapper[4824]: E1211 03:11:21.824528 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="224f20e3-2b4a-4e0c-bbad-906eb92a8ce9" containerName="extract-content" Dec 11 03:11:21 crc kubenswrapper[4824]: I1211 03:11:21.824544 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="224f20e3-2b4a-4e0c-bbad-906eb92a8ce9" containerName="extract-content" Dec 11 03:11:21 crc kubenswrapper[4824]: E1211 03:11:21.824559 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="224f20e3-2b4a-4e0c-bbad-906eb92a8ce9" containerName="registry-server" Dec 11 03:11:21 crc kubenswrapper[4824]: I1211 03:11:21.824567 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="224f20e3-2b4a-4e0c-bbad-906eb92a8ce9" containerName="registry-server" Dec 11 03:11:21 crc kubenswrapper[4824]: E1211 03:11:21.824587 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb60c05d-1bda-4726-a2f2-5967992fd930" containerName="copy" Dec 11 03:11:21 crc kubenswrapper[4824]: I1211 03:11:21.824596 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb60c05d-1bda-4726-a2f2-5967992fd930" containerName="copy" Dec 11 03:11:21 crc kubenswrapper[4824]: E1211 03:11:21.824614 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="224f20e3-2b4a-4e0c-bbad-906eb92a8ce9" containerName="extract-utilities" Dec 11 03:11:21 crc kubenswrapper[4824]: I1211 03:11:21.824622 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="224f20e3-2b4a-4e0c-bbad-906eb92a8ce9" containerName="extract-utilities" Dec 11 03:11:21 crc kubenswrapper[4824]: E1211 03:11:21.824640 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb60c05d-1bda-4726-a2f2-5967992fd930" containerName="gather" Dec 11 03:11:21 crc kubenswrapper[4824]: I1211 03:11:21.824647 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb60c05d-1bda-4726-a2f2-5967992fd930" containerName="gather" Dec 11 03:11:21 crc kubenswrapper[4824]: I1211 03:11:21.824860 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb60c05d-1bda-4726-a2f2-5967992fd930" containerName="copy" Dec 11 03:11:21 crc kubenswrapper[4824]: I1211 03:11:21.824872 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="224f20e3-2b4a-4e0c-bbad-906eb92a8ce9" containerName="registry-server" Dec 11 03:11:21 crc kubenswrapper[4824]: I1211 03:11:21.824903 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb60c05d-1bda-4726-a2f2-5967992fd930" containerName="gather" Dec 11 03:11:21 crc kubenswrapper[4824]: I1211 03:11:21.826003 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2pxpn/must-gather-8jk9m" Dec 11 03:11:21 crc kubenswrapper[4824]: I1211 03:11:21.828575 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-2pxpn"/"openshift-service-ca.crt" Dec 11 03:11:21 crc kubenswrapper[4824]: I1211 03:11:21.828817 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-2pxpn"/"kube-root-ca.crt" Dec 11 03:11:21 crc kubenswrapper[4824]: I1211 03:11:21.833965 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-2pxpn"/"default-dockercfg-922n7" Dec 11 03:11:21 crc kubenswrapper[4824]: I1211 03:11:21.852648 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-2pxpn/must-gather-8jk9m"] Dec 11 03:11:21 crc kubenswrapper[4824]: I1211 03:11:21.950961 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8jjzm\" (UniqueName: \"kubernetes.io/projected/9a78e53b-6cf8-44aa-abbd-3f5ab6fe156b-kube-api-access-8jjzm\") pod \"must-gather-8jk9m\" (UID: \"9a78e53b-6cf8-44aa-abbd-3f5ab6fe156b\") " pod="openshift-must-gather-2pxpn/must-gather-8jk9m" Dec 11 03:11:21 crc kubenswrapper[4824]: I1211 03:11:21.951066 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/9a78e53b-6cf8-44aa-abbd-3f5ab6fe156b-must-gather-output\") pod \"must-gather-8jk9m\" (UID: \"9a78e53b-6cf8-44aa-abbd-3f5ab6fe156b\") " pod="openshift-must-gather-2pxpn/must-gather-8jk9m" Dec 11 03:11:22 crc kubenswrapper[4824]: I1211 03:11:22.052727 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/9a78e53b-6cf8-44aa-abbd-3f5ab6fe156b-must-gather-output\") pod \"must-gather-8jk9m\" (UID: \"9a78e53b-6cf8-44aa-abbd-3f5ab6fe156b\") " pod="openshift-must-gather-2pxpn/must-gather-8jk9m" Dec 11 03:11:22 crc kubenswrapper[4824]: I1211 03:11:22.052857 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8jjzm\" (UniqueName: \"kubernetes.io/projected/9a78e53b-6cf8-44aa-abbd-3f5ab6fe156b-kube-api-access-8jjzm\") pod \"must-gather-8jk9m\" (UID: \"9a78e53b-6cf8-44aa-abbd-3f5ab6fe156b\") " pod="openshift-must-gather-2pxpn/must-gather-8jk9m" Dec 11 03:11:22 crc kubenswrapper[4824]: I1211 03:11:22.053225 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/9a78e53b-6cf8-44aa-abbd-3f5ab6fe156b-must-gather-output\") pod \"must-gather-8jk9m\" (UID: \"9a78e53b-6cf8-44aa-abbd-3f5ab6fe156b\") " pod="openshift-must-gather-2pxpn/must-gather-8jk9m" Dec 11 03:11:22 crc kubenswrapper[4824]: I1211 03:11:22.074107 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8jjzm\" (UniqueName: \"kubernetes.io/projected/9a78e53b-6cf8-44aa-abbd-3f5ab6fe156b-kube-api-access-8jjzm\") pod \"must-gather-8jk9m\" (UID: \"9a78e53b-6cf8-44aa-abbd-3f5ab6fe156b\") " pod="openshift-must-gather-2pxpn/must-gather-8jk9m" Dec 11 03:11:22 crc kubenswrapper[4824]: I1211 03:11:22.151411 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2pxpn/must-gather-8jk9m" Dec 11 03:11:22 crc kubenswrapper[4824]: I1211 03:11:22.465462 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-2pxpn/must-gather-8jk9m"] Dec 11 03:11:23 crc kubenswrapper[4824]: I1211 03:11:23.445305 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2pxpn/must-gather-8jk9m" event={"ID":"9a78e53b-6cf8-44aa-abbd-3f5ab6fe156b","Type":"ContainerStarted","Data":"840585396fdd658dee8442e9090dd91c78f38255e7946543986306315a99228d"} Dec 11 03:11:23 crc kubenswrapper[4824]: I1211 03:11:23.445625 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2pxpn/must-gather-8jk9m" event={"ID":"9a78e53b-6cf8-44aa-abbd-3f5ab6fe156b","Type":"ContainerStarted","Data":"565bc55d51103690977681b6054c6bfe409d8f1c588376ce223d412aafbd75bd"} Dec 11 03:11:23 crc kubenswrapper[4824]: I1211 03:11:23.445635 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2pxpn/must-gather-8jk9m" event={"ID":"9a78e53b-6cf8-44aa-abbd-3f5ab6fe156b","Type":"ContainerStarted","Data":"186bdea14d193b1c3699fd988179eacb4fc2dc2eaecbd7aa3554ff365131dd6b"} Dec 11 03:11:23 crc kubenswrapper[4824]: I1211 03:11:23.467528 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-2pxpn/must-gather-8jk9m" podStartSLOduration=2.467504663 podStartE2EDuration="2.467504663s" podCreationTimestamp="2025-12-11 03:11:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 03:11:23.46139934 +0000 UTC m=+4225.150436729" watchObservedRunningTime="2025-12-11 03:11:23.467504663 +0000 UTC m=+4225.156542052" Dec 11 03:11:26 crc kubenswrapper[4824]: I1211 03:11:26.256230 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-2pxpn/crc-debug-lphhr"] Dec 11 03:11:26 crc kubenswrapper[4824]: I1211 03:11:26.257717 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2pxpn/crc-debug-lphhr" Dec 11 03:11:26 crc kubenswrapper[4824]: I1211 03:11:26.341292 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/feac62ad-eb56-44e6-ab03-e2a7bd8dd4bf-host\") pod \"crc-debug-lphhr\" (UID: \"feac62ad-eb56-44e6-ab03-e2a7bd8dd4bf\") " pod="openshift-must-gather-2pxpn/crc-debug-lphhr" Dec 11 03:11:26 crc kubenswrapper[4824]: I1211 03:11:26.341481 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cd5zx\" (UniqueName: \"kubernetes.io/projected/feac62ad-eb56-44e6-ab03-e2a7bd8dd4bf-kube-api-access-cd5zx\") pod \"crc-debug-lphhr\" (UID: \"feac62ad-eb56-44e6-ab03-e2a7bd8dd4bf\") " pod="openshift-must-gather-2pxpn/crc-debug-lphhr" Dec 11 03:11:26 crc kubenswrapper[4824]: I1211 03:11:26.443833 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/feac62ad-eb56-44e6-ab03-e2a7bd8dd4bf-host\") pod \"crc-debug-lphhr\" (UID: \"feac62ad-eb56-44e6-ab03-e2a7bd8dd4bf\") " pod="openshift-must-gather-2pxpn/crc-debug-lphhr" Dec 11 03:11:26 crc kubenswrapper[4824]: I1211 03:11:26.443956 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cd5zx\" (UniqueName: \"kubernetes.io/projected/feac62ad-eb56-44e6-ab03-e2a7bd8dd4bf-kube-api-access-cd5zx\") pod \"crc-debug-lphhr\" (UID: \"feac62ad-eb56-44e6-ab03-e2a7bd8dd4bf\") " pod="openshift-must-gather-2pxpn/crc-debug-lphhr" Dec 11 03:11:26 crc kubenswrapper[4824]: I1211 03:11:26.444036 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/feac62ad-eb56-44e6-ab03-e2a7bd8dd4bf-host\") pod \"crc-debug-lphhr\" (UID: \"feac62ad-eb56-44e6-ab03-e2a7bd8dd4bf\") " pod="openshift-must-gather-2pxpn/crc-debug-lphhr" Dec 11 03:11:26 crc kubenswrapper[4824]: I1211 03:11:26.474243 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cd5zx\" (UniqueName: \"kubernetes.io/projected/feac62ad-eb56-44e6-ab03-e2a7bd8dd4bf-kube-api-access-cd5zx\") pod \"crc-debug-lphhr\" (UID: \"feac62ad-eb56-44e6-ab03-e2a7bd8dd4bf\") " pod="openshift-must-gather-2pxpn/crc-debug-lphhr" Dec 11 03:11:26 crc kubenswrapper[4824]: I1211 03:11:26.576009 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2pxpn/crc-debug-lphhr" Dec 11 03:11:27 crc kubenswrapper[4824]: I1211 03:11:27.488978 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2pxpn/crc-debug-lphhr" event={"ID":"feac62ad-eb56-44e6-ab03-e2a7bd8dd4bf","Type":"ContainerStarted","Data":"5d0b0836a4a3dd0a08c418f770678040e24414d98b8c2f2ed5032d3a008794d5"} Dec 11 03:11:27 crc kubenswrapper[4824]: I1211 03:11:27.489549 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2pxpn/crc-debug-lphhr" event={"ID":"feac62ad-eb56-44e6-ab03-e2a7bd8dd4bf","Type":"ContainerStarted","Data":"2755b8dca02f208ba7e4a10db3d17bbd8c247d4695bbe981b3db92e0551f5830"} Dec 11 03:11:27 crc kubenswrapper[4824]: I1211 03:11:27.508454 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-2pxpn/crc-debug-lphhr" podStartSLOduration=1.5084398 podStartE2EDuration="1.5084398s" podCreationTimestamp="2025-12-11 03:11:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 03:11:27.505197033 +0000 UTC m=+4229.194234412" watchObservedRunningTime="2025-12-11 03:11:27.5084398 +0000 UTC m=+4229.197477179" Dec 11 03:11:37 crc kubenswrapper[4824]: I1211 03:11:37.251038 4824 patch_prober.go:28] interesting pod/machine-config-daemon-gx6xt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 03:11:37 crc kubenswrapper[4824]: I1211 03:11:37.251606 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 03:11:59 crc kubenswrapper[4824]: I1211 03:11:59.808779 4824 generic.go:334] "Generic (PLEG): container finished" podID="feac62ad-eb56-44e6-ab03-e2a7bd8dd4bf" containerID="5d0b0836a4a3dd0a08c418f770678040e24414d98b8c2f2ed5032d3a008794d5" exitCode=0 Dec 11 03:11:59 crc kubenswrapper[4824]: I1211 03:11:59.808854 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2pxpn/crc-debug-lphhr" event={"ID":"feac62ad-eb56-44e6-ab03-e2a7bd8dd4bf","Type":"ContainerDied","Data":"5d0b0836a4a3dd0a08c418f770678040e24414d98b8c2f2ed5032d3a008794d5"} Dec 11 03:12:00 crc kubenswrapper[4824]: I1211 03:12:00.921273 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2pxpn/crc-debug-lphhr" Dec 11 03:12:00 crc kubenswrapper[4824]: I1211 03:12:00.954564 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-2pxpn/crc-debug-lphhr"] Dec 11 03:12:00 crc kubenswrapper[4824]: I1211 03:12:00.965028 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-2pxpn/crc-debug-lphhr"] Dec 11 03:12:01 crc kubenswrapper[4824]: I1211 03:12:01.124959 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/feac62ad-eb56-44e6-ab03-e2a7bd8dd4bf-host\") pod \"feac62ad-eb56-44e6-ab03-e2a7bd8dd4bf\" (UID: \"feac62ad-eb56-44e6-ab03-e2a7bd8dd4bf\") " Dec 11 03:12:01 crc kubenswrapper[4824]: I1211 03:12:01.125179 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cd5zx\" (UniqueName: \"kubernetes.io/projected/feac62ad-eb56-44e6-ab03-e2a7bd8dd4bf-kube-api-access-cd5zx\") pod \"feac62ad-eb56-44e6-ab03-e2a7bd8dd4bf\" (UID: \"feac62ad-eb56-44e6-ab03-e2a7bd8dd4bf\") " Dec 11 03:12:01 crc kubenswrapper[4824]: I1211 03:12:01.125501 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/feac62ad-eb56-44e6-ab03-e2a7bd8dd4bf-host" (OuterVolumeSpecName: "host") pod "feac62ad-eb56-44e6-ab03-e2a7bd8dd4bf" (UID: "feac62ad-eb56-44e6-ab03-e2a7bd8dd4bf"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 03:12:01 crc kubenswrapper[4824]: I1211 03:12:01.125806 4824 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/feac62ad-eb56-44e6-ab03-e2a7bd8dd4bf-host\") on node \"crc\" DevicePath \"\"" Dec 11 03:12:01 crc kubenswrapper[4824]: I1211 03:12:01.144762 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/feac62ad-eb56-44e6-ab03-e2a7bd8dd4bf-kube-api-access-cd5zx" (OuterVolumeSpecName: "kube-api-access-cd5zx") pod "feac62ad-eb56-44e6-ab03-e2a7bd8dd4bf" (UID: "feac62ad-eb56-44e6-ab03-e2a7bd8dd4bf"). InnerVolumeSpecName "kube-api-access-cd5zx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 03:12:01 crc kubenswrapper[4824]: I1211 03:12:01.227357 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cd5zx\" (UniqueName: \"kubernetes.io/projected/feac62ad-eb56-44e6-ab03-e2a7bd8dd4bf-kube-api-access-cd5zx\") on node \"crc\" DevicePath \"\"" Dec 11 03:12:01 crc kubenswrapper[4824]: I1211 03:12:01.831291 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2755b8dca02f208ba7e4a10db3d17bbd8c247d4695bbe981b3db92e0551f5830" Dec 11 03:12:01 crc kubenswrapper[4824]: I1211 03:12:01.831368 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2pxpn/crc-debug-lphhr" Dec 11 03:12:02 crc kubenswrapper[4824]: I1211 03:12:02.206877 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-2pxpn/crc-debug-xsfqz"] Dec 11 03:12:02 crc kubenswrapper[4824]: E1211 03:12:02.207344 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="feac62ad-eb56-44e6-ab03-e2a7bd8dd4bf" containerName="container-00" Dec 11 03:12:02 crc kubenswrapper[4824]: I1211 03:12:02.207357 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="feac62ad-eb56-44e6-ab03-e2a7bd8dd4bf" containerName="container-00" Dec 11 03:12:02 crc kubenswrapper[4824]: I1211 03:12:02.207531 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="feac62ad-eb56-44e6-ab03-e2a7bd8dd4bf" containerName="container-00" Dec 11 03:12:02 crc kubenswrapper[4824]: I1211 03:12:02.208164 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2pxpn/crc-debug-xsfqz" Dec 11 03:12:02 crc kubenswrapper[4824]: I1211 03:12:02.247502 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8d5mv\" (UniqueName: \"kubernetes.io/projected/37dde005-b9ad-4223-8996-ce62500f6c3c-kube-api-access-8d5mv\") pod \"crc-debug-xsfqz\" (UID: \"37dde005-b9ad-4223-8996-ce62500f6c3c\") " pod="openshift-must-gather-2pxpn/crc-debug-xsfqz" Dec 11 03:12:02 crc kubenswrapper[4824]: I1211 03:12:02.247664 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/37dde005-b9ad-4223-8996-ce62500f6c3c-host\") pod \"crc-debug-xsfqz\" (UID: \"37dde005-b9ad-4223-8996-ce62500f6c3c\") " pod="openshift-must-gather-2pxpn/crc-debug-xsfqz" Dec 11 03:12:02 crc kubenswrapper[4824]: I1211 03:12:02.349821 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8d5mv\" (UniqueName: \"kubernetes.io/projected/37dde005-b9ad-4223-8996-ce62500f6c3c-kube-api-access-8d5mv\") pod \"crc-debug-xsfqz\" (UID: \"37dde005-b9ad-4223-8996-ce62500f6c3c\") " pod="openshift-must-gather-2pxpn/crc-debug-xsfqz" Dec 11 03:12:02 crc kubenswrapper[4824]: I1211 03:12:02.349886 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/37dde005-b9ad-4223-8996-ce62500f6c3c-host\") pod \"crc-debug-xsfqz\" (UID: \"37dde005-b9ad-4223-8996-ce62500f6c3c\") " pod="openshift-must-gather-2pxpn/crc-debug-xsfqz" Dec 11 03:12:02 crc kubenswrapper[4824]: I1211 03:12:02.350044 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/37dde005-b9ad-4223-8996-ce62500f6c3c-host\") pod \"crc-debug-xsfqz\" (UID: \"37dde005-b9ad-4223-8996-ce62500f6c3c\") " pod="openshift-must-gather-2pxpn/crc-debug-xsfqz" Dec 11 03:12:02 crc kubenswrapper[4824]: I1211 03:12:02.367544 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8d5mv\" (UniqueName: \"kubernetes.io/projected/37dde005-b9ad-4223-8996-ce62500f6c3c-kube-api-access-8d5mv\") pod \"crc-debug-xsfqz\" (UID: \"37dde005-b9ad-4223-8996-ce62500f6c3c\") " pod="openshift-must-gather-2pxpn/crc-debug-xsfqz" Dec 11 03:12:02 crc kubenswrapper[4824]: I1211 03:12:02.541388 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2pxpn/crc-debug-xsfqz" Dec 11 03:12:02 crc kubenswrapper[4824]: W1211 03:12:02.579960 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod37dde005_b9ad_4223_8996_ce62500f6c3c.slice/crio-4b7ea7a6f4b61c3c99c2c311775c638e394ce9cd9274262635cde54d28fe6a7d WatchSource:0}: Error finding container 4b7ea7a6f4b61c3c99c2c311775c638e394ce9cd9274262635cde54d28fe6a7d: Status 404 returned error can't find the container with id 4b7ea7a6f4b61c3c99c2c311775c638e394ce9cd9274262635cde54d28fe6a7d Dec 11 03:12:02 crc kubenswrapper[4824]: I1211 03:12:02.647213 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="feac62ad-eb56-44e6-ab03-e2a7bd8dd4bf" path="/var/lib/kubelet/pods/feac62ad-eb56-44e6-ab03-e2a7bd8dd4bf/volumes" Dec 11 03:12:02 crc kubenswrapper[4824]: I1211 03:12:02.842858 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2pxpn/crc-debug-xsfqz" event={"ID":"37dde005-b9ad-4223-8996-ce62500f6c3c","Type":"ContainerStarted","Data":"d648bb6134869e6b36beef509d5393ef4854b7351ae0a020516e478a5d56ddd3"} Dec 11 03:12:02 crc kubenswrapper[4824]: I1211 03:12:02.843289 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2pxpn/crc-debug-xsfqz" event={"ID":"37dde005-b9ad-4223-8996-ce62500f6c3c","Type":"ContainerStarted","Data":"4b7ea7a6f4b61c3c99c2c311775c638e394ce9cd9274262635cde54d28fe6a7d"} Dec 11 03:12:02 crc kubenswrapper[4824]: I1211 03:12:02.859895 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-2pxpn/crc-debug-xsfqz" podStartSLOduration=0.859877079 podStartE2EDuration="859.877079ms" podCreationTimestamp="2025-12-11 03:12:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 03:12:02.859217522 +0000 UTC m=+4264.548254901" watchObservedRunningTime="2025-12-11 03:12:02.859877079 +0000 UTC m=+4264.548914478" Dec 11 03:12:03 crc kubenswrapper[4824]: I1211 03:12:03.857242 4824 generic.go:334] "Generic (PLEG): container finished" podID="37dde005-b9ad-4223-8996-ce62500f6c3c" containerID="d648bb6134869e6b36beef509d5393ef4854b7351ae0a020516e478a5d56ddd3" exitCode=0 Dec 11 03:12:03 crc kubenswrapper[4824]: I1211 03:12:03.857434 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2pxpn/crc-debug-xsfqz" event={"ID":"37dde005-b9ad-4223-8996-ce62500f6c3c","Type":"ContainerDied","Data":"d648bb6134869e6b36beef509d5393ef4854b7351ae0a020516e478a5d56ddd3"} Dec 11 03:12:04 crc kubenswrapper[4824]: I1211 03:12:04.979601 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2pxpn/crc-debug-xsfqz" Dec 11 03:12:05 crc kubenswrapper[4824]: I1211 03:12:05.017835 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-2pxpn/crc-debug-xsfqz"] Dec 11 03:12:05 crc kubenswrapper[4824]: I1211 03:12:05.025236 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-2pxpn/crc-debug-xsfqz"] Dec 11 03:12:05 crc kubenswrapper[4824]: I1211 03:12:05.116933 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/37dde005-b9ad-4223-8996-ce62500f6c3c-host\") pod \"37dde005-b9ad-4223-8996-ce62500f6c3c\" (UID: \"37dde005-b9ad-4223-8996-ce62500f6c3c\") " Dec 11 03:12:05 crc kubenswrapper[4824]: I1211 03:12:05.117100 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/37dde005-b9ad-4223-8996-ce62500f6c3c-host" (OuterVolumeSpecName: "host") pod "37dde005-b9ad-4223-8996-ce62500f6c3c" (UID: "37dde005-b9ad-4223-8996-ce62500f6c3c"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 03:12:05 crc kubenswrapper[4824]: I1211 03:12:05.117226 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8d5mv\" (UniqueName: \"kubernetes.io/projected/37dde005-b9ad-4223-8996-ce62500f6c3c-kube-api-access-8d5mv\") pod \"37dde005-b9ad-4223-8996-ce62500f6c3c\" (UID: \"37dde005-b9ad-4223-8996-ce62500f6c3c\") " Dec 11 03:12:05 crc kubenswrapper[4824]: I1211 03:12:05.117806 4824 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/37dde005-b9ad-4223-8996-ce62500f6c3c-host\") on node \"crc\" DevicePath \"\"" Dec 11 03:12:05 crc kubenswrapper[4824]: I1211 03:12:05.160083 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37dde005-b9ad-4223-8996-ce62500f6c3c-kube-api-access-8d5mv" (OuterVolumeSpecName: "kube-api-access-8d5mv") pod "37dde005-b9ad-4223-8996-ce62500f6c3c" (UID: "37dde005-b9ad-4223-8996-ce62500f6c3c"). InnerVolumeSpecName "kube-api-access-8d5mv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 03:12:05 crc kubenswrapper[4824]: I1211 03:12:05.219497 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8d5mv\" (UniqueName: \"kubernetes.io/projected/37dde005-b9ad-4223-8996-ce62500f6c3c-kube-api-access-8d5mv\") on node \"crc\" DevicePath \"\"" Dec 11 03:12:05 crc kubenswrapper[4824]: I1211 03:12:05.879828 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4b7ea7a6f4b61c3c99c2c311775c638e394ce9cd9274262635cde54d28fe6a7d" Dec 11 03:12:05 crc kubenswrapper[4824]: I1211 03:12:05.879875 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2pxpn/crc-debug-xsfqz" Dec 11 03:12:06 crc kubenswrapper[4824]: I1211 03:12:06.177606 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-2pxpn/crc-debug-djg86"] Dec 11 03:12:06 crc kubenswrapper[4824]: E1211 03:12:06.178307 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37dde005-b9ad-4223-8996-ce62500f6c3c" containerName="container-00" Dec 11 03:12:06 crc kubenswrapper[4824]: I1211 03:12:06.178322 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="37dde005-b9ad-4223-8996-ce62500f6c3c" containerName="container-00" Dec 11 03:12:06 crc kubenswrapper[4824]: I1211 03:12:06.178559 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="37dde005-b9ad-4223-8996-ce62500f6c3c" containerName="container-00" Dec 11 03:12:06 crc kubenswrapper[4824]: I1211 03:12:06.179419 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2pxpn/crc-debug-djg86" Dec 11 03:12:06 crc kubenswrapper[4824]: I1211 03:12:06.340081 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jx8nk\" (UniqueName: \"kubernetes.io/projected/b05370ff-d631-451a-918f-6689b00a0f4f-kube-api-access-jx8nk\") pod \"crc-debug-djg86\" (UID: \"b05370ff-d631-451a-918f-6689b00a0f4f\") " pod="openshift-must-gather-2pxpn/crc-debug-djg86" Dec 11 03:12:06 crc kubenswrapper[4824]: I1211 03:12:06.340687 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b05370ff-d631-451a-918f-6689b00a0f4f-host\") pod \"crc-debug-djg86\" (UID: \"b05370ff-d631-451a-918f-6689b00a0f4f\") " pod="openshift-must-gather-2pxpn/crc-debug-djg86" Dec 11 03:12:06 crc kubenswrapper[4824]: I1211 03:12:06.443211 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b05370ff-d631-451a-918f-6689b00a0f4f-host\") pod \"crc-debug-djg86\" (UID: \"b05370ff-d631-451a-918f-6689b00a0f4f\") " pod="openshift-must-gather-2pxpn/crc-debug-djg86" Dec 11 03:12:06 crc kubenswrapper[4824]: I1211 03:12:06.443340 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jx8nk\" (UniqueName: \"kubernetes.io/projected/b05370ff-d631-451a-918f-6689b00a0f4f-kube-api-access-jx8nk\") pod \"crc-debug-djg86\" (UID: \"b05370ff-d631-451a-918f-6689b00a0f4f\") " pod="openshift-must-gather-2pxpn/crc-debug-djg86" Dec 11 03:12:06 crc kubenswrapper[4824]: I1211 03:12:06.444094 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b05370ff-d631-451a-918f-6689b00a0f4f-host\") pod \"crc-debug-djg86\" (UID: \"b05370ff-d631-451a-918f-6689b00a0f4f\") " pod="openshift-must-gather-2pxpn/crc-debug-djg86" Dec 11 03:12:06 crc kubenswrapper[4824]: I1211 03:12:06.461777 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jx8nk\" (UniqueName: \"kubernetes.io/projected/b05370ff-d631-451a-918f-6689b00a0f4f-kube-api-access-jx8nk\") pod \"crc-debug-djg86\" (UID: \"b05370ff-d631-451a-918f-6689b00a0f4f\") " pod="openshift-must-gather-2pxpn/crc-debug-djg86" Dec 11 03:12:06 crc kubenswrapper[4824]: I1211 03:12:06.506890 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2pxpn/crc-debug-djg86" Dec 11 03:12:06 crc kubenswrapper[4824]: W1211 03:12:06.547410 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb05370ff_d631_451a_918f_6689b00a0f4f.slice/crio-a61df7403334386f3d5adbf0dc69aaf2ff33b935ec5be009c64f33fddfd9ae19 WatchSource:0}: Error finding container a61df7403334386f3d5adbf0dc69aaf2ff33b935ec5be009c64f33fddfd9ae19: Status 404 returned error can't find the container with id a61df7403334386f3d5adbf0dc69aaf2ff33b935ec5be009c64f33fddfd9ae19 Dec 11 03:12:06 crc kubenswrapper[4824]: I1211 03:12:06.642310 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37dde005-b9ad-4223-8996-ce62500f6c3c" path="/var/lib/kubelet/pods/37dde005-b9ad-4223-8996-ce62500f6c3c/volumes" Dec 11 03:12:06 crc kubenswrapper[4824]: I1211 03:12:06.888069 4824 generic.go:334] "Generic (PLEG): container finished" podID="b05370ff-d631-451a-918f-6689b00a0f4f" containerID="4f434e4fb2c2bec59ede89a6559c8187e677b1336805793ad9f78deb9f74499b" exitCode=0 Dec 11 03:12:06 crc kubenswrapper[4824]: I1211 03:12:06.888334 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2pxpn/crc-debug-djg86" event={"ID":"b05370ff-d631-451a-918f-6689b00a0f4f","Type":"ContainerDied","Data":"4f434e4fb2c2bec59ede89a6559c8187e677b1336805793ad9f78deb9f74499b"} Dec 11 03:12:06 crc kubenswrapper[4824]: I1211 03:12:06.888369 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2pxpn/crc-debug-djg86" event={"ID":"b05370ff-d631-451a-918f-6689b00a0f4f","Type":"ContainerStarted","Data":"a61df7403334386f3d5adbf0dc69aaf2ff33b935ec5be009c64f33fddfd9ae19"} Dec 11 03:12:06 crc kubenswrapper[4824]: I1211 03:12:06.918448 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-2pxpn/crc-debug-djg86"] Dec 11 03:12:06 crc kubenswrapper[4824]: I1211 03:12:06.925397 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-2pxpn/crc-debug-djg86"] Dec 11 03:12:07 crc kubenswrapper[4824]: I1211 03:12:07.251588 4824 patch_prober.go:28] interesting pod/machine-config-daemon-gx6xt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 03:12:07 crc kubenswrapper[4824]: I1211 03:12:07.251673 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 03:12:07 crc kubenswrapper[4824]: I1211 03:12:07.251741 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" Dec 11 03:12:07 crc kubenswrapper[4824]: I1211 03:12:07.252822 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fdad360b2681cc88a198f2fc151daab36d27e39aa44b3ec03ace588804769464"} pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 11 03:12:07 crc kubenswrapper[4824]: I1211 03:12:07.252919 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" containerName="machine-config-daemon" containerID="cri-o://fdad360b2681cc88a198f2fc151daab36d27e39aa44b3ec03ace588804769464" gracePeriod=600 Dec 11 03:12:07 crc kubenswrapper[4824]: E1211 03:12:07.382180 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 03:12:07 crc kubenswrapper[4824]: I1211 03:12:07.903264 4824 generic.go:334] "Generic (PLEG): container finished" podID="44b156e4-64a4-4d45-aa5e-9b10a862faed" containerID="fdad360b2681cc88a198f2fc151daab36d27e39aa44b3ec03ace588804769464" exitCode=0 Dec 11 03:12:07 crc kubenswrapper[4824]: I1211 03:12:07.903870 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" event={"ID":"44b156e4-64a4-4d45-aa5e-9b10a862faed","Type":"ContainerDied","Data":"fdad360b2681cc88a198f2fc151daab36d27e39aa44b3ec03ace588804769464"} Dec 11 03:12:07 crc kubenswrapper[4824]: I1211 03:12:07.903928 4824 scope.go:117] "RemoveContainer" containerID="a0748aabb594d1c62d2ac36a888d8e889bc711466d2b3fd7120852453140e334" Dec 11 03:12:07 crc kubenswrapper[4824]: I1211 03:12:07.904799 4824 scope.go:117] "RemoveContainer" containerID="fdad360b2681cc88a198f2fc151daab36d27e39aa44b3ec03ace588804769464" Dec 11 03:12:07 crc kubenswrapper[4824]: E1211 03:12:07.905242 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 03:12:08 crc kubenswrapper[4824]: I1211 03:12:08.037152 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2pxpn/crc-debug-djg86" Dec 11 03:12:08 crc kubenswrapper[4824]: I1211 03:12:08.175997 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jx8nk\" (UniqueName: \"kubernetes.io/projected/b05370ff-d631-451a-918f-6689b00a0f4f-kube-api-access-jx8nk\") pod \"b05370ff-d631-451a-918f-6689b00a0f4f\" (UID: \"b05370ff-d631-451a-918f-6689b00a0f4f\") " Dec 11 03:12:08 crc kubenswrapper[4824]: I1211 03:12:08.176098 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b05370ff-d631-451a-918f-6689b00a0f4f-host\") pod \"b05370ff-d631-451a-918f-6689b00a0f4f\" (UID: \"b05370ff-d631-451a-918f-6689b00a0f4f\") " Dec 11 03:12:08 crc kubenswrapper[4824]: I1211 03:12:08.178288 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b05370ff-d631-451a-918f-6689b00a0f4f-host" (OuterVolumeSpecName: "host") pod "b05370ff-d631-451a-918f-6689b00a0f4f" (UID: "b05370ff-d631-451a-918f-6689b00a0f4f"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 03:12:08 crc kubenswrapper[4824]: I1211 03:12:08.185528 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b05370ff-d631-451a-918f-6689b00a0f4f-kube-api-access-jx8nk" (OuterVolumeSpecName: "kube-api-access-jx8nk") pod "b05370ff-d631-451a-918f-6689b00a0f4f" (UID: "b05370ff-d631-451a-918f-6689b00a0f4f"). InnerVolumeSpecName "kube-api-access-jx8nk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 03:12:08 crc kubenswrapper[4824]: I1211 03:12:08.279888 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jx8nk\" (UniqueName: \"kubernetes.io/projected/b05370ff-d631-451a-918f-6689b00a0f4f-kube-api-access-jx8nk\") on node \"crc\" DevicePath \"\"" Dec 11 03:12:08 crc kubenswrapper[4824]: I1211 03:12:08.279943 4824 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b05370ff-d631-451a-918f-6689b00a0f4f-host\") on node \"crc\" DevicePath \"\"" Dec 11 03:12:08 crc kubenswrapper[4824]: I1211 03:12:08.653435 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b05370ff-d631-451a-918f-6689b00a0f4f" path="/var/lib/kubelet/pods/b05370ff-d631-451a-918f-6689b00a0f4f/volumes" Dec 11 03:12:08 crc kubenswrapper[4824]: I1211 03:12:08.916172 4824 scope.go:117] "RemoveContainer" containerID="4f434e4fb2c2bec59ede89a6559c8187e677b1336805793ad9f78deb9f74499b" Dec 11 03:12:08 crc kubenswrapper[4824]: I1211 03:12:08.916507 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2pxpn/crc-debug-djg86" Dec 11 03:12:22 crc kubenswrapper[4824]: I1211 03:12:22.632989 4824 scope.go:117] "RemoveContainer" containerID="fdad360b2681cc88a198f2fc151daab36d27e39aa44b3ec03ace588804769464" Dec 11 03:12:22 crc kubenswrapper[4824]: E1211 03:12:22.633740 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 03:12:34 crc kubenswrapper[4824]: I1211 03:12:34.634179 4824 scope.go:117] "RemoveContainer" containerID="fdad360b2681cc88a198f2fc151daab36d27e39aa44b3ec03ace588804769464" Dec 11 03:12:34 crc kubenswrapper[4824]: E1211 03:12:34.635108 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 03:12:37 crc kubenswrapper[4824]: I1211 03:12:37.664689 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-c88ccd6d6-bg6lk_bd669057-4834-42ed-83ea-f7a454c0a013/barbican-api/0.log" Dec 11 03:12:37 crc kubenswrapper[4824]: I1211 03:12:37.833608 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-c88ccd6d6-bg6lk_bd669057-4834-42ed-83ea-f7a454c0a013/barbican-api-log/0.log" Dec 11 03:12:37 crc kubenswrapper[4824]: I1211 03:12:37.881606 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-664bd5db46-xt8tr_cd747b97-a287-4df6-8b35-f30121c8c0b9/barbican-keystone-listener/0.log" Dec 11 03:12:37 crc kubenswrapper[4824]: I1211 03:12:37.954882 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-664bd5db46-xt8tr_cd747b97-a287-4df6-8b35-f30121c8c0b9/barbican-keystone-listener-log/0.log" Dec 11 03:12:38 crc kubenswrapper[4824]: I1211 03:12:38.040899 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-6967b7c86c-47gl7_33482b7a-6aaa-48c4-9a46-c3de05cacebe/barbican-worker/0.log" Dec 11 03:12:38 crc kubenswrapper[4824]: I1211 03:12:38.132596 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-6967b7c86c-47gl7_33482b7a-6aaa-48c4-9a46-c3de05cacebe/barbican-worker-log/0.log" Dec 11 03:12:38 crc kubenswrapper[4824]: I1211 03:12:38.701844 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-bsknw_1bf6cf36-d7eb-42be-aeee-c6824de8b6ba/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 11 03:12:38 crc kubenswrapper[4824]: I1211 03:12:38.787670 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_c65e7e08-e393-4cd6-b17e-3d9e81783e39/ceilometer-central-agent/0.log" Dec 11 03:12:38 crc kubenswrapper[4824]: I1211 03:12:38.894553 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_c65e7e08-e393-4cd6-b17e-3d9e81783e39/proxy-httpd/0.log" Dec 11 03:12:38 crc kubenswrapper[4824]: I1211 03:12:38.924184 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_c65e7e08-e393-4cd6-b17e-3d9e81783e39/ceilometer-notification-agent/0.log" Dec 11 03:12:38 crc kubenswrapper[4824]: I1211 03:12:38.926198 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_c65e7e08-e393-4cd6-b17e-3d9e81783e39/sg-core/0.log" Dec 11 03:12:39 crc kubenswrapper[4824]: I1211 03:12:39.101759 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_717f2bc0-db62-483f-83e6-4d2f247eb2d6/cinder-api-log/0.log" Dec 11 03:12:39 crc kubenswrapper[4824]: I1211 03:12:39.118301 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_717f2bc0-db62-483f-83e6-4d2f247eb2d6/cinder-api/0.log" Dec 11 03:12:39 crc kubenswrapper[4824]: I1211 03:12:39.287544 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_0dde14f9-85a8-4731-94ef-8199ea6d8a4b/cinder-scheduler/0.log" Dec 11 03:12:39 crc kubenswrapper[4824]: I1211 03:12:39.320252 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_0dde14f9-85a8-4731-94ef-8199ea6d8a4b/probe/0.log" Dec 11 03:12:39 crc kubenswrapper[4824]: I1211 03:12:39.349182 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-v4fz9_aeee6ae2-66ac-45e6-9a98-753649777abe/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 11 03:12:39 crc kubenswrapper[4824]: I1211 03:12:39.486712 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-4qkdg_6b2de157-739b-48d1-8bcb-585e4e6f475a/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 11 03:12:39 crc kubenswrapper[4824]: I1211 03:12:39.626448 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-mptxc_71c2c849-fc4f-44ac-92d4-75848ea56874/init/0.log" Dec 11 03:12:40 crc kubenswrapper[4824]: I1211 03:12:40.554355 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-mptxc_71c2c849-fc4f-44ac-92d4-75848ea56874/init/0.log" Dec 11 03:12:40 crc kubenswrapper[4824]: I1211 03:12:40.636856 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-mptxc_71c2c849-fc4f-44ac-92d4-75848ea56874/dnsmasq-dns/0.log" Dec 11 03:12:40 crc kubenswrapper[4824]: I1211 03:12:40.684695 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-47r5p_6c6c5ca9-ed78-4f51-a88b-12ec3df9efa2/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Dec 11 03:12:40 crc kubenswrapper[4824]: I1211 03:12:40.841092 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_ac4513fb-74e7-4bf2-8766-ef96066f6c13/glance-httpd/0.log" Dec 11 03:12:40 crc kubenswrapper[4824]: I1211 03:12:40.856216 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_ac4513fb-74e7-4bf2-8766-ef96066f6c13/glance-log/0.log" Dec 11 03:12:40 crc kubenswrapper[4824]: I1211 03:12:40.992488 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_e6776fbd-a091-4715-9942-4dafac0630e2/glance-httpd/0.log" Dec 11 03:12:41 crc kubenswrapper[4824]: I1211 03:12:41.042262 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_e6776fbd-a091-4715-9942-4dafac0630e2/glance-log/0.log" Dec 11 03:12:41 crc kubenswrapper[4824]: I1211 03:12:41.136767 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-74cc6c88d8-zvlk7_1da5eb01-d709-42ad-b5f1-e7ea6310cc9e/horizon/0.log" Dec 11 03:12:41 crc kubenswrapper[4824]: I1211 03:12:41.262684 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-lw68c_20ff7897-84b7-4c8c-b9f9-fc7721488ea7/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Dec 11 03:12:41 crc kubenswrapper[4824]: I1211 03:12:41.424253 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-9k74z_a786c754-b2c4-4f19-b082-e8d64d47ac44/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 11 03:12:41 crc kubenswrapper[4824]: I1211 03:12:41.506333 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-74cc6c88d8-zvlk7_1da5eb01-d709-42ad-b5f1-e7ea6310cc9e/horizon-log/0.log" Dec 11 03:12:41 crc kubenswrapper[4824]: I1211 03:12:41.614829 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-6c99457c66-dcp4p_55bc7333-3e36-4d58-9451-fe018b085c0b/keystone-api/0.log" Dec 11 03:12:41 crc kubenswrapper[4824]: I1211 03:12:41.659189 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29423701-hlhbx_a20415c4-3d24-4424-b32f-9e88eb6305e3/keystone-cron/0.log" Dec 11 03:12:41 crc kubenswrapper[4824]: I1211 03:12:41.715435 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_1ed2f6d5-a61f-4c5b-9b73-1089d13cf9b6/kube-state-metrics/0.log" Dec 11 03:12:41 crc kubenswrapper[4824]: I1211 03:12:41.817299 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-rs2vt_bc0c56ce-23bd-4883-b3dd-879d58c540b5/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Dec 11 03:12:42 crc kubenswrapper[4824]: I1211 03:12:42.177696 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5b766ddd67-lflrw_cde933a7-1f3a-4567-b6fd-7644a02cb745/neutron-httpd/0.log" Dec 11 03:12:42 crc kubenswrapper[4824]: I1211 03:12:42.221559 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5b766ddd67-lflrw_cde933a7-1f3a-4567-b6fd-7644a02cb745/neutron-api/0.log" Dec 11 03:12:42 crc kubenswrapper[4824]: I1211 03:12:42.372168 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-8rbcc_8736a409-d9d8-437e-965a-630f1ee6ae85/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Dec 11 03:12:42 crc kubenswrapper[4824]: I1211 03:12:42.909935 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_a36b8e48-3fc4-4132-a824-a0078424150f/nova-api-log/0.log" Dec 11 03:12:42 crc kubenswrapper[4824]: I1211 03:12:42.966406 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_1443d344-6dc0-4aab-96fd-522ae53f31cc/nova-cell0-conductor-conductor/0.log" Dec 11 03:12:43 crc kubenswrapper[4824]: I1211 03:12:43.163235 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_a36b8e48-3fc4-4132-a824-a0078424150f/nova-api-api/0.log" Dec 11 03:12:43 crc kubenswrapper[4824]: I1211 03:12:43.189833 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_fddad47b-d680-4be3-8bfa-a7708daef973/nova-cell1-conductor-conductor/0.log" Dec 11 03:12:43 crc kubenswrapper[4824]: I1211 03:12:43.330013 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_c246a287-b22f-405b-8a24-c50765f5b153/nova-cell1-novncproxy-novncproxy/0.log" Dec 11 03:12:43 crc kubenswrapper[4824]: I1211 03:12:43.408631 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-7x98s_f844f569-89fd-410c-b2e1-5f9883d9127a/nova-edpm-deployment-openstack-edpm-ipam/0.log" Dec 11 03:12:43 crc kubenswrapper[4824]: I1211 03:12:43.615263 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_e0e1ab61-fe6b-4b2e-aa01-2bf77e119ce4/nova-metadata-log/0.log" Dec 11 03:12:43 crc kubenswrapper[4824]: I1211 03:12:43.829972 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_f35efb32-a031-4c75-9358-3f5143335131/mysql-bootstrap/0.log" Dec 11 03:12:43 crc kubenswrapper[4824]: I1211 03:12:43.930793 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_bc3713cb-287a-4350-a147-e13a5151ac71/nova-scheduler-scheduler/0.log" Dec 11 03:12:44 crc kubenswrapper[4824]: I1211 03:12:44.056477 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_f35efb32-a031-4c75-9358-3f5143335131/mysql-bootstrap/0.log" Dec 11 03:12:44 crc kubenswrapper[4824]: I1211 03:12:44.102059 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_f35efb32-a031-4c75-9358-3f5143335131/galera/0.log" Dec 11 03:12:44 crc kubenswrapper[4824]: I1211 03:12:44.265065 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_6e5994f4-2c9e-4e8c-b7d3-84a15af846cd/mysql-bootstrap/0.log" Dec 11 03:12:44 crc kubenswrapper[4824]: I1211 03:12:44.500964 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_6e5994f4-2c9e-4e8c-b7d3-84a15af846cd/galera/0.log" Dec 11 03:12:44 crc kubenswrapper[4824]: I1211 03:12:44.503037 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_6e5994f4-2c9e-4e8c-b7d3-84a15af846cd/mysql-bootstrap/0.log" Dec 11 03:12:44 crc kubenswrapper[4824]: I1211 03:12:44.744702 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_a66f0fef-8984-4c67-b797-6762b404cd36/openstackclient/0.log" Dec 11 03:12:44 crc kubenswrapper[4824]: I1211 03:12:44.803770 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-779x4_148c69f8-3121-4a55-b3a6-be44b816f643/ovn-controller/0.log" Dec 11 03:12:44 crc kubenswrapper[4824]: I1211 03:12:44.918728 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-9fj44_d4fa84ff-af7c-4d69-9dfd-e4c1f20b0ce9/openstack-network-exporter/0.log" Dec 11 03:12:44 crc kubenswrapper[4824]: I1211 03:12:44.977446 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_e0e1ab61-fe6b-4b2e-aa01-2bf77e119ce4/nova-metadata-metadata/0.log" Dec 11 03:12:45 crc kubenswrapper[4824]: I1211 03:12:45.120639 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-xzkzd_17cc80fc-a51f-41e4-be0f-f593fc23476e/ovsdb-server-init/0.log" Dec 11 03:12:45 crc kubenswrapper[4824]: I1211 03:12:45.341193 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-xzkzd_17cc80fc-a51f-41e4-be0f-f593fc23476e/ovsdb-server/0.log" Dec 11 03:12:45 crc kubenswrapper[4824]: I1211 03:12:45.373984 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-xzkzd_17cc80fc-a51f-41e4-be0f-f593fc23476e/ovsdb-server-init/0.log" Dec 11 03:12:45 crc kubenswrapper[4824]: I1211 03:12:45.393866 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-xzkzd_17cc80fc-a51f-41e4-be0f-f593fc23476e/ovs-vswitchd/0.log" Dec 11 03:12:45 crc kubenswrapper[4824]: I1211 03:12:45.581039 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_964729d8-30e0-4d0b-ae8e-6f9bfc2536d0/openstack-network-exporter/0.log" Dec 11 03:12:45 crc kubenswrapper[4824]: I1211 03:12:45.614369 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_964729d8-30e0-4d0b-ae8e-6f9bfc2536d0/ovn-northd/0.log" Dec 11 03:12:45 crc kubenswrapper[4824]: I1211 03:12:45.670210 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-grms5_c8ca447b-581a-45ba-af85-a1a4ff83687f/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Dec 11 03:12:45 crc kubenswrapper[4824]: I1211 03:12:45.800941 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_068df61c-d193-4911-b077-fd80d25fa4de/ovsdbserver-nb/0.log" Dec 11 03:12:45 crc kubenswrapper[4824]: I1211 03:12:45.806387 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_068df61c-d193-4911-b077-fd80d25fa4de/openstack-network-exporter/0.log" Dec 11 03:12:46 crc kubenswrapper[4824]: I1211 03:12:46.212902 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_402ac70d-73e7-4697-ac46-a92bbbaf1aac/openstack-network-exporter/0.log" Dec 11 03:12:46 crc kubenswrapper[4824]: I1211 03:12:46.240761 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_402ac70d-73e7-4697-ac46-a92bbbaf1aac/ovsdbserver-sb/0.log" Dec 11 03:12:46 crc kubenswrapper[4824]: I1211 03:12:46.375683 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-78b68944cb-4p8xq_d4663f55-6344-4154-9df6-96a17ca59bdb/placement-api/0.log" Dec 11 03:12:46 crc kubenswrapper[4824]: I1211 03:12:46.460595 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-78b68944cb-4p8xq_d4663f55-6344-4154-9df6-96a17ca59bdb/placement-log/0.log" Dec 11 03:12:46 crc kubenswrapper[4824]: I1211 03:12:46.469418 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_89eddaec-a9d6-4b73-b654-2f176d45d162/setup-container/0.log" Dec 11 03:12:46 crc kubenswrapper[4824]: I1211 03:12:46.636098 4824 scope.go:117] "RemoveContainer" containerID="fdad360b2681cc88a198f2fc151daab36d27e39aa44b3ec03ace588804769464" Dec 11 03:12:46 crc kubenswrapper[4824]: E1211 03:12:46.636420 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 03:12:46 crc kubenswrapper[4824]: I1211 03:12:46.856398 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_89eddaec-a9d6-4b73-b654-2f176d45d162/setup-container/0.log" Dec 11 03:12:46 crc kubenswrapper[4824]: I1211 03:12:46.866758 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_8799d853-613d-4376-a372-7dedd968f264/setup-container/0.log" Dec 11 03:12:46 crc kubenswrapper[4824]: I1211 03:12:46.868999 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_89eddaec-a9d6-4b73-b654-2f176d45d162/rabbitmq/0.log" Dec 11 03:12:47 crc kubenswrapper[4824]: I1211 03:12:47.442585 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_8799d853-613d-4376-a372-7dedd968f264/setup-container/0.log" Dec 11 03:12:47 crc kubenswrapper[4824]: I1211 03:12:47.491488 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_8799d853-613d-4376-a372-7dedd968f264/rabbitmq/0.log" Dec 11 03:12:47 crc kubenswrapper[4824]: I1211 03:12:47.615016 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-wx57q_a24e9b38-d375-4b6b-a641-12aa6bf0fe18/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 11 03:12:47 crc kubenswrapper[4824]: I1211 03:12:47.688596 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-4292s_17d30a7d-e050-463d-ba60-bc8ae69cb21e/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Dec 11 03:12:47 crc kubenswrapper[4824]: I1211 03:12:47.896808 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-nl76s_09a89381-9670-4aff-965b-bd6d0d3e5ae7/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 11 03:12:47 crc kubenswrapper[4824]: I1211 03:12:47.967977 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-gk2qc_ab605768-f69e-4079-892a-a5b78e8dfc7d/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 11 03:12:48 crc kubenswrapper[4824]: I1211 03:12:48.095322 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-pkkbb_bf0de509-4023-4c82-b39a-f42d78d535a7/ssh-known-hosts-edpm-deployment/0.log" Dec 11 03:12:48 crc kubenswrapper[4824]: I1211 03:12:48.280037 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-6458cc7487-d2555_ead63438-3159-4a5e-889f-5f7c8a8d6a30/proxy-server/0.log" Dec 11 03:12:48 crc kubenswrapper[4824]: I1211 03:12:48.409011 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-zbgp5_5719adff-b313-426e-afd1-69986bdc81bc/swift-ring-rebalance/0.log" Dec 11 03:12:48 crc kubenswrapper[4824]: I1211 03:12:48.423260 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-6458cc7487-d2555_ead63438-3159-4a5e-889f-5f7c8a8d6a30/proxy-httpd/0.log" Dec 11 03:12:49 crc kubenswrapper[4824]: I1211 03:12:49.149984 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7913c5de-48cb-4e9a-8b9b-ee976476bab2/account-auditor/0.log" Dec 11 03:12:49 crc kubenswrapper[4824]: I1211 03:12:49.299539 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7913c5de-48cb-4e9a-8b9b-ee976476bab2/account-reaper/0.log" Dec 11 03:12:49 crc kubenswrapper[4824]: I1211 03:12:49.315872 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7913c5de-48cb-4e9a-8b9b-ee976476bab2/account-replicator/0.log" Dec 11 03:12:49 crc kubenswrapper[4824]: I1211 03:12:49.337558 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7913c5de-48cb-4e9a-8b9b-ee976476bab2/account-server/0.log" Dec 11 03:12:49 crc kubenswrapper[4824]: I1211 03:12:49.435234 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7913c5de-48cb-4e9a-8b9b-ee976476bab2/container-auditor/0.log" Dec 11 03:12:49 crc kubenswrapper[4824]: I1211 03:12:49.512737 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7913c5de-48cb-4e9a-8b9b-ee976476bab2/container-server/0.log" Dec 11 03:12:49 crc kubenswrapper[4824]: I1211 03:12:49.553768 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7913c5de-48cb-4e9a-8b9b-ee976476bab2/container-updater/0.log" Dec 11 03:12:49 crc kubenswrapper[4824]: I1211 03:12:49.561843 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7913c5de-48cb-4e9a-8b9b-ee976476bab2/container-replicator/0.log" Dec 11 03:12:49 crc kubenswrapper[4824]: I1211 03:12:49.625352 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7913c5de-48cb-4e9a-8b9b-ee976476bab2/object-auditor/0.log" Dec 11 03:12:49 crc kubenswrapper[4824]: I1211 03:12:49.696977 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7913c5de-48cb-4e9a-8b9b-ee976476bab2/object-expirer/0.log" Dec 11 03:12:49 crc kubenswrapper[4824]: I1211 03:12:49.807389 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7913c5de-48cb-4e9a-8b9b-ee976476bab2/object-server/0.log" Dec 11 03:12:49 crc kubenswrapper[4824]: I1211 03:12:49.821050 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7913c5de-48cb-4e9a-8b9b-ee976476bab2/object-replicator/0.log" Dec 11 03:12:49 crc kubenswrapper[4824]: I1211 03:12:49.875975 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7913c5de-48cb-4e9a-8b9b-ee976476bab2/rsync/0.log" Dec 11 03:12:49 crc kubenswrapper[4824]: I1211 03:12:49.884680 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7913c5de-48cb-4e9a-8b9b-ee976476bab2/object-updater/0.log" Dec 11 03:12:50 crc kubenswrapper[4824]: I1211 03:12:50.164635 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7913c5de-48cb-4e9a-8b9b-ee976476bab2/swift-recon-cron/0.log" Dec 11 03:12:50 crc kubenswrapper[4824]: I1211 03:12:50.381572 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-czhq8_f7ed3232-8b86-4804-85e1-0b87715ed1f2/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Dec 11 03:12:50 crc kubenswrapper[4824]: I1211 03:12:50.431184 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_ab73a048-199c-464e-98ab-2e87063161b6/tempest-tests-tempest-tests-runner/0.log" Dec 11 03:12:50 crc kubenswrapper[4824]: I1211 03:12:50.537728 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_1283daba-11e7-4232-81dd-c4316478f297/test-operator-logs-container/0.log" Dec 11 03:12:50 crc kubenswrapper[4824]: I1211 03:12:50.660586 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-t4pfx_f694423f-b8e7-49bd-9032-4b29a64fda4e/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 11 03:12:56 crc kubenswrapper[4824]: I1211 03:12:56.994838 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_709f3d23-a475-443a-8787-7f018c486b56/memcached/0.log" Dec 11 03:12:57 crc kubenswrapper[4824]: I1211 03:12:57.633875 4824 scope.go:117] "RemoveContainer" containerID="fdad360b2681cc88a198f2fc151daab36d27e39aa44b3ec03ace588804769464" Dec 11 03:12:57 crc kubenswrapper[4824]: E1211 03:12:57.634060 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 03:13:10 crc kubenswrapper[4824]: I1211 03:13:10.334471 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-4vgr2"] Dec 11 03:13:10 crc kubenswrapper[4824]: E1211 03:13:10.335754 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b05370ff-d631-451a-918f-6689b00a0f4f" containerName="container-00" Dec 11 03:13:10 crc kubenswrapper[4824]: I1211 03:13:10.335777 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="b05370ff-d631-451a-918f-6689b00a0f4f" containerName="container-00" Dec 11 03:13:10 crc kubenswrapper[4824]: I1211 03:13:10.336103 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="b05370ff-d631-451a-918f-6689b00a0f4f" containerName="container-00" Dec 11 03:13:10 crc kubenswrapper[4824]: I1211 03:13:10.338082 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4vgr2" Dec 11 03:13:10 crc kubenswrapper[4824]: I1211 03:13:10.346417 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p9qcd\" (UniqueName: \"kubernetes.io/projected/ad4e10f1-aa90-4ae8-a1f6-8de2202ebbc6-kube-api-access-p9qcd\") pod \"certified-operators-4vgr2\" (UID: \"ad4e10f1-aa90-4ae8-a1f6-8de2202ebbc6\") " pod="openshift-marketplace/certified-operators-4vgr2" Dec 11 03:13:10 crc kubenswrapper[4824]: I1211 03:13:10.346485 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad4e10f1-aa90-4ae8-a1f6-8de2202ebbc6-catalog-content\") pod \"certified-operators-4vgr2\" (UID: \"ad4e10f1-aa90-4ae8-a1f6-8de2202ebbc6\") " pod="openshift-marketplace/certified-operators-4vgr2" Dec 11 03:13:10 crc kubenswrapper[4824]: I1211 03:13:10.346632 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad4e10f1-aa90-4ae8-a1f6-8de2202ebbc6-utilities\") pod \"certified-operators-4vgr2\" (UID: \"ad4e10f1-aa90-4ae8-a1f6-8de2202ebbc6\") " pod="openshift-marketplace/certified-operators-4vgr2" Dec 11 03:13:10 crc kubenswrapper[4824]: I1211 03:13:10.359186 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4vgr2"] Dec 11 03:13:10 crc kubenswrapper[4824]: I1211 03:13:10.449202 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p9qcd\" (UniqueName: \"kubernetes.io/projected/ad4e10f1-aa90-4ae8-a1f6-8de2202ebbc6-kube-api-access-p9qcd\") pod \"certified-operators-4vgr2\" (UID: \"ad4e10f1-aa90-4ae8-a1f6-8de2202ebbc6\") " pod="openshift-marketplace/certified-operators-4vgr2" Dec 11 03:13:10 crc kubenswrapper[4824]: I1211 03:13:10.449257 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad4e10f1-aa90-4ae8-a1f6-8de2202ebbc6-catalog-content\") pod \"certified-operators-4vgr2\" (UID: \"ad4e10f1-aa90-4ae8-a1f6-8de2202ebbc6\") " pod="openshift-marketplace/certified-operators-4vgr2" Dec 11 03:13:10 crc kubenswrapper[4824]: I1211 03:13:10.449309 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad4e10f1-aa90-4ae8-a1f6-8de2202ebbc6-utilities\") pod \"certified-operators-4vgr2\" (UID: \"ad4e10f1-aa90-4ae8-a1f6-8de2202ebbc6\") " pod="openshift-marketplace/certified-operators-4vgr2" Dec 11 03:13:10 crc kubenswrapper[4824]: I1211 03:13:10.449826 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad4e10f1-aa90-4ae8-a1f6-8de2202ebbc6-utilities\") pod \"certified-operators-4vgr2\" (UID: \"ad4e10f1-aa90-4ae8-a1f6-8de2202ebbc6\") " pod="openshift-marketplace/certified-operators-4vgr2" Dec 11 03:13:10 crc kubenswrapper[4824]: I1211 03:13:10.449973 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad4e10f1-aa90-4ae8-a1f6-8de2202ebbc6-catalog-content\") pod \"certified-operators-4vgr2\" (UID: \"ad4e10f1-aa90-4ae8-a1f6-8de2202ebbc6\") " pod="openshift-marketplace/certified-operators-4vgr2" Dec 11 03:13:10 crc kubenswrapper[4824]: I1211 03:13:10.482212 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p9qcd\" (UniqueName: \"kubernetes.io/projected/ad4e10f1-aa90-4ae8-a1f6-8de2202ebbc6-kube-api-access-p9qcd\") pod \"certified-operators-4vgr2\" (UID: \"ad4e10f1-aa90-4ae8-a1f6-8de2202ebbc6\") " pod="openshift-marketplace/certified-operators-4vgr2" Dec 11 03:13:10 crc kubenswrapper[4824]: I1211 03:13:10.633224 4824 scope.go:117] "RemoveContainer" containerID="fdad360b2681cc88a198f2fc151daab36d27e39aa44b3ec03ace588804769464" Dec 11 03:13:10 crc kubenswrapper[4824]: E1211 03:13:10.633559 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 03:13:10 crc kubenswrapper[4824]: I1211 03:13:10.666075 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4vgr2" Dec 11 03:13:11 crc kubenswrapper[4824]: I1211 03:13:11.196012 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4vgr2"] Dec 11 03:13:11 crc kubenswrapper[4824]: I1211 03:13:11.481282 4824 generic.go:334] "Generic (PLEG): container finished" podID="ad4e10f1-aa90-4ae8-a1f6-8de2202ebbc6" containerID="60d24b793410d78bf8bae6e8af046ae57b0b2051471ac238486745dcf6b0ea00" exitCode=0 Dec 11 03:13:11 crc kubenswrapper[4824]: I1211 03:13:11.481324 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4vgr2" event={"ID":"ad4e10f1-aa90-4ae8-a1f6-8de2202ebbc6","Type":"ContainerDied","Data":"60d24b793410d78bf8bae6e8af046ae57b0b2051471ac238486745dcf6b0ea00"} Dec 11 03:13:11 crc kubenswrapper[4824]: I1211 03:13:11.481526 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4vgr2" event={"ID":"ad4e10f1-aa90-4ae8-a1f6-8de2202ebbc6","Type":"ContainerStarted","Data":"3e99d26edbaffedb2026739f575163e2c918a608b23e5fa25b71e2f229e75bc0"} Dec 11 03:13:11 crc kubenswrapper[4824]: I1211 03:13:11.483015 4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 11 03:13:13 crc kubenswrapper[4824]: I1211 03:13:13.503022 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4vgr2" event={"ID":"ad4e10f1-aa90-4ae8-a1f6-8de2202ebbc6","Type":"ContainerStarted","Data":"bf07fe44178d7dd29c7c5b49bb8b9727c8532af79d1d9c8e2dad1019bbaf5049"} Dec 11 03:13:14 crc kubenswrapper[4824]: I1211 03:13:14.518374 4824 generic.go:334] "Generic (PLEG): container finished" podID="ad4e10f1-aa90-4ae8-a1f6-8de2202ebbc6" containerID="bf07fe44178d7dd29c7c5b49bb8b9727c8532af79d1d9c8e2dad1019bbaf5049" exitCode=0 Dec 11 03:13:14 crc kubenswrapper[4824]: I1211 03:13:14.518489 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4vgr2" event={"ID":"ad4e10f1-aa90-4ae8-a1f6-8de2202ebbc6","Type":"ContainerDied","Data":"bf07fe44178d7dd29c7c5b49bb8b9727c8532af79d1d9c8e2dad1019bbaf5049"} Dec 11 03:13:16 crc kubenswrapper[4824]: I1211 03:13:16.546546 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4vgr2" event={"ID":"ad4e10f1-aa90-4ae8-a1f6-8de2202ebbc6","Type":"ContainerStarted","Data":"a42b1d548f9d14b9cfcce83870a905b090f8cb7a4ee077fe7e3e2c99495f5819"} Dec 11 03:13:16 crc kubenswrapper[4824]: I1211 03:13:16.584049 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-4vgr2" podStartSLOduration=3.09293385 podStartE2EDuration="6.584030745s" podCreationTimestamp="2025-12-11 03:13:10 +0000 UTC" firstStartedPulling="2025-12-11 03:13:11.482781473 +0000 UTC m=+4333.171818852" lastFinishedPulling="2025-12-11 03:13:14.973878368 +0000 UTC m=+4336.662915747" observedRunningTime="2025-12-11 03:13:16.57634846 +0000 UTC m=+4338.265385839" watchObservedRunningTime="2025-12-11 03:13:16.584030745 +0000 UTC m=+4338.273068134" Dec 11 03:13:18 crc kubenswrapper[4824]: I1211 03:13:18.836151 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05vd4cz_487cf0fc-4368-4f0e-878a-d3cd3d1e54f4/util/0.log" Dec 11 03:13:18 crc kubenswrapper[4824]: I1211 03:13:18.949988 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05vd4cz_487cf0fc-4368-4f0e-878a-d3cd3d1e54f4/util/0.log" Dec 11 03:13:19 crc kubenswrapper[4824]: I1211 03:13:19.024165 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05vd4cz_487cf0fc-4368-4f0e-878a-d3cd3d1e54f4/pull/0.log" Dec 11 03:13:19 crc kubenswrapper[4824]: I1211 03:13:19.025965 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05vd4cz_487cf0fc-4368-4f0e-878a-d3cd3d1e54f4/pull/0.log" Dec 11 03:13:19 crc kubenswrapper[4824]: I1211 03:13:19.196409 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05vd4cz_487cf0fc-4368-4f0e-878a-d3cd3d1e54f4/util/0.log" Dec 11 03:13:19 crc kubenswrapper[4824]: I1211 03:13:19.244343 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05vd4cz_487cf0fc-4368-4f0e-878a-d3cd3d1e54f4/extract/0.log" Dec 11 03:13:19 crc kubenswrapper[4824]: I1211 03:13:19.244790 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05vd4cz_487cf0fc-4368-4f0e-878a-d3cd3d1e54f4/pull/0.log" Dec 11 03:13:19 crc kubenswrapper[4824]: I1211 03:13:19.442290 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-r8tnj_df6864cf-9559-4398-bb3c-0f58f4c45563/manager/0.log" Dec 11 03:13:19 crc kubenswrapper[4824]: I1211 03:13:19.461736 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-r8tnj_df6864cf-9559-4398-bb3c-0f58f4c45563/kube-rbac-proxy/0.log" Dec 11 03:13:19 crc kubenswrapper[4824]: I1211 03:13:19.516242 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6c677c69b-8hvxh_0b7b142b-04e2-4bb3-98ff-ed3b4da63746/kube-rbac-proxy/0.log" Dec 11 03:13:19 crc kubenswrapper[4824]: I1211 03:13:19.625531 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6c677c69b-8hvxh_0b7b142b-04e2-4bb3-98ff-ed3b4da63746/manager/0.log" Dec 11 03:13:19 crc kubenswrapper[4824]: I1211 03:13:19.715914 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-697fb699cf-4b2cb_6743e907-9570-4dde-bb99-83e67202b224/kube-rbac-proxy/0.log" Dec 11 03:13:19 crc kubenswrapper[4824]: I1211 03:13:19.824053 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-697fb699cf-4b2cb_6743e907-9570-4dde-bb99-83e67202b224/manager/0.log" Dec 11 03:13:19 crc kubenswrapper[4824]: I1211 03:13:19.878996 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5697bb5779-4cjld_0eeea3b6-c514-4547-a2e0-52870ef572c6/kube-rbac-proxy/0.log" Dec 11 03:13:19 crc kubenswrapper[4824]: I1211 03:13:19.963247 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5697bb5779-4cjld_0eeea3b6-c514-4547-a2e0-52870ef572c6/manager/0.log" Dec 11 03:13:20 crc kubenswrapper[4824]: I1211 03:13:20.077980 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-l9jrd_db4453b9-ead1-45fc-91c5-d90541fa9c78/kube-rbac-proxy/0.log" Dec 11 03:13:20 crc kubenswrapper[4824]: I1211 03:13:20.107733 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-l9jrd_db4453b9-ead1-45fc-91c5-d90541fa9c78/manager/0.log" Dec 11 03:13:20 crc kubenswrapper[4824]: I1211 03:13:20.210535 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-5qp6v_1fc0b6e3-15c9-4f14-90f8-3f39a04ebc2c/kube-rbac-proxy/0.log" Dec 11 03:13:20 crc kubenswrapper[4824]: I1211 03:13:20.255768 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-5qp6v_1fc0b6e3-15c9-4f14-90f8-3f39a04ebc2c/manager/0.log" Dec 11 03:13:20 crc kubenswrapper[4824]: I1211 03:13:20.374796 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-78d48bff9d-j68mf_f219b9bf-34b1-48c4-96c8-20580b6ec9a4/kube-rbac-proxy/0.log" Dec 11 03:13:20 crc kubenswrapper[4824]: I1211 03:13:20.508504 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-967d97867-5tpj5_aea47aaf-ee88-45a0-87b7-375b9f656828/kube-rbac-proxy/0.log" Dec 11 03:13:20 crc kubenswrapper[4824]: I1211 03:13:20.601100 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-967d97867-5tpj5_aea47aaf-ee88-45a0-87b7-375b9f656828/manager/0.log" Dec 11 03:13:20 crc kubenswrapper[4824]: I1211 03:13:20.652561 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-78d48bff9d-j68mf_f219b9bf-34b1-48c4-96c8-20580b6ec9a4/manager/0.log" Dec 11 03:13:20 crc kubenswrapper[4824]: I1211 03:13:20.666235 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-4vgr2" Dec 11 03:13:20 crc kubenswrapper[4824]: I1211 03:13:20.666491 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-4vgr2" Dec 11 03:13:20 crc kubenswrapper[4824]: I1211 03:13:20.714524 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-4vgr2" Dec 11 03:13:20 crc kubenswrapper[4824]: I1211 03:13:20.730701 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-4np4x_cec1f29f-a1db-495f-b59c-e9308a9b53a2/kube-rbac-proxy/0.log" Dec 11 03:13:20 crc kubenswrapper[4824]: I1211 03:13:20.866002 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-4np4x_cec1f29f-a1db-495f-b59c-e9308a9b53a2/manager/0.log" Dec 11 03:13:20 crc kubenswrapper[4824]: I1211 03:13:20.946819 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5b5fd79c9c-w24mx_46bf5c63-42b4-45bd-ade3-446a4b2e4f6e/kube-rbac-proxy/0.log" Dec 11 03:13:20 crc kubenswrapper[4824]: I1211 03:13:20.951573 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5b5fd79c9c-w24mx_46bf5c63-42b4-45bd-ade3-446a4b2e4f6e/manager/0.log" Dec 11 03:13:21 crc kubenswrapper[4824]: I1211 03:13:21.057512 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-79c8c4686c-hpjng_fd3368a6-77e2-4c8f-b535-1d602b794d7d/kube-rbac-proxy/0.log" Dec 11 03:13:21 crc kubenswrapper[4824]: I1211 03:13:21.137894 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-79c8c4686c-hpjng_fd3368a6-77e2-4c8f-b535-1d602b794d7d/manager/0.log" Dec 11 03:13:21 crc kubenswrapper[4824]: I1211 03:13:21.272404 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-lmszx_dca85063-ba7a-45ac-ab88-5498a44834ba/kube-rbac-proxy/0.log" Dec 11 03:13:21 crc kubenswrapper[4824]: I1211 03:13:21.326464 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-lmszx_dca85063-ba7a-45ac-ab88-5498a44834ba/manager/0.log" Dec 11 03:13:21 crc kubenswrapper[4824]: I1211 03:13:21.384418 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-zkz72_b206f71d-32cd-4b7d-ae38-a5125ab481dd/kube-rbac-proxy/0.log" Dec 11 03:13:21 crc kubenswrapper[4824]: I1211 03:13:21.541420 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-zkz72_b206f71d-32cd-4b7d-ae38-a5125ab481dd/manager/0.log" Dec 11 03:13:21 crc kubenswrapper[4824]: I1211 03:13:21.622162 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-qqpf2_f1183bb7-356d-43b7-8ee7-55ab60dcbe54/kube-rbac-proxy/0.log" Dec 11 03:13:21 crc kubenswrapper[4824]: I1211 03:13:21.643294 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-4vgr2" Dec 11 03:13:21 crc kubenswrapper[4824]: I1211 03:13:21.661131 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-qqpf2_f1183bb7-356d-43b7-8ee7-55ab60dcbe54/manager/0.log" Dec 11 03:13:21 crc kubenswrapper[4824]: I1211 03:13:21.705207 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4vgr2"] Dec 11 03:13:21 crc kubenswrapper[4824]: I1211 03:13:21.787312 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-84b575879f4k8rk_cb370fee-74a4-4f43-98a9-3df8c6e61335/kube-rbac-proxy/0.log" Dec 11 03:13:21 crc kubenswrapper[4824]: I1211 03:13:21.797792 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-84b575879f4k8rk_cb370fee-74a4-4f43-98a9-3df8c6e61335/manager/0.log" Dec 11 03:13:22 crc kubenswrapper[4824]: I1211 03:13:22.101201 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-nb526_ab26e0f0-f195-481f-84bd-0a7bca8d0432/registry-server/0.log" Dec 11 03:13:22 crc kubenswrapper[4824]: I1211 03:13:22.217238 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-7795b79658-tzqvv_52218f50-e21e-4566-b323-9b5ece7a5e35/operator/0.log" Dec 11 03:13:22 crc kubenswrapper[4824]: I1211 03:13:22.307102 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-r2bsg_4d26d90a-05a8-4960-b2ca-12fa220383b7/kube-rbac-proxy/0.log" Dec 11 03:13:22 crc kubenswrapper[4824]: I1211 03:13:22.386498 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-r2bsg_4d26d90a-05a8-4960-b2ca-12fa220383b7/manager/0.log" Dec 11 03:13:22 crc kubenswrapper[4824]: I1211 03:13:22.492531 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-xx9gc_10325d55-7efd-49e8-95e9-46d68c7cd7ae/kube-rbac-proxy/0.log" Dec 11 03:13:22 crc kubenswrapper[4824]: I1211 03:13:22.565588 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-xx9gc_10325d55-7efd-49e8-95e9-46d68c7cd7ae/manager/0.log" Dec 11 03:13:22 crc kubenswrapper[4824]: I1211 03:13:22.637968 4824 scope.go:117] "RemoveContainer" containerID="fdad360b2681cc88a198f2fc151daab36d27e39aa44b3ec03ace588804769464" Dec 11 03:13:22 crc kubenswrapper[4824]: E1211 03:13:22.638399 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 03:13:22 crc kubenswrapper[4824]: I1211 03:13:22.698320 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-wsjbh_34bc4184-9cfb-4278-a9b9-80bf2f1ae612/operator/0.log" Dec 11 03:13:22 crc kubenswrapper[4824]: I1211 03:13:22.824158 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-9d58d64bc-fmvm7_8047af9c-3a6d-4b2e-a088-df967e42671c/kube-rbac-proxy/0.log" Dec 11 03:13:22 crc kubenswrapper[4824]: I1211 03:13:22.846455 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-56cf947496-nhjjr_7180c14a-0c3a-4267-b122-5c30c5685d60/manager/0.log" Dec 11 03:13:22 crc kubenswrapper[4824]: I1211 03:13:22.917358 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-9d58d64bc-fmvm7_8047af9c-3a6d-4b2e-a088-df967e42671c/manager/0.log" Dec 11 03:13:22 crc kubenswrapper[4824]: I1211 03:13:22.935069 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-58d5ff84df-qhhzn_aff71eda-47bc-4b47-89df-fe9bf8357fba/kube-rbac-proxy/0.log" Dec 11 03:13:23 crc kubenswrapper[4824]: I1211 03:13:23.085105 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-58d5ff84df-qhhzn_aff71eda-47bc-4b47-89df-fe9bf8357fba/manager/0.log" Dec 11 03:13:23 crc kubenswrapper[4824]: I1211 03:13:23.110471 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-4zfbk_cf19e8a0-8f29-48a4-b715-f499644fc896/kube-rbac-proxy/0.log" Dec 11 03:13:23 crc kubenswrapper[4824]: I1211 03:13:23.144645 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-4zfbk_cf19e8a0-8f29-48a4-b715-f499644fc896/manager/0.log" Dec 11 03:13:23 crc kubenswrapper[4824]: I1211 03:13:23.254452 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-75944c9b7-h6bgf_9793f7fe-2e1b-46c0-985e-1f77733f925f/kube-rbac-proxy/0.log" Dec 11 03:13:23 crc kubenswrapper[4824]: I1211 03:13:23.259863 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-75944c9b7-h6bgf_9793f7fe-2e1b-46c0-985e-1f77733f925f/manager/0.log" Dec 11 03:13:23 crc kubenswrapper[4824]: I1211 03:13:23.607168 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-4vgr2" podUID="ad4e10f1-aa90-4ae8-a1f6-8de2202ebbc6" containerName="registry-server" containerID="cri-o://a42b1d548f9d14b9cfcce83870a905b090f8cb7a4ee077fe7e3e2c99495f5819" gracePeriod=2 Dec 11 03:13:24 crc kubenswrapper[4824]: I1211 03:13:24.590158 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4vgr2" Dec 11 03:13:24 crc kubenswrapper[4824]: I1211 03:13:24.624528 4824 generic.go:334] "Generic (PLEG): container finished" podID="ad4e10f1-aa90-4ae8-a1f6-8de2202ebbc6" containerID="a42b1d548f9d14b9cfcce83870a905b090f8cb7a4ee077fe7e3e2c99495f5819" exitCode=0 Dec 11 03:13:24 crc kubenswrapper[4824]: I1211 03:13:24.624612 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4vgr2" event={"ID":"ad4e10f1-aa90-4ae8-a1f6-8de2202ebbc6","Type":"ContainerDied","Data":"a42b1d548f9d14b9cfcce83870a905b090f8cb7a4ee077fe7e3e2c99495f5819"} Dec 11 03:13:24 crc kubenswrapper[4824]: I1211 03:13:24.624655 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4vgr2" event={"ID":"ad4e10f1-aa90-4ae8-a1f6-8de2202ebbc6","Type":"ContainerDied","Data":"3e99d26edbaffedb2026739f575163e2c918a608b23e5fa25b71e2f229e75bc0"} Dec 11 03:13:24 crc kubenswrapper[4824]: I1211 03:13:24.624683 4824 scope.go:117] "RemoveContainer" containerID="a42b1d548f9d14b9cfcce83870a905b090f8cb7a4ee077fe7e3e2c99495f5819" Dec 11 03:13:24 crc kubenswrapper[4824]: I1211 03:13:24.624881 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4vgr2" Dec 11 03:13:24 crc kubenswrapper[4824]: I1211 03:13:24.647313 4824 scope.go:117] "RemoveContainer" containerID="bf07fe44178d7dd29c7c5b49bb8b9727c8532af79d1d9c8e2dad1019bbaf5049" Dec 11 03:13:24 crc kubenswrapper[4824]: I1211 03:13:24.691935 4824 scope.go:117] "RemoveContainer" containerID="60d24b793410d78bf8bae6e8af046ae57b0b2051471ac238486745dcf6b0ea00" Dec 11 03:13:24 crc kubenswrapper[4824]: I1211 03:13:24.705347 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad4e10f1-aa90-4ae8-a1f6-8de2202ebbc6-utilities\") pod \"ad4e10f1-aa90-4ae8-a1f6-8de2202ebbc6\" (UID: \"ad4e10f1-aa90-4ae8-a1f6-8de2202ebbc6\") " Dec 11 03:13:24 crc kubenswrapper[4824]: I1211 03:13:24.705402 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p9qcd\" (UniqueName: \"kubernetes.io/projected/ad4e10f1-aa90-4ae8-a1f6-8de2202ebbc6-kube-api-access-p9qcd\") pod \"ad4e10f1-aa90-4ae8-a1f6-8de2202ebbc6\" (UID: \"ad4e10f1-aa90-4ae8-a1f6-8de2202ebbc6\") " Dec 11 03:13:24 crc kubenswrapper[4824]: I1211 03:13:24.705638 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad4e10f1-aa90-4ae8-a1f6-8de2202ebbc6-catalog-content\") pod \"ad4e10f1-aa90-4ae8-a1f6-8de2202ebbc6\" (UID: \"ad4e10f1-aa90-4ae8-a1f6-8de2202ebbc6\") " Dec 11 03:13:24 crc kubenswrapper[4824]: I1211 03:13:24.706260 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ad4e10f1-aa90-4ae8-a1f6-8de2202ebbc6-utilities" (OuterVolumeSpecName: "utilities") pod "ad4e10f1-aa90-4ae8-a1f6-8de2202ebbc6" (UID: "ad4e10f1-aa90-4ae8-a1f6-8de2202ebbc6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 03:13:24 crc kubenswrapper[4824]: I1211 03:13:24.714484 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad4e10f1-aa90-4ae8-a1f6-8de2202ebbc6-kube-api-access-p9qcd" (OuterVolumeSpecName: "kube-api-access-p9qcd") pod "ad4e10f1-aa90-4ae8-a1f6-8de2202ebbc6" (UID: "ad4e10f1-aa90-4ae8-a1f6-8de2202ebbc6"). InnerVolumeSpecName "kube-api-access-p9qcd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 03:13:24 crc kubenswrapper[4824]: I1211 03:13:24.728354 4824 scope.go:117] "RemoveContainer" containerID="a42b1d548f9d14b9cfcce83870a905b090f8cb7a4ee077fe7e3e2c99495f5819" Dec 11 03:13:24 crc kubenswrapper[4824]: E1211 03:13:24.728782 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a42b1d548f9d14b9cfcce83870a905b090f8cb7a4ee077fe7e3e2c99495f5819\": container with ID starting with a42b1d548f9d14b9cfcce83870a905b090f8cb7a4ee077fe7e3e2c99495f5819 not found: ID does not exist" containerID="a42b1d548f9d14b9cfcce83870a905b090f8cb7a4ee077fe7e3e2c99495f5819" Dec 11 03:13:24 crc kubenswrapper[4824]: I1211 03:13:24.728837 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a42b1d548f9d14b9cfcce83870a905b090f8cb7a4ee077fe7e3e2c99495f5819"} err="failed to get container status \"a42b1d548f9d14b9cfcce83870a905b090f8cb7a4ee077fe7e3e2c99495f5819\": rpc error: code = NotFound desc = could not find container \"a42b1d548f9d14b9cfcce83870a905b090f8cb7a4ee077fe7e3e2c99495f5819\": container with ID starting with a42b1d548f9d14b9cfcce83870a905b090f8cb7a4ee077fe7e3e2c99495f5819 not found: ID does not exist" Dec 11 03:13:24 crc kubenswrapper[4824]: I1211 03:13:24.728869 4824 scope.go:117] "RemoveContainer" containerID="bf07fe44178d7dd29c7c5b49bb8b9727c8532af79d1d9c8e2dad1019bbaf5049" Dec 11 03:13:24 crc kubenswrapper[4824]: E1211 03:13:24.729206 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bf07fe44178d7dd29c7c5b49bb8b9727c8532af79d1d9c8e2dad1019bbaf5049\": container with ID starting with bf07fe44178d7dd29c7c5b49bb8b9727c8532af79d1d9c8e2dad1019bbaf5049 not found: ID does not exist" containerID="bf07fe44178d7dd29c7c5b49bb8b9727c8532af79d1d9c8e2dad1019bbaf5049" Dec 11 03:13:24 crc kubenswrapper[4824]: I1211 03:13:24.729231 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf07fe44178d7dd29c7c5b49bb8b9727c8532af79d1d9c8e2dad1019bbaf5049"} err="failed to get container status \"bf07fe44178d7dd29c7c5b49bb8b9727c8532af79d1d9c8e2dad1019bbaf5049\": rpc error: code = NotFound desc = could not find container \"bf07fe44178d7dd29c7c5b49bb8b9727c8532af79d1d9c8e2dad1019bbaf5049\": container with ID starting with bf07fe44178d7dd29c7c5b49bb8b9727c8532af79d1d9c8e2dad1019bbaf5049 not found: ID does not exist" Dec 11 03:13:24 crc kubenswrapper[4824]: I1211 03:13:24.729247 4824 scope.go:117] "RemoveContainer" containerID="60d24b793410d78bf8bae6e8af046ae57b0b2051471ac238486745dcf6b0ea00" Dec 11 03:13:24 crc kubenswrapper[4824]: E1211 03:13:24.729438 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"60d24b793410d78bf8bae6e8af046ae57b0b2051471ac238486745dcf6b0ea00\": container with ID starting with 60d24b793410d78bf8bae6e8af046ae57b0b2051471ac238486745dcf6b0ea00 not found: ID does not exist" containerID="60d24b793410d78bf8bae6e8af046ae57b0b2051471ac238486745dcf6b0ea00" Dec 11 03:13:24 crc kubenswrapper[4824]: I1211 03:13:24.729463 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60d24b793410d78bf8bae6e8af046ae57b0b2051471ac238486745dcf6b0ea00"} err="failed to get container status \"60d24b793410d78bf8bae6e8af046ae57b0b2051471ac238486745dcf6b0ea00\": rpc error: code = NotFound desc = could not find container \"60d24b793410d78bf8bae6e8af046ae57b0b2051471ac238486745dcf6b0ea00\": container with ID starting with 60d24b793410d78bf8bae6e8af046ae57b0b2051471ac238486745dcf6b0ea00 not found: ID does not exist" Dec 11 03:13:24 crc kubenswrapper[4824]: I1211 03:13:24.775335 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ad4e10f1-aa90-4ae8-a1f6-8de2202ebbc6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ad4e10f1-aa90-4ae8-a1f6-8de2202ebbc6" (UID: "ad4e10f1-aa90-4ae8-a1f6-8de2202ebbc6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 03:13:24 crc kubenswrapper[4824]: I1211 03:13:24.807790 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad4e10f1-aa90-4ae8-a1f6-8de2202ebbc6-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 03:13:24 crc kubenswrapper[4824]: I1211 03:13:24.807817 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad4e10f1-aa90-4ae8-a1f6-8de2202ebbc6-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 03:13:24 crc kubenswrapper[4824]: I1211 03:13:24.807826 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p9qcd\" (UniqueName: \"kubernetes.io/projected/ad4e10f1-aa90-4ae8-a1f6-8de2202ebbc6-kube-api-access-p9qcd\") on node \"crc\" DevicePath \"\"" Dec 11 03:13:25 crc kubenswrapper[4824]: I1211 03:13:25.534144 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4vgr2"] Dec 11 03:13:25 crc kubenswrapper[4824]: I1211 03:13:25.543773 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-4vgr2"] Dec 11 03:13:26 crc kubenswrapper[4824]: I1211 03:13:26.643612 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad4e10f1-aa90-4ae8-a1f6-8de2202ebbc6" path="/var/lib/kubelet/pods/ad4e10f1-aa90-4ae8-a1f6-8de2202ebbc6/volumes" Dec 11 03:13:36 crc kubenswrapper[4824]: I1211 03:13:36.633804 4824 scope.go:117] "RemoveContainer" containerID="fdad360b2681cc88a198f2fc151daab36d27e39aa44b3ec03ace588804769464" Dec 11 03:13:36 crc kubenswrapper[4824]: E1211 03:13:36.634771 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 03:13:45 crc kubenswrapper[4824]: I1211 03:13:45.570240 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-gp8dt_c70805eb-e095-4462-a202-e01860559471/control-plane-machine-set-operator/0.log" Dec 11 03:13:45 crc kubenswrapper[4824]: I1211 03:13:45.717576 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-d89dq_7704cded-5ba1-4474-8a31-c0ba947b7679/kube-rbac-proxy/0.log" Dec 11 03:13:45 crc kubenswrapper[4824]: I1211 03:13:45.726691 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-d89dq_7704cded-5ba1-4474-8a31-c0ba947b7679/machine-api-operator/0.log" Dec 11 03:13:50 crc kubenswrapper[4824]: I1211 03:13:50.633062 4824 scope.go:117] "RemoveContainer" containerID="fdad360b2681cc88a198f2fc151daab36d27e39aa44b3ec03ace588804769464" Dec 11 03:13:50 crc kubenswrapper[4824]: E1211 03:13:50.633939 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 03:14:00 crc kubenswrapper[4824]: I1211 03:14:00.984780 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-7jv9b_81bcc53d-3664-4987-8724-765fda6c8c09/cert-manager-controller/0.log" Dec 11 03:14:01 crc kubenswrapper[4824]: I1211 03:14:01.510689 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-bstmj_919ec639-d2fa-4447-86a4-a88b39d05ec3/cert-manager-cainjector/0.log" Dec 11 03:14:01 crc kubenswrapper[4824]: I1211 03:14:01.587398 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-hxtz9_b9bca511-59e3-4650-86c8-a82c0175d51a/cert-manager-webhook/0.log" Dec 11 03:14:01 crc kubenswrapper[4824]: I1211 03:14:01.632495 4824 scope.go:117] "RemoveContainer" containerID="fdad360b2681cc88a198f2fc151daab36d27e39aa44b3ec03ace588804769464" Dec 11 03:14:01 crc kubenswrapper[4824]: E1211 03:14:01.632740 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 03:14:12 crc kubenswrapper[4824]: I1211 03:14:12.632431 4824 scope.go:117] "RemoveContainer" containerID="fdad360b2681cc88a198f2fc151daab36d27e39aa44b3ec03ace588804769464" Dec 11 03:14:12 crc kubenswrapper[4824]: E1211 03:14:12.633095 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 03:14:15 crc kubenswrapper[4824]: I1211 03:14:15.386545 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-6hcp5_ffe586ab-061d-47f6-a3e8-220bf9ffebe2/nmstate-console-plugin/0.log" Dec 11 03:14:15 crc kubenswrapper[4824]: I1211 03:14:15.568462 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-b4h4m_22b4f565-3eff-4432-8171-56fa15cf7fcb/nmstate-handler/0.log" Dec 11 03:14:15 crc kubenswrapper[4824]: I1211 03:14:15.650766 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-qpjg9_731b5801-36ca-4f1d-8a34-4ad8554a0c3b/nmstate-metrics/0.log" Dec 11 03:14:15 crc kubenswrapper[4824]: I1211 03:14:15.664662 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-qpjg9_731b5801-36ca-4f1d-8a34-4ad8554a0c3b/kube-rbac-proxy/0.log" Dec 11 03:14:15 crc kubenswrapper[4824]: I1211 03:14:15.773412 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-fd644_4ad81e5d-ce6e-4e3c-ae63-7a87ce2a4def/nmstate-operator/0.log" Dec 11 03:14:15 crc kubenswrapper[4824]: I1211 03:14:15.838785 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-8v6lh_2b946a34-6de7-4afe-8876-18b15e09ea98/nmstate-webhook/0.log" Dec 11 03:14:24 crc kubenswrapper[4824]: I1211 03:14:24.633017 4824 scope.go:117] "RemoveContainer" containerID="fdad360b2681cc88a198f2fc151daab36d27e39aa44b3ec03ace588804769464" Dec 11 03:14:24 crc kubenswrapper[4824]: E1211 03:14:24.634083 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 03:14:33 crc kubenswrapper[4824]: I1211 03:14:33.275840 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-r49c8_7fbb3013-54ad-421f-8022-26b0ffad44f6/kube-rbac-proxy/0.log" Dec 11 03:14:33 crc kubenswrapper[4824]: I1211 03:14:33.316952 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-r49c8_7fbb3013-54ad-421f-8022-26b0ffad44f6/controller/0.log" Dec 11 03:14:33 crc kubenswrapper[4824]: I1211 03:14:33.464138 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-s7hz7_57d0325f-b062-4148-8b2a-c21f42de39a3/cp-frr-files/0.log" Dec 11 03:14:33 crc kubenswrapper[4824]: I1211 03:14:33.658584 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-s7hz7_57d0325f-b062-4148-8b2a-c21f42de39a3/cp-metrics/0.log" Dec 11 03:14:33 crc kubenswrapper[4824]: I1211 03:14:33.702893 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-s7hz7_57d0325f-b062-4148-8b2a-c21f42de39a3/cp-reloader/0.log" Dec 11 03:14:33 crc kubenswrapper[4824]: I1211 03:14:33.717613 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-s7hz7_57d0325f-b062-4148-8b2a-c21f42de39a3/cp-frr-files/0.log" Dec 11 03:14:33 crc kubenswrapper[4824]: I1211 03:14:33.741336 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-s7hz7_57d0325f-b062-4148-8b2a-c21f42de39a3/cp-reloader/0.log" Dec 11 03:14:33 crc kubenswrapper[4824]: I1211 03:14:33.874596 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-s7hz7_57d0325f-b062-4148-8b2a-c21f42de39a3/cp-reloader/0.log" Dec 11 03:14:33 crc kubenswrapper[4824]: I1211 03:14:33.897097 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-s7hz7_57d0325f-b062-4148-8b2a-c21f42de39a3/cp-frr-files/0.log" Dec 11 03:14:33 crc kubenswrapper[4824]: I1211 03:14:33.929413 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-s7hz7_57d0325f-b062-4148-8b2a-c21f42de39a3/cp-metrics/0.log" Dec 11 03:14:33 crc kubenswrapper[4824]: I1211 03:14:33.959515 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-s7hz7_57d0325f-b062-4148-8b2a-c21f42de39a3/cp-metrics/0.log" Dec 11 03:14:34 crc kubenswrapper[4824]: I1211 03:14:34.089162 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-s7hz7_57d0325f-b062-4148-8b2a-c21f42de39a3/cp-reloader/0.log" Dec 11 03:14:34 crc kubenswrapper[4824]: I1211 03:14:34.114375 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-s7hz7_57d0325f-b062-4148-8b2a-c21f42de39a3/cp-frr-files/0.log" Dec 11 03:14:34 crc kubenswrapper[4824]: I1211 03:14:34.142085 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-s7hz7_57d0325f-b062-4148-8b2a-c21f42de39a3/cp-metrics/0.log" Dec 11 03:14:34 crc kubenswrapper[4824]: I1211 03:14:34.170318 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-s7hz7_57d0325f-b062-4148-8b2a-c21f42de39a3/controller/0.log" Dec 11 03:14:34 crc kubenswrapper[4824]: I1211 03:14:34.300351 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-s7hz7_57d0325f-b062-4148-8b2a-c21f42de39a3/frr-metrics/0.log" Dec 11 03:14:34 crc kubenswrapper[4824]: I1211 03:14:34.373242 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-s7hz7_57d0325f-b062-4148-8b2a-c21f42de39a3/kube-rbac-proxy/0.log" Dec 11 03:14:34 crc kubenswrapper[4824]: I1211 03:14:34.459520 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-s7hz7_57d0325f-b062-4148-8b2a-c21f42de39a3/kube-rbac-proxy-frr/0.log" Dec 11 03:14:34 crc kubenswrapper[4824]: I1211 03:14:34.539278 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-s7hz7_57d0325f-b062-4148-8b2a-c21f42de39a3/reloader/0.log" Dec 11 03:14:34 crc kubenswrapper[4824]: I1211 03:14:34.647665 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-mnh62_053662c4-f168-42b3-a9c9-dc21ca9908ea/frr-k8s-webhook-server/0.log" Dec 11 03:14:35 crc kubenswrapper[4824]: I1211 03:14:35.470626 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-s7hz7_57d0325f-b062-4148-8b2a-c21f42de39a3/frr/0.log" Dec 11 03:14:35 crc kubenswrapper[4824]: I1211 03:14:35.580509 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-7d6477f866-c4dqb_47581f0c-dffe-42e8-8211-bb40625aadf1/manager/0.log" Dec 11 03:14:35 crc kubenswrapper[4824]: I1211 03:14:35.667821 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-6cb87f7d4c-jt2ns_0c1d110c-0bd5-410f-af71-7af964dfb17d/webhook-server/0.log" Dec 11 03:14:35 crc kubenswrapper[4824]: I1211 03:14:35.786844 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-d99pv_6e59f040-b574-48e8-9a0d-05072fc1a3f6/kube-rbac-proxy/0.log" Dec 11 03:14:36 crc kubenswrapper[4824]: I1211 03:14:36.060203 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-d99pv_6e59f040-b574-48e8-9a0d-05072fc1a3f6/speaker/0.log" Dec 11 03:14:37 crc kubenswrapper[4824]: I1211 03:14:37.633159 4824 scope.go:117] "RemoveContainer" containerID="fdad360b2681cc88a198f2fc151daab36d27e39aa44b3ec03ace588804769464" Dec 11 03:14:37 crc kubenswrapper[4824]: E1211 03:14:37.634275 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 03:14:50 crc kubenswrapper[4824]: I1211 03:14:50.573663 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f97jkp_bb38965f-b3bb-4d2e-8430-8c29298730e4/util/0.log" Dec 11 03:14:50 crc kubenswrapper[4824]: I1211 03:14:50.632287 4824 scope.go:117] "RemoveContainer" containerID="fdad360b2681cc88a198f2fc151daab36d27e39aa44b3ec03ace588804769464" Dec 11 03:14:50 crc kubenswrapper[4824]: E1211 03:14:50.632546 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 03:14:50 crc kubenswrapper[4824]: I1211 03:14:50.744698 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f97jkp_bb38965f-b3bb-4d2e-8430-8c29298730e4/util/0.log" Dec 11 03:14:50 crc kubenswrapper[4824]: I1211 03:14:50.769805 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f97jkp_bb38965f-b3bb-4d2e-8430-8c29298730e4/pull/0.log" Dec 11 03:14:50 crc kubenswrapper[4824]: I1211 03:14:50.791735 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f97jkp_bb38965f-b3bb-4d2e-8430-8c29298730e4/pull/0.log" Dec 11 03:14:50 crc kubenswrapper[4824]: I1211 03:14:50.926339 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f97jkp_bb38965f-b3bb-4d2e-8430-8c29298730e4/util/0.log" Dec 11 03:14:50 crc kubenswrapper[4824]: I1211 03:14:50.940967 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f97jkp_bb38965f-b3bb-4d2e-8430-8c29298730e4/pull/0.log" Dec 11 03:14:50 crc kubenswrapper[4824]: I1211 03:14:50.962621 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f97jkp_bb38965f-b3bb-4d2e-8430-8c29298730e4/extract/0.log" Dec 11 03:14:51 crc kubenswrapper[4824]: I1211 03:14:51.067904 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83x97sf_55c9ce15-2335-433b-a7ac-cfdf1a3ef599/util/0.log" Dec 11 03:14:51 crc kubenswrapper[4824]: I1211 03:14:51.234249 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83x97sf_55c9ce15-2335-433b-a7ac-cfdf1a3ef599/util/0.log" Dec 11 03:14:51 crc kubenswrapper[4824]: I1211 03:14:51.237013 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83x97sf_55c9ce15-2335-433b-a7ac-cfdf1a3ef599/pull/0.log" Dec 11 03:14:51 crc kubenswrapper[4824]: I1211 03:14:51.261568 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83x97sf_55c9ce15-2335-433b-a7ac-cfdf1a3ef599/pull/0.log" Dec 11 03:14:51 crc kubenswrapper[4824]: I1211 03:14:51.429613 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83x97sf_55c9ce15-2335-433b-a7ac-cfdf1a3ef599/util/0.log" Dec 11 03:14:51 crc kubenswrapper[4824]: I1211 03:14:51.430810 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83x97sf_55c9ce15-2335-433b-a7ac-cfdf1a3ef599/pull/0.log" Dec 11 03:14:51 crc kubenswrapper[4824]: I1211 03:14:51.447354 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83x97sf_55c9ce15-2335-433b-a7ac-cfdf1a3ef599/extract/0.log" Dec 11 03:14:51 crc kubenswrapper[4824]: I1211 03:14:51.593278 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-nqwwb_a61fd179-f3bd-4a79-ae61-d338e7994602/extract-utilities/0.log" Dec 11 03:14:51 crc kubenswrapper[4824]: I1211 03:14:51.753598 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-nqwwb_a61fd179-f3bd-4a79-ae61-d338e7994602/extract-utilities/0.log" Dec 11 03:14:51 crc kubenswrapper[4824]: I1211 03:14:51.757834 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-nqwwb_a61fd179-f3bd-4a79-ae61-d338e7994602/extract-content/0.log" Dec 11 03:14:51 crc kubenswrapper[4824]: I1211 03:14:51.798884 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-nqwwb_a61fd179-f3bd-4a79-ae61-d338e7994602/extract-content/0.log" Dec 11 03:14:51 crc kubenswrapper[4824]: I1211 03:14:51.917608 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-nqwwb_a61fd179-f3bd-4a79-ae61-d338e7994602/extract-utilities/0.log" Dec 11 03:14:51 crc kubenswrapper[4824]: I1211 03:14:51.925661 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-nqwwb_a61fd179-f3bd-4a79-ae61-d338e7994602/extract-content/0.log" Dec 11 03:14:52 crc kubenswrapper[4824]: I1211 03:14:52.129654 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4qn4g_d5e0df10-f888-47f3-87a9-889529a89473/extract-utilities/0.log" Dec 11 03:14:52 crc kubenswrapper[4824]: I1211 03:14:52.371581 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4qn4g_d5e0df10-f888-47f3-87a9-889529a89473/extract-content/0.log" Dec 11 03:14:52 crc kubenswrapper[4824]: I1211 03:14:52.396328 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-nqwwb_a61fd179-f3bd-4a79-ae61-d338e7994602/registry-server/0.log" Dec 11 03:14:52 crc kubenswrapper[4824]: I1211 03:14:52.402748 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4qn4g_d5e0df10-f888-47f3-87a9-889529a89473/extract-content/0.log" Dec 11 03:14:52 crc kubenswrapper[4824]: I1211 03:14:52.428877 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4qn4g_d5e0df10-f888-47f3-87a9-889529a89473/extract-utilities/0.log" Dec 11 03:14:52 crc kubenswrapper[4824]: I1211 03:14:52.538061 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4qn4g_d5e0df10-f888-47f3-87a9-889529a89473/extract-utilities/0.log" Dec 11 03:14:52 crc kubenswrapper[4824]: I1211 03:14:52.579560 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4qn4g_d5e0df10-f888-47f3-87a9-889529a89473/extract-content/0.log" Dec 11 03:14:52 crc kubenswrapper[4824]: I1211 03:14:52.749869 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-xj7qj_19528072-20f2-4741-bfd9-b7c44430b834/marketplace-operator/0.log" Dec 11 03:14:52 crc kubenswrapper[4824]: I1211 03:14:52.848820 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-k4msm_38e53ad7-42f0-4b59-a721-b5e38b6e1904/extract-utilities/0.log" Dec 11 03:14:53 crc kubenswrapper[4824]: I1211 03:14:53.001930 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-k4msm_38e53ad7-42f0-4b59-a721-b5e38b6e1904/extract-utilities/0.log" Dec 11 03:14:53 crc kubenswrapper[4824]: I1211 03:14:53.038925 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-k4msm_38e53ad7-42f0-4b59-a721-b5e38b6e1904/extract-content/0.log" Dec 11 03:14:53 crc kubenswrapper[4824]: I1211 03:14:53.085438 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-k4msm_38e53ad7-42f0-4b59-a721-b5e38b6e1904/extract-content/0.log" Dec 11 03:14:53 crc kubenswrapper[4824]: I1211 03:14:53.106720 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4qn4g_d5e0df10-f888-47f3-87a9-889529a89473/registry-server/0.log" Dec 11 03:14:53 crc kubenswrapper[4824]: I1211 03:14:53.280200 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-k4msm_38e53ad7-42f0-4b59-a721-b5e38b6e1904/extract-utilities/0.log" Dec 11 03:14:53 crc kubenswrapper[4824]: I1211 03:14:53.281579 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-k4msm_38e53ad7-42f0-4b59-a721-b5e38b6e1904/extract-content/0.log" Dec 11 03:14:53 crc kubenswrapper[4824]: I1211 03:14:53.445066 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-k4msm_38e53ad7-42f0-4b59-a721-b5e38b6e1904/registry-server/0.log" Dec 11 03:14:53 crc kubenswrapper[4824]: I1211 03:14:53.482395 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-wjmvj_351ca498-5ccd-4d80-a99f-3d56f7c5dbf2/extract-utilities/0.log" Dec 11 03:14:53 crc kubenswrapper[4824]: I1211 03:14:53.669212 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-wjmvj_351ca498-5ccd-4d80-a99f-3d56f7c5dbf2/extract-content/0.log" Dec 11 03:14:53 crc kubenswrapper[4824]: I1211 03:14:53.694772 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-wjmvj_351ca498-5ccd-4d80-a99f-3d56f7c5dbf2/extract-content/0.log" Dec 11 03:14:53 crc kubenswrapper[4824]: I1211 03:14:53.699796 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-wjmvj_351ca498-5ccd-4d80-a99f-3d56f7c5dbf2/extract-utilities/0.log" Dec 11 03:14:53 crc kubenswrapper[4824]: I1211 03:14:53.858303 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-wjmvj_351ca498-5ccd-4d80-a99f-3d56f7c5dbf2/extract-utilities/0.log" Dec 11 03:14:53 crc kubenswrapper[4824]: I1211 03:14:53.880265 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-wjmvj_351ca498-5ccd-4d80-a99f-3d56f7c5dbf2/extract-content/0.log" Dec 11 03:14:53 crc kubenswrapper[4824]: I1211 03:14:53.978705 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-wjmvj_351ca498-5ccd-4d80-a99f-3d56f7c5dbf2/registry-server/0.log" Dec 11 03:15:00 crc kubenswrapper[4824]: I1211 03:15:00.213327 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423715-gfhlc"] Dec 11 03:15:00 crc kubenswrapper[4824]: E1211 03:15:00.214258 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad4e10f1-aa90-4ae8-a1f6-8de2202ebbc6" containerName="registry-server" Dec 11 03:15:00 crc kubenswrapper[4824]: I1211 03:15:00.214274 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad4e10f1-aa90-4ae8-a1f6-8de2202ebbc6" containerName="registry-server" Dec 11 03:15:00 crc kubenswrapper[4824]: E1211 03:15:00.214285 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad4e10f1-aa90-4ae8-a1f6-8de2202ebbc6" containerName="extract-utilities" Dec 11 03:15:00 crc kubenswrapper[4824]: I1211 03:15:00.214293 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad4e10f1-aa90-4ae8-a1f6-8de2202ebbc6" containerName="extract-utilities" Dec 11 03:15:00 crc kubenswrapper[4824]: E1211 03:15:00.214315 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad4e10f1-aa90-4ae8-a1f6-8de2202ebbc6" containerName="extract-content" Dec 11 03:15:00 crc kubenswrapper[4824]: I1211 03:15:00.214324 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad4e10f1-aa90-4ae8-a1f6-8de2202ebbc6" containerName="extract-content" Dec 11 03:15:00 crc kubenswrapper[4824]: I1211 03:15:00.214580 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad4e10f1-aa90-4ae8-a1f6-8de2202ebbc6" containerName="registry-server" Dec 11 03:15:00 crc kubenswrapper[4824]: I1211 03:15:00.215317 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423715-gfhlc" Dec 11 03:15:00 crc kubenswrapper[4824]: I1211 03:15:00.218457 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 11 03:15:00 crc kubenswrapper[4824]: I1211 03:15:00.220376 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 11 03:15:00 crc kubenswrapper[4824]: I1211 03:15:00.228280 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423715-gfhlc"] Dec 11 03:15:00 crc kubenswrapper[4824]: I1211 03:15:00.394248 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f41fe791-cd27-4076-99b9-22766ee116ef-config-volume\") pod \"collect-profiles-29423715-gfhlc\" (UID: \"f41fe791-cd27-4076-99b9-22766ee116ef\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423715-gfhlc" Dec 11 03:15:00 crc kubenswrapper[4824]: I1211 03:15:00.394412 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f41fe791-cd27-4076-99b9-22766ee116ef-secret-volume\") pod \"collect-profiles-29423715-gfhlc\" (UID: \"f41fe791-cd27-4076-99b9-22766ee116ef\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423715-gfhlc" Dec 11 03:15:00 crc kubenswrapper[4824]: I1211 03:15:00.394453 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dgx8g\" (UniqueName: \"kubernetes.io/projected/f41fe791-cd27-4076-99b9-22766ee116ef-kube-api-access-dgx8g\") pod \"collect-profiles-29423715-gfhlc\" (UID: \"f41fe791-cd27-4076-99b9-22766ee116ef\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423715-gfhlc" Dec 11 03:15:00 crc kubenswrapper[4824]: I1211 03:15:00.495882 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f41fe791-cd27-4076-99b9-22766ee116ef-config-volume\") pod \"collect-profiles-29423715-gfhlc\" (UID: \"f41fe791-cd27-4076-99b9-22766ee116ef\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423715-gfhlc" Dec 11 03:15:00 crc kubenswrapper[4824]: I1211 03:15:00.496018 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f41fe791-cd27-4076-99b9-22766ee116ef-secret-volume\") pod \"collect-profiles-29423715-gfhlc\" (UID: \"f41fe791-cd27-4076-99b9-22766ee116ef\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423715-gfhlc" Dec 11 03:15:00 crc kubenswrapper[4824]: I1211 03:15:00.496044 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dgx8g\" (UniqueName: \"kubernetes.io/projected/f41fe791-cd27-4076-99b9-22766ee116ef-kube-api-access-dgx8g\") pod \"collect-profiles-29423715-gfhlc\" (UID: \"f41fe791-cd27-4076-99b9-22766ee116ef\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423715-gfhlc" Dec 11 03:15:00 crc kubenswrapper[4824]: I1211 03:15:00.497294 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f41fe791-cd27-4076-99b9-22766ee116ef-config-volume\") pod \"collect-profiles-29423715-gfhlc\" (UID: \"f41fe791-cd27-4076-99b9-22766ee116ef\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423715-gfhlc" Dec 11 03:15:01 crc kubenswrapper[4824]: I1211 03:15:01.159588 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f41fe791-cd27-4076-99b9-22766ee116ef-secret-volume\") pod \"collect-profiles-29423715-gfhlc\" (UID: \"f41fe791-cd27-4076-99b9-22766ee116ef\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423715-gfhlc" Dec 11 03:15:01 crc kubenswrapper[4824]: I1211 03:15:01.161793 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dgx8g\" (UniqueName: \"kubernetes.io/projected/f41fe791-cd27-4076-99b9-22766ee116ef-kube-api-access-dgx8g\") pod \"collect-profiles-29423715-gfhlc\" (UID: \"f41fe791-cd27-4076-99b9-22766ee116ef\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423715-gfhlc" Dec 11 03:15:01 crc kubenswrapper[4824]: I1211 03:15:01.436317 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423715-gfhlc" Dec 11 03:15:01 crc kubenswrapper[4824]: I1211 03:15:01.908554 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423715-gfhlc"] Dec 11 03:15:02 crc kubenswrapper[4824]: I1211 03:15:02.632380 4824 scope.go:117] "RemoveContainer" containerID="fdad360b2681cc88a198f2fc151daab36d27e39aa44b3ec03ace588804769464" Dec 11 03:15:02 crc kubenswrapper[4824]: E1211 03:15:02.633053 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 03:15:02 crc kubenswrapper[4824]: I1211 03:15:02.720822 4824 generic.go:334] "Generic (PLEG): container finished" podID="f41fe791-cd27-4076-99b9-22766ee116ef" containerID="1a108434ce42455a84ca63527637dde1b4d52e4f46bab14aa6da1186b596cc53" exitCode=0 Dec 11 03:15:02 crc kubenswrapper[4824]: I1211 03:15:02.720896 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423715-gfhlc" event={"ID":"f41fe791-cd27-4076-99b9-22766ee116ef","Type":"ContainerDied","Data":"1a108434ce42455a84ca63527637dde1b4d52e4f46bab14aa6da1186b596cc53"} Dec 11 03:15:02 crc kubenswrapper[4824]: I1211 03:15:02.720924 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423715-gfhlc" event={"ID":"f41fe791-cd27-4076-99b9-22766ee116ef","Type":"ContainerStarted","Data":"96d580e59c06cd053d07e990b49ee1dfe2e82898ebe9d4b020223b668311db5f"} Dec 11 03:15:04 crc kubenswrapper[4824]: I1211 03:15:04.098655 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423715-gfhlc" Dec 11 03:15:04 crc kubenswrapper[4824]: I1211 03:15:04.280550 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dgx8g\" (UniqueName: \"kubernetes.io/projected/f41fe791-cd27-4076-99b9-22766ee116ef-kube-api-access-dgx8g\") pod \"f41fe791-cd27-4076-99b9-22766ee116ef\" (UID: \"f41fe791-cd27-4076-99b9-22766ee116ef\") " Dec 11 03:15:04 crc kubenswrapper[4824]: I1211 03:15:04.281275 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f41fe791-cd27-4076-99b9-22766ee116ef-config-volume\") pod \"f41fe791-cd27-4076-99b9-22766ee116ef\" (UID: \"f41fe791-cd27-4076-99b9-22766ee116ef\") " Dec 11 03:15:04 crc kubenswrapper[4824]: I1211 03:15:04.281640 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f41fe791-cd27-4076-99b9-22766ee116ef-secret-volume\") pod \"f41fe791-cd27-4076-99b9-22766ee116ef\" (UID: \"f41fe791-cd27-4076-99b9-22766ee116ef\") " Dec 11 03:15:04 crc kubenswrapper[4824]: I1211 03:15:04.281881 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f41fe791-cd27-4076-99b9-22766ee116ef-config-volume" (OuterVolumeSpecName: "config-volume") pod "f41fe791-cd27-4076-99b9-22766ee116ef" (UID: "f41fe791-cd27-4076-99b9-22766ee116ef"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 03:15:04 crc kubenswrapper[4824]: I1211 03:15:04.284295 4824 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f41fe791-cd27-4076-99b9-22766ee116ef-config-volume\") on node \"crc\" DevicePath \"\"" Dec 11 03:15:04 crc kubenswrapper[4824]: I1211 03:15:04.298336 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f41fe791-cd27-4076-99b9-22766ee116ef-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "f41fe791-cd27-4076-99b9-22766ee116ef" (UID: "f41fe791-cd27-4076-99b9-22766ee116ef"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 03:15:04 crc kubenswrapper[4824]: I1211 03:15:04.298423 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f41fe791-cd27-4076-99b9-22766ee116ef-kube-api-access-dgx8g" (OuterVolumeSpecName: "kube-api-access-dgx8g") pod "f41fe791-cd27-4076-99b9-22766ee116ef" (UID: "f41fe791-cd27-4076-99b9-22766ee116ef"). InnerVolumeSpecName "kube-api-access-dgx8g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 03:15:04 crc kubenswrapper[4824]: I1211 03:15:04.385682 4824 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f41fe791-cd27-4076-99b9-22766ee116ef-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 11 03:15:04 crc kubenswrapper[4824]: I1211 03:15:04.386001 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dgx8g\" (UniqueName: \"kubernetes.io/projected/f41fe791-cd27-4076-99b9-22766ee116ef-kube-api-access-dgx8g\") on node \"crc\" DevicePath \"\"" Dec 11 03:15:04 crc kubenswrapper[4824]: I1211 03:15:04.738720 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423715-gfhlc" event={"ID":"f41fe791-cd27-4076-99b9-22766ee116ef","Type":"ContainerDied","Data":"96d580e59c06cd053d07e990b49ee1dfe2e82898ebe9d4b020223b668311db5f"} Dec 11 03:15:04 crc kubenswrapper[4824]: I1211 03:15:04.738765 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="96d580e59c06cd053d07e990b49ee1dfe2e82898ebe9d4b020223b668311db5f" Dec 11 03:15:04 crc kubenswrapper[4824]: I1211 03:15:04.738786 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423715-gfhlc" Dec 11 03:15:05 crc kubenswrapper[4824]: I1211 03:15:05.174244 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423670-szq5v"] Dec 11 03:15:05 crc kubenswrapper[4824]: I1211 03:15:05.187519 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423670-szq5v"] Dec 11 03:15:06 crc kubenswrapper[4824]: I1211 03:15:06.653964 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0104cb73-80d6-4553-9bab-1ead3adeaca2" path="/var/lib/kubelet/pods/0104cb73-80d6-4553-9bab-1ead3adeaca2/volumes" Dec 11 03:15:11 crc kubenswrapper[4824]: I1211 03:15:11.739412 4824 scope.go:117] "RemoveContainer" containerID="3e60c204f2b971f9917d13f901f25bd11bc76d14ad92661319c0c2c3553c70cf" Dec 11 03:15:17 crc kubenswrapper[4824]: I1211 03:15:17.632187 4824 scope.go:117] "RemoveContainer" containerID="fdad360b2681cc88a198f2fc151daab36d27e39aa44b3ec03ace588804769464" Dec 11 03:15:17 crc kubenswrapper[4824]: E1211 03:15:17.633040 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 03:15:31 crc kubenswrapper[4824]: I1211 03:15:31.633535 4824 scope.go:117] "RemoveContainer" containerID="fdad360b2681cc88a198f2fc151daab36d27e39aa44b3ec03ace588804769464" Dec 11 03:15:31 crc kubenswrapper[4824]: E1211 03:15:31.634430 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 03:15:42 crc kubenswrapper[4824]: I1211 03:15:42.633928 4824 scope.go:117] "RemoveContainer" containerID="fdad360b2681cc88a198f2fc151daab36d27e39aa44b3ec03ace588804769464" Dec 11 03:15:42 crc kubenswrapper[4824]: E1211 03:15:42.634941 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 03:15:55 crc kubenswrapper[4824]: I1211 03:15:55.632986 4824 scope.go:117] "RemoveContainer" containerID="fdad360b2681cc88a198f2fc151daab36d27e39aa44b3ec03ace588804769464" Dec 11 03:15:55 crc kubenswrapper[4824]: E1211 03:15:55.634374 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 03:16:09 crc kubenswrapper[4824]: I1211 03:16:09.632412 4824 scope.go:117] "RemoveContainer" containerID="fdad360b2681cc88a198f2fc151daab36d27e39aa44b3ec03ace588804769464" Dec 11 03:16:09 crc kubenswrapper[4824]: E1211 03:16:09.633233 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 03:16:23 crc kubenswrapper[4824]: I1211 03:16:23.633232 4824 scope.go:117] "RemoveContainer" containerID="fdad360b2681cc88a198f2fc151daab36d27e39aa44b3ec03ace588804769464" Dec 11 03:16:23 crc kubenswrapper[4824]: E1211 03:16:23.634145 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 03:16:34 crc kubenswrapper[4824]: I1211 03:16:34.633994 4824 scope.go:117] "RemoveContainer" containerID="fdad360b2681cc88a198f2fc151daab36d27e39aa44b3ec03ace588804769464" Dec 11 03:16:34 crc kubenswrapper[4824]: E1211 03:16:34.635494 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 03:16:34 crc kubenswrapper[4824]: I1211 03:16:34.690935 4824 generic.go:334] "Generic (PLEG): container finished" podID="9a78e53b-6cf8-44aa-abbd-3f5ab6fe156b" containerID="565bc55d51103690977681b6054c6bfe409d8f1c588376ce223d412aafbd75bd" exitCode=0 Dec 11 03:16:34 crc kubenswrapper[4824]: I1211 03:16:34.690981 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2pxpn/must-gather-8jk9m" event={"ID":"9a78e53b-6cf8-44aa-abbd-3f5ab6fe156b","Type":"ContainerDied","Data":"565bc55d51103690977681b6054c6bfe409d8f1c588376ce223d412aafbd75bd"} Dec 11 03:16:34 crc kubenswrapper[4824]: I1211 03:16:34.691513 4824 scope.go:117] "RemoveContainer" containerID="565bc55d51103690977681b6054c6bfe409d8f1c588376ce223d412aafbd75bd" Dec 11 03:16:34 crc kubenswrapper[4824]: I1211 03:16:34.998897 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-2pxpn_must-gather-8jk9m_9a78e53b-6cf8-44aa-abbd-3f5ab6fe156b/gather/0.log" Dec 11 03:16:44 crc kubenswrapper[4824]: I1211 03:16:44.861203 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-2pxpn/must-gather-8jk9m"] Dec 11 03:16:44 crc kubenswrapper[4824]: I1211 03:16:44.862190 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-2pxpn/must-gather-8jk9m" podUID="9a78e53b-6cf8-44aa-abbd-3f5ab6fe156b" containerName="copy" containerID="cri-o://840585396fdd658dee8442e9090dd91c78f38255e7946543986306315a99228d" gracePeriod=2 Dec 11 03:16:44 crc kubenswrapper[4824]: I1211 03:16:44.872045 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-2pxpn/must-gather-8jk9m"] Dec 11 03:16:45 crc kubenswrapper[4824]: I1211 03:16:45.289221 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-2pxpn_must-gather-8jk9m_9a78e53b-6cf8-44aa-abbd-3f5ab6fe156b/copy/0.log" Dec 11 03:16:45 crc kubenswrapper[4824]: I1211 03:16:45.290034 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2pxpn/must-gather-8jk9m" Dec 11 03:16:45 crc kubenswrapper[4824]: I1211 03:16:45.417481 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8jjzm\" (UniqueName: \"kubernetes.io/projected/9a78e53b-6cf8-44aa-abbd-3f5ab6fe156b-kube-api-access-8jjzm\") pod \"9a78e53b-6cf8-44aa-abbd-3f5ab6fe156b\" (UID: \"9a78e53b-6cf8-44aa-abbd-3f5ab6fe156b\") " Dec 11 03:16:45 crc kubenswrapper[4824]: I1211 03:16:45.417610 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/9a78e53b-6cf8-44aa-abbd-3f5ab6fe156b-must-gather-output\") pod \"9a78e53b-6cf8-44aa-abbd-3f5ab6fe156b\" (UID: \"9a78e53b-6cf8-44aa-abbd-3f5ab6fe156b\") " Dec 11 03:16:45 crc kubenswrapper[4824]: I1211 03:16:45.436085 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a78e53b-6cf8-44aa-abbd-3f5ab6fe156b-kube-api-access-8jjzm" (OuterVolumeSpecName: "kube-api-access-8jjzm") pod "9a78e53b-6cf8-44aa-abbd-3f5ab6fe156b" (UID: "9a78e53b-6cf8-44aa-abbd-3f5ab6fe156b"). InnerVolumeSpecName "kube-api-access-8jjzm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 03:16:45 crc kubenswrapper[4824]: I1211 03:16:45.520456 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8jjzm\" (UniqueName: \"kubernetes.io/projected/9a78e53b-6cf8-44aa-abbd-3f5ab6fe156b-kube-api-access-8jjzm\") on node \"crc\" DevicePath \"\"" Dec 11 03:16:45 crc kubenswrapper[4824]: I1211 03:16:45.588823 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9a78e53b-6cf8-44aa-abbd-3f5ab6fe156b-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "9a78e53b-6cf8-44aa-abbd-3f5ab6fe156b" (UID: "9a78e53b-6cf8-44aa-abbd-3f5ab6fe156b"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 03:16:45 crc kubenswrapper[4824]: I1211 03:16:45.622675 4824 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/9a78e53b-6cf8-44aa-abbd-3f5ab6fe156b-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 11 03:16:45 crc kubenswrapper[4824]: I1211 03:16:45.814024 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-2pxpn_must-gather-8jk9m_9a78e53b-6cf8-44aa-abbd-3f5ab6fe156b/copy/0.log" Dec 11 03:16:45 crc kubenswrapper[4824]: I1211 03:16:45.814580 4824 generic.go:334] "Generic (PLEG): container finished" podID="9a78e53b-6cf8-44aa-abbd-3f5ab6fe156b" containerID="840585396fdd658dee8442e9090dd91c78f38255e7946543986306315a99228d" exitCode=143 Dec 11 03:16:45 crc kubenswrapper[4824]: I1211 03:16:45.814688 4824 scope.go:117] "RemoveContainer" containerID="840585396fdd658dee8442e9090dd91c78f38255e7946543986306315a99228d" Dec 11 03:16:45 crc kubenswrapper[4824]: I1211 03:16:45.814912 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2pxpn/must-gather-8jk9m" Dec 11 03:16:45 crc kubenswrapper[4824]: I1211 03:16:45.849355 4824 scope.go:117] "RemoveContainer" containerID="565bc55d51103690977681b6054c6bfe409d8f1c588376ce223d412aafbd75bd" Dec 11 03:16:45 crc kubenswrapper[4824]: I1211 03:16:45.924696 4824 scope.go:117] "RemoveContainer" containerID="840585396fdd658dee8442e9090dd91c78f38255e7946543986306315a99228d" Dec 11 03:16:45 crc kubenswrapper[4824]: E1211 03:16:45.926951 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"840585396fdd658dee8442e9090dd91c78f38255e7946543986306315a99228d\": container with ID starting with 840585396fdd658dee8442e9090dd91c78f38255e7946543986306315a99228d not found: ID does not exist" containerID="840585396fdd658dee8442e9090dd91c78f38255e7946543986306315a99228d" Dec 11 03:16:45 crc kubenswrapper[4824]: I1211 03:16:45.926997 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"840585396fdd658dee8442e9090dd91c78f38255e7946543986306315a99228d"} err="failed to get container status \"840585396fdd658dee8442e9090dd91c78f38255e7946543986306315a99228d\": rpc error: code = NotFound desc = could not find container \"840585396fdd658dee8442e9090dd91c78f38255e7946543986306315a99228d\": container with ID starting with 840585396fdd658dee8442e9090dd91c78f38255e7946543986306315a99228d not found: ID does not exist" Dec 11 03:16:45 crc kubenswrapper[4824]: I1211 03:16:45.927022 4824 scope.go:117] "RemoveContainer" containerID="565bc55d51103690977681b6054c6bfe409d8f1c588376ce223d412aafbd75bd" Dec 11 03:16:45 crc kubenswrapper[4824]: E1211 03:16:45.927340 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"565bc55d51103690977681b6054c6bfe409d8f1c588376ce223d412aafbd75bd\": container with ID starting with 565bc55d51103690977681b6054c6bfe409d8f1c588376ce223d412aafbd75bd not found: ID does not exist" containerID="565bc55d51103690977681b6054c6bfe409d8f1c588376ce223d412aafbd75bd" Dec 11 03:16:45 crc kubenswrapper[4824]: I1211 03:16:45.927383 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"565bc55d51103690977681b6054c6bfe409d8f1c588376ce223d412aafbd75bd"} err="failed to get container status \"565bc55d51103690977681b6054c6bfe409d8f1c588376ce223d412aafbd75bd\": rpc error: code = NotFound desc = could not find container \"565bc55d51103690977681b6054c6bfe409d8f1c588376ce223d412aafbd75bd\": container with ID starting with 565bc55d51103690977681b6054c6bfe409d8f1c588376ce223d412aafbd75bd not found: ID does not exist" Dec 11 03:16:46 crc kubenswrapper[4824]: I1211 03:16:46.644265 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a78e53b-6cf8-44aa-abbd-3f5ab6fe156b" path="/var/lib/kubelet/pods/9a78e53b-6cf8-44aa-abbd-3f5ab6fe156b/volumes" Dec 11 03:16:47 crc kubenswrapper[4824]: I1211 03:16:47.632473 4824 scope.go:117] "RemoveContainer" containerID="fdad360b2681cc88a198f2fc151daab36d27e39aa44b3ec03ace588804769464" Dec 11 03:16:47 crc kubenswrapper[4824]: E1211 03:16:47.633035 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 03:16:58 crc kubenswrapper[4824]: I1211 03:16:58.641646 4824 scope.go:117] "RemoveContainer" containerID="fdad360b2681cc88a198f2fc151daab36d27e39aa44b3ec03ace588804769464" Dec 11 03:16:58 crc kubenswrapper[4824]: E1211 03:16:58.642461 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gx6xt_openshift-machine-config-operator(44b156e4-64a4-4d45-aa5e-9b10a862faed)\"" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" Dec 11 03:17:09 crc kubenswrapper[4824]: I1211 03:17:09.632976 4824 scope.go:117] "RemoveContainer" containerID="fdad360b2681cc88a198f2fc151daab36d27e39aa44b3ec03ace588804769464" Dec 11 03:17:10 crc kubenswrapper[4824]: I1211 03:17:10.121938 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" event={"ID":"44b156e4-64a4-4d45-aa5e-9b10a862faed","Type":"ContainerStarted","Data":"b575259a1a7fa7f8fce57992718f48a43eac4df98dadf6265e0879b8d98aec24"} Dec 11 03:17:10 crc kubenswrapper[4824]: I1211 03:17:10.386716 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-k9zwd"] Dec 11 03:17:10 crc kubenswrapper[4824]: E1211 03:17:10.387225 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a78e53b-6cf8-44aa-abbd-3f5ab6fe156b" containerName="copy" Dec 11 03:17:10 crc kubenswrapper[4824]: I1211 03:17:10.387247 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a78e53b-6cf8-44aa-abbd-3f5ab6fe156b" containerName="copy" Dec 11 03:17:10 crc kubenswrapper[4824]: E1211 03:17:10.387267 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a78e53b-6cf8-44aa-abbd-3f5ab6fe156b" containerName="gather" Dec 11 03:17:10 crc kubenswrapper[4824]: I1211 03:17:10.387276 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a78e53b-6cf8-44aa-abbd-3f5ab6fe156b" containerName="gather" Dec 11 03:17:10 crc kubenswrapper[4824]: E1211 03:17:10.387299 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f41fe791-cd27-4076-99b9-22766ee116ef" containerName="collect-profiles" Dec 11 03:17:10 crc kubenswrapper[4824]: I1211 03:17:10.387309 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f41fe791-cd27-4076-99b9-22766ee116ef" containerName="collect-profiles" Dec 11 03:17:10 crc kubenswrapper[4824]: I1211 03:17:10.387523 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a78e53b-6cf8-44aa-abbd-3f5ab6fe156b" containerName="gather" Dec 11 03:17:10 crc kubenswrapper[4824]: I1211 03:17:10.387538 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a78e53b-6cf8-44aa-abbd-3f5ab6fe156b" containerName="copy" Dec 11 03:17:10 crc kubenswrapper[4824]: I1211 03:17:10.387550 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f41fe791-cd27-4076-99b9-22766ee116ef" containerName="collect-profiles" Dec 11 03:17:10 crc kubenswrapper[4824]: I1211 03:17:10.391398 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k9zwd" Dec 11 03:17:10 crc kubenswrapper[4824]: I1211 03:17:10.412154 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-k9zwd"] Dec 11 03:17:10 crc kubenswrapper[4824]: I1211 03:17:10.512299 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/102d6abf-7e9f-4a69-8027-5b0dd00d257a-utilities\") pod \"redhat-marketplace-k9zwd\" (UID: \"102d6abf-7e9f-4a69-8027-5b0dd00d257a\") " pod="openshift-marketplace/redhat-marketplace-k9zwd" Dec 11 03:17:10 crc kubenswrapper[4824]: I1211 03:17:10.512444 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/102d6abf-7e9f-4a69-8027-5b0dd00d257a-catalog-content\") pod \"redhat-marketplace-k9zwd\" (UID: \"102d6abf-7e9f-4a69-8027-5b0dd00d257a\") " pod="openshift-marketplace/redhat-marketplace-k9zwd" Dec 11 03:17:10 crc kubenswrapper[4824]: I1211 03:17:10.512528 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rwg7w\" (UniqueName: \"kubernetes.io/projected/102d6abf-7e9f-4a69-8027-5b0dd00d257a-kube-api-access-rwg7w\") pod \"redhat-marketplace-k9zwd\" (UID: \"102d6abf-7e9f-4a69-8027-5b0dd00d257a\") " pod="openshift-marketplace/redhat-marketplace-k9zwd" Dec 11 03:17:10 crc kubenswrapper[4824]: I1211 03:17:10.613954 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rwg7w\" (UniqueName: \"kubernetes.io/projected/102d6abf-7e9f-4a69-8027-5b0dd00d257a-kube-api-access-rwg7w\") pod \"redhat-marketplace-k9zwd\" (UID: \"102d6abf-7e9f-4a69-8027-5b0dd00d257a\") " pod="openshift-marketplace/redhat-marketplace-k9zwd" Dec 11 03:17:10 crc kubenswrapper[4824]: I1211 03:17:10.614044 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/102d6abf-7e9f-4a69-8027-5b0dd00d257a-utilities\") pod \"redhat-marketplace-k9zwd\" (UID: \"102d6abf-7e9f-4a69-8027-5b0dd00d257a\") " pod="openshift-marketplace/redhat-marketplace-k9zwd" Dec 11 03:17:10 crc kubenswrapper[4824]: I1211 03:17:10.614135 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/102d6abf-7e9f-4a69-8027-5b0dd00d257a-catalog-content\") pod \"redhat-marketplace-k9zwd\" (UID: \"102d6abf-7e9f-4a69-8027-5b0dd00d257a\") " pod="openshift-marketplace/redhat-marketplace-k9zwd" Dec 11 03:17:10 crc kubenswrapper[4824]: I1211 03:17:10.614564 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/102d6abf-7e9f-4a69-8027-5b0dd00d257a-catalog-content\") pod \"redhat-marketplace-k9zwd\" (UID: \"102d6abf-7e9f-4a69-8027-5b0dd00d257a\") " pod="openshift-marketplace/redhat-marketplace-k9zwd" Dec 11 03:17:10 crc kubenswrapper[4824]: I1211 03:17:10.615060 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/102d6abf-7e9f-4a69-8027-5b0dd00d257a-utilities\") pod \"redhat-marketplace-k9zwd\" (UID: \"102d6abf-7e9f-4a69-8027-5b0dd00d257a\") " pod="openshift-marketplace/redhat-marketplace-k9zwd" Dec 11 03:17:10 crc kubenswrapper[4824]: I1211 03:17:10.960784 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rwg7w\" (UniqueName: \"kubernetes.io/projected/102d6abf-7e9f-4a69-8027-5b0dd00d257a-kube-api-access-rwg7w\") pod \"redhat-marketplace-k9zwd\" (UID: \"102d6abf-7e9f-4a69-8027-5b0dd00d257a\") " pod="openshift-marketplace/redhat-marketplace-k9zwd" Dec 11 03:17:11 crc kubenswrapper[4824]: I1211 03:17:11.017633 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k9zwd" Dec 11 03:17:11 crc kubenswrapper[4824]: I1211 03:17:11.510818 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-k9zwd"] Dec 11 03:17:12 crc kubenswrapper[4824]: I1211 03:17:12.154324 4824 generic.go:334] "Generic (PLEG): container finished" podID="102d6abf-7e9f-4a69-8027-5b0dd00d257a" containerID="3a9c3c17606dc38603a4aacfc44ff8d2e0413e7bc41249ced46b6a21a7dd1010" exitCode=0 Dec 11 03:17:12 crc kubenswrapper[4824]: I1211 03:17:12.154409 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k9zwd" event={"ID":"102d6abf-7e9f-4a69-8027-5b0dd00d257a","Type":"ContainerDied","Data":"3a9c3c17606dc38603a4aacfc44ff8d2e0413e7bc41249ced46b6a21a7dd1010"} Dec 11 03:17:12 crc kubenswrapper[4824]: I1211 03:17:12.155226 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k9zwd" event={"ID":"102d6abf-7e9f-4a69-8027-5b0dd00d257a","Type":"ContainerStarted","Data":"fbe36fb922bf84599608e9ff6d9a28f14741341f1e2e5a91ec7b43369c9573db"} Dec 11 03:17:14 crc kubenswrapper[4824]: I1211 03:17:14.199695 4824 generic.go:334] "Generic (PLEG): container finished" podID="102d6abf-7e9f-4a69-8027-5b0dd00d257a" containerID="126844a75fa378bd6492511b5e954caa2ae380ff7bc91e0c5bbeb49bf75d6c18" exitCode=0 Dec 11 03:17:14 crc kubenswrapper[4824]: I1211 03:17:14.199811 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k9zwd" event={"ID":"102d6abf-7e9f-4a69-8027-5b0dd00d257a","Type":"ContainerDied","Data":"126844a75fa378bd6492511b5e954caa2ae380ff7bc91e0c5bbeb49bf75d6c18"} Dec 11 03:17:15 crc kubenswrapper[4824]: I1211 03:17:15.218232 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k9zwd" event={"ID":"102d6abf-7e9f-4a69-8027-5b0dd00d257a","Type":"ContainerStarted","Data":"89d2bf445e62f67636a979ea2705b637573ffee4e9775cd976fcb0e1a47871f7"} Dec 11 03:17:15 crc kubenswrapper[4824]: I1211 03:17:15.243250 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-k9zwd" podStartSLOduration=2.695499469 podStartE2EDuration="5.243231092s" podCreationTimestamp="2025-12-11 03:17:10 +0000 UTC" firstStartedPulling="2025-12-11 03:17:12.157336218 +0000 UTC m=+4573.846373647" lastFinishedPulling="2025-12-11 03:17:14.705067881 +0000 UTC m=+4576.394105270" observedRunningTime="2025-12-11 03:17:15.236525974 +0000 UTC m=+4576.925563393" watchObservedRunningTime="2025-12-11 03:17:15.243231092 +0000 UTC m=+4576.932268481" Dec 11 03:17:21 crc kubenswrapper[4824]: I1211 03:17:21.018461 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-k9zwd" Dec 11 03:17:21 crc kubenswrapper[4824]: I1211 03:17:21.019107 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-k9zwd" Dec 11 03:17:21 crc kubenswrapper[4824]: I1211 03:17:21.097895 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-k9zwd" Dec 11 03:17:21 crc kubenswrapper[4824]: I1211 03:17:21.346742 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-k9zwd" Dec 11 03:17:21 crc kubenswrapper[4824]: I1211 03:17:21.430596 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-k9zwd"] Dec 11 03:17:23 crc kubenswrapper[4824]: I1211 03:17:23.308950 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-k9zwd" podUID="102d6abf-7e9f-4a69-8027-5b0dd00d257a" containerName="registry-server" containerID="cri-o://89d2bf445e62f67636a979ea2705b637573ffee4e9775cd976fcb0e1a47871f7" gracePeriod=2 Dec 11 03:17:23 crc kubenswrapper[4824]: I1211 03:17:23.805568 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k9zwd" Dec 11 03:17:23 crc kubenswrapper[4824]: I1211 03:17:23.877831 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/102d6abf-7e9f-4a69-8027-5b0dd00d257a-catalog-content\") pod \"102d6abf-7e9f-4a69-8027-5b0dd00d257a\" (UID: \"102d6abf-7e9f-4a69-8027-5b0dd00d257a\") " Dec 11 03:17:23 crc kubenswrapper[4824]: I1211 03:17:23.878130 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/102d6abf-7e9f-4a69-8027-5b0dd00d257a-utilities\") pod \"102d6abf-7e9f-4a69-8027-5b0dd00d257a\" (UID: \"102d6abf-7e9f-4a69-8027-5b0dd00d257a\") " Dec 11 03:17:23 crc kubenswrapper[4824]: I1211 03:17:23.878209 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rwg7w\" (UniqueName: \"kubernetes.io/projected/102d6abf-7e9f-4a69-8027-5b0dd00d257a-kube-api-access-rwg7w\") pod \"102d6abf-7e9f-4a69-8027-5b0dd00d257a\" (UID: \"102d6abf-7e9f-4a69-8027-5b0dd00d257a\") " Dec 11 03:17:23 crc kubenswrapper[4824]: I1211 03:17:23.879605 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/102d6abf-7e9f-4a69-8027-5b0dd00d257a-utilities" (OuterVolumeSpecName: "utilities") pod "102d6abf-7e9f-4a69-8027-5b0dd00d257a" (UID: "102d6abf-7e9f-4a69-8027-5b0dd00d257a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 03:17:23 crc kubenswrapper[4824]: I1211 03:17:23.885374 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/102d6abf-7e9f-4a69-8027-5b0dd00d257a-kube-api-access-rwg7w" (OuterVolumeSpecName: "kube-api-access-rwg7w") pod "102d6abf-7e9f-4a69-8027-5b0dd00d257a" (UID: "102d6abf-7e9f-4a69-8027-5b0dd00d257a"). InnerVolumeSpecName "kube-api-access-rwg7w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 03:17:23 crc kubenswrapper[4824]: I1211 03:17:23.912383 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/102d6abf-7e9f-4a69-8027-5b0dd00d257a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "102d6abf-7e9f-4a69-8027-5b0dd00d257a" (UID: "102d6abf-7e9f-4a69-8027-5b0dd00d257a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 03:17:23 crc kubenswrapper[4824]: I1211 03:17:23.980609 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rwg7w\" (UniqueName: \"kubernetes.io/projected/102d6abf-7e9f-4a69-8027-5b0dd00d257a-kube-api-access-rwg7w\") on node \"crc\" DevicePath \"\"" Dec 11 03:17:23 crc kubenswrapper[4824]: I1211 03:17:23.980632 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/102d6abf-7e9f-4a69-8027-5b0dd00d257a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 03:17:23 crc kubenswrapper[4824]: I1211 03:17:23.980641 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/102d6abf-7e9f-4a69-8027-5b0dd00d257a-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 03:17:24 crc kubenswrapper[4824]: I1211 03:17:24.324255 4824 generic.go:334] "Generic (PLEG): container finished" podID="102d6abf-7e9f-4a69-8027-5b0dd00d257a" containerID="89d2bf445e62f67636a979ea2705b637573ffee4e9775cd976fcb0e1a47871f7" exitCode=0 Dec 11 03:17:24 crc kubenswrapper[4824]: I1211 03:17:24.324305 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k9zwd" event={"ID":"102d6abf-7e9f-4a69-8027-5b0dd00d257a","Type":"ContainerDied","Data":"89d2bf445e62f67636a979ea2705b637573ffee4e9775cd976fcb0e1a47871f7"} Dec 11 03:17:24 crc kubenswrapper[4824]: I1211 03:17:24.324346 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k9zwd" event={"ID":"102d6abf-7e9f-4a69-8027-5b0dd00d257a","Type":"ContainerDied","Data":"fbe36fb922bf84599608e9ff6d9a28f14741341f1e2e5a91ec7b43369c9573db"} Dec 11 03:17:24 crc kubenswrapper[4824]: I1211 03:17:24.324364 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k9zwd" Dec 11 03:17:24 crc kubenswrapper[4824]: I1211 03:17:24.324370 4824 scope.go:117] "RemoveContainer" containerID="89d2bf445e62f67636a979ea2705b637573ffee4e9775cd976fcb0e1a47871f7" Dec 11 03:17:24 crc kubenswrapper[4824]: I1211 03:17:24.365909 4824 scope.go:117] "RemoveContainer" containerID="126844a75fa378bd6492511b5e954caa2ae380ff7bc91e0c5bbeb49bf75d6c18" Dec 11 03:17:24 crc kubenswrapper[4824]: I1211 03:17:24.393781 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-k9zwd"] Dec 11 03:17:24 crc kubenswrapper[4824]: I1211 03:17:24.405477 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-k9zwd"] Dec 11 03:17:24 crc kubenswrapper[4824]: I1211 03:17:24.581346 4824 scope.go:117] "RemoveContainer" containerID="3a9c3c17606dc38603a4aacfc44ff8d2e0413e7bc41249ced46b6a21a7dd1010" Dec 11 03:17:24 crc kubenswrapper[4824]: I1211 03:17:24.641471 4824 scope.go:117] "RemoveContainer" containerID="89d2bf445e62f67636a979ea2705b637573ffee4e9775cd976fcb0e1a47871f7" Dec 11 03:17:24 crc kubenswrapper[4824]: E1211 03:17:24.642569 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"89d2bf445e62f67636a979ea2705b637573ffee4e9775cd976fcb0e1a47871f7\": container with ID starting with 89d2bf445e62f67636a979ea2705b637573ffee4e9775cd976fcb0e1a47871f7 not found: ID does not exist" containerID="89d2bf445e62f67636a979ea2705b637573ffee4e9775cd976fcb0e1a47871f7" Dec 11 03:17:24 crc kubenswrapper[4824]: I1211 03:17:24.642610 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"89d2bf445e62f67636a979ea2705b637573ffee4e9775cd976fcb0e1a47871f7"} err="failed to get container status \"89d2bf445e62f67636a979ea2705b637573ffee4e9775cd976fcb0e1a47871f7\": rpc error: code = NotFound desc = could not find container \"89d2bf445e62f67636a979ea2705b637573ffee4e9775cd976fcb0e1a47871f7\": container with ID starting with 89d2bf445e62f67636a979ea2705b637573ffee4e9775cd976fcb0e1a47871f7 not found: ID does not exist" Dec 11 03:17:24 crc kubenswrapper[4824]: I1211 03:17:24.642636 4824 scope.go:117] "RemoveContainer" containerID="126844a75fa378bd6492511b5e954caa2ae380ff7bc91e0c5bbeb49bf75d6c18" Dec 11 03:17:24 crc kubenswrapper[4824]: E1211 03:17:24.642981 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"126844a75fa378bd6492511b5e954caa2ae380ff7bc91e0c5bbeb49bf75d6c18\": container with ID starting with 126844a75fa378bd6492511b5e954caa2ae380ff7bc91e0c5bbeb49bf75d6c18 not found: ID does not exist" containerID="126844a75fa378bd6492511b5e954caa2ae380ff7bc91e0c5bbeb49bf75d6c18" Dec 11 03:17:24 crc kubenswrapper[4824]: I1211 03:17:24.643013 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"126844a75fa378bd6492511b5e954caa2ae380ff7bc91e0c5bbeb49bf75d6c18"} err="failed to get container status \"126844a75fa378bd6492511b5e954caa2ae380ff7bc91e0c5bbeb49bf75d6c18\": rpc error: code = NotFound desc = could not find container \"126844a75fa378bd6492511b5e954caa2ae380ff7bc91e0c5bbeb49bf75d6c18\": container with ID starting with 126844a75fa378bd6492511b5e954caa2ae380ff7bc91e0c5bbeb49bf75d6c18 not found: ID does not exist" Dec 11 03:17:24 crc kubenswrapper[4824]: I1211 03:17:24.643033 4824 scope.go:117] "RemoveContainer" containerID="3a9c3c17606dc38603a4aacfc44ff8d2e0413e7bc41249ced46b6a21a7dd1010" Dec 11 03:17:24 crc kubenswrapper[4824]: E1211 03:17:24.645192 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a9c3c17606dc38603a4aacfc44ff8d2e0413e7bc41249ced46b6a21a7dd1010\": container with ID starting with 3a9c3c17606dc38603a4aacfc44ff8d2e0413e7bc41249ced46b6a21a7dd1010 not found: ID does not exist" containerID="3a9c3c17606dc38603a4aacfc44ff8d2e0413e7bc41249ced46b6a21a7dd1010" Dec 11 03:17:24 crc kubenswrapper[4824]: I1211 03:17:24.645254 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a9c3c17606dc38603a4aacfc44ff8d2e0413e7bc41249ced46b6a21a7dd1010"} err="failed to get container status \"3a9c3c17606dc38603a4aacfc44ff8d2e0413e7bc41249ced46b6a21a7dd1010\": rpc error: code = NotFound desc = could not find container \"3a9c3c17606dc38603a4aacfc44ff8d2e0413e7bc41249ced46b6a21a7dd1010\": container with ID starting with 3a9c3c17606dc38603a4aacfc44ff8d2e0413e7bc41249ced46b6a21a7dd1010 not found: ID does not exist" Dec 11 03:17:24 crc kubenswrapper[4824]: I1211 03:17:24.658586 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="102d6abf-7e9f-4a69-8027-5b0dd00d257a" path="/var/lib/kubelet/pods/102d6abf-7e9f-4a69-8027-5b0dd00d257a/volumes" Dec 11 03:18:04 crc kubenswrapper[4824]: I1211 03:18:04.402744 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-k6z4z"] Dec 11 03:18:04 crc kubenswrapper[4824]: E1211 03:18:04.404322 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="102d6abf-7e9f-4a69-8027-5b0dd00d257a" containerName="extract-utilities" Dec 11 03:18:04 crc kubenswrapper[4824]: I1211 03:18:04.404349 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="102d6abf-7e9f-4a69-8027-5b0dd00d257a" containerName="extract-utilities" Dec 11 03:18:04 crc kubenswrapper[4824]: E1211 03:18:04.404393 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="102d6abf-7e9f-4a69-8027-5b0dd00d257a" containerName="registry-server" Dec 11 03:18:04 crc kubenswrapper[4824]: I1211 03:18:04.404408 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="102d6abf-7e9f-4a69-8027-5b0dd00d257a" containerName="registry-server" Dec 11 03:18:04 crc kubenswrapper[4824]: E1211 03:18:04.404438 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="102d6abf-7e9f-4a69-8027-5b0dd00d257a" containerName="extract-content" Dec 11 03:18:04 crc kubenswrapper[4824]: I1211 03:18:04.404454 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="102d6abf-7e9f-4a69-8027-5b0dd00d257a" containerName="extract-content" Dec 11 03:18:04 crc kubenswrapper[4824]: I1211 03:18:04.404841 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="102d6abf-7e9f-4a69-8027-5b0dd00d257a" containerName="registry-server" Dec 11 03:18:04 crc kubenswrapper[4824]: I1211 03:18:04.407975 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k6z4z" Dec 11 03:18:04 crc kubenswrapper[4824]: I1211 03:18:04.415676 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-k6z4z"] Dec 11 03:18:04 crc kubenswrapper[4824]: I1211 03:18:04.565216 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b75794c-f41d-442d-8d2b-674841a8f02e-catalog-content\") pod \"community-operators-k6z4z\" (UID: \"2b75794c-f41d-442d-8d2b-674841a8f02e\") " pod="openshift-marketplace/community-operators-k6z4z" Dec 11 03:18:04 crc kubenswrapper[4824]: I1211 03:18:04.565264 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2wrlw\" (UniqueName: \"kubernetes.io/projected/2b75794c-f41d-442d-8d2b-674841a8f02e-kube-api-access-2wrlw\") pod \"community-operators-k6z4z\" (UID: \"2b75794c-f41d-442d-8d2b-674841a8f02e\") " pod="openshift-marketplace/community-operators-k6z4z" Dec 11 03:18:04 crc kubenswrapper[4824]: I1211 03:18:04.565694 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b75794c-f41d-442d-8d2b-674841a8f02e-utilities\") pod \"community-operators-k6z4z\" (UID: \"2b75794c-f41d-442d-8d2b-674841a8f02e\") " pod="openshift-marketplace/community-operators-k6z4z" Dec 11 03:18:04 crc kubenswrapper[4824]: I1211 03:18:04.667432 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b75794c-f41d-442d-8d2b-674841a8f02e-catalog-content\") pod \"community-operators-k6z4z\" (UID: \"2b75794c-f41d-442d-8d2b-674841a8f02e\") " pod="openshift-marketplace/community-operators-k6z4z" Dec 11 03:18:04 crc kubenswrapper[4824]: I1211 03:18:04.667476 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2wrlw\" (UniqueName: \"kubernetes.io/projected/2b75794c-f41d-442d-8d2b-674841a8f02e-kube-api-access-2wrlw\") pod \"community-operators-k6z4z\" (UID: \"2b75794c-f41d-442d-8d2b-674841a8f02e\") " pod="openshift-marketplace/community-operators-k6z4z" Dec 11 03:18:04 crc kubenswrapper[4824]: I1211 03:18:04.667616 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b75794c-f41d-442d-8d2b-674841a8f02e-utilities\") pod \"community-operators-k6z4z\" (UID: \"2b75794c-f41d-442d-8d2b-674841a8f02e\") " pod="openshift-marketplace/community-operators-k6z4z" Dec 11 03:18:04 crc kubenswrapper[4824]: I1211 03:18:04.668070 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b75794c-f41d-442d-8d2b-674841a8f02e-utilities\") pod \"community-operators-k6z4z\" (UID: \"2b75794c-f41d-442d-8d2b-674841a8f02e\") " pod="openshift-marketplace/community-operators-k6z4z" Dec 11 03:18:04 crc kubenswrapper[4824]: I1211 03:18:04.668524 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b75794c-f41d-442d-8d2b-674841a8f02e-catalog-content\") pod \"community-operators-k6z4z\" (UID: \"2b75794c-f41d-442d-8d2b-674841a8f02e\") " pod="openshift-marketplace/community-operators-k6z4z" Dec 11 03:18:04 crc kubenswrapper[4824]: I1211 03:18:04.687102 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2wrlw\" (UniqueName: \"kubernetes.io/projected/2b75794c-f41d-442d-8d2b-674841a8f02e-kube-api-access-2wrlw\") pod \"community-operators-k6z4z\" (UID: \"2b75794c-f41d-442d-8d2b-674841a8f02e\") " pod="openshift-marketplace/community-operators-k6z4z" Dec 11 03:18:04 crc kubenswrapper[4824]: I1211 03:18:04.740436 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k6z4z" Dec 11 03:18:05 crc kubenswrapper[4824]: I1211 03:18:05.260979 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-k6z4z"] Dec 11 03:18:05 crc kubenswrapper[4824]: I1211 03:18:05.875407 4824 generic.go:334] "Generic (PLEG): container finished" podID="2b75794c-f41d-442d-8d2b-674841a8f02e" containerID="6bef3134006b3e8d708c8050a171ff3784987d0975f3de1d62a94684745856b0" exitCode=0 Dec 11 03:18:05 crc kubenswrapper[4824]: I1211 03:18:05.875589 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k6z4z" event={"ID":"2b75794c-f41d-442d-8d2b-674841a8f02e","Type":"ContainerDied","Data":"6bef3134006b3e8d708c8050a171ff3784987d0975f3de1d62a94684745856b0"} Dec 11 03:18:05 crc kubenswrapper[4824]: I1211 03:18:05.875789 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k6z4z" event={"ID":"2b75794c-f41d-442d-8d2b-674841a8f02e","Type":"ContainerStarted","Data":"348d1da4706eda3a613107965cb6c69df7a3c54a920fe6ca36ac012e71acb60c"} Dec 11 03:18:07 crc kubenswrapper[4824]: I1211 03:18:07.903096 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k6z4z" event={"ID":"2b75794c-f41d-442d-8d2b-674841a8f02e","Type":"ContainerStarted","Data":"483ce8d257588ef9b69792a991f62fccdd4b5bf0eadbc5d47b45614121ab8748"} Dec 11 03:18:08 crc kubenswrapper[4824]: I1211 03:18:08.918746 4824 generic.go:334] "Generic (PLEG): container finished" podID="2b75794c-f41d-442d-8d2b-674841a8f02e" containerID="483ce8d257588ef9b69792a991f62fccdd4b5bf0eadbc5d47b45614121ab8748" exitCode=0 Dec 11 03:18:08 crc kubenswrapper[4824]: I1211 03:18:08.918824 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k6z4z" event={"ID":"2b75794c-f41d-442d-8d2b-674841a8f02e","Type":"ContainerDied","Data":"483ce8d257588ef9b69792a991f62fccdd4b5bf0eadbc5d47b45614121ab8748"} Dec 11 03:18:10 crc kubenswrapper[4824]: I1211 03:18:10.954449 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k6z4z" event={"ID":"2b75794c-f41d-442d-8d2b-674841a8f02e","Type":"ContainerStarted","Data":"8a5e3a54242c369530e676e6e95e4eb46e70c8dc0bc9d47199ef89850a758d0d"} Dec 11 03:18:10 crc kubenswrapper[4824]: I1211 03:18:10.992757 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-k6z4z" podStartSLOduration=3.491156813 podStartE2EDuration="6.992728987s" podCreationTimestamp="2025-12-11 03:18:04 +0000 UTC" firstStartedPulling="2025-12-11 03:18:05.877869092 +0000 UTC m=+4627.566906491" lastFinishedPulling="2025-12-11 03:18:09.379441246 +0000 UTC m=+4631.068478665" observedRunningTime="2025-12-11 03:18:10.986286315 +0000 UTC m=+4632.675323704" watchObservedRunningTime="2025-12-11 03:18:10.992728987 +0000 UTC m=+4632.681766406" Dec 11 03:18:12 crc kubenswrapper[4824]: I1211 03:18:12.446787 4824 scope.go:117] "RemoveContainer" containerID="d648bb6134869e6b36beef509d5393ef4854b7351ae0a020516e478a5d56ddd3" Dec 11 03:18:12 crc kubenswrapper[4824]: I1211 03:18:12.472964 4824 scope.go:117] "RemoveContainer" containerID="5d0b0836a4a3dd0a08c418f770678040e24414d98b8c2f2ed5032d3a008794d5" Dec 11 03:18:14 crc kubenswrapper[4824]: I1211 03:18:14.741311 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-k6z4z" Dec 11 03:18:14 crc kubenswrapper[4824]: I1211 03:18:14.741735 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-k6z4z" Dec 11 03:18:14 crc kubenswrapper[4824]: I1211 03:18:14.818688 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-k6z4z" Dec 11 03:18:25 crc kubenswrapper[4824]: I1211 03:18:25.033642 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-k6z4z" Dec 11 03:18:25 crc kubenswrapper[4824]: I1211 03:18:25.131326 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-k6z4z"] Dec 11 03:18:25 crc kubenswrapper[4824]: I1211 03:18:25.132383 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-k6z4z" podUID="2b75794c-f41d-442d-8d2b-674841a8f02e" containerName="registry-server" containerID="cri-o://8a5e3a54242c369530e676e6e95e4eb46e70c8dc0bc9d47199ef89850a758d0d" gracePeriod=2 Dec 11 03:18:25 crc kubenswrapper[4824]: E1211 03:18:25.182430 4824 log.go:32] "ExecSync cmd from runtime service failed" err=< Dec 11 03:18:25 crc kubenswrapper[4824]: rpc error: code = Unknown desc = command error: setns `mnt`: Bad file descriptor Dec 11 03:18:25 crc kubenswrapper[4824]: fail startup Dec 11 03:18:25 crc kubenswrapper[4824]: , stdout: , stderr: , exit code -1 Dec 11 03:18:25 crc kubenswrapper[4824]: > containerID="8a5e3a54242c369530e676e6e95e4eb46e70c8dc0bc9d47199ef89850a758d0d" cmd=["grpc_health_probe","-addr=:50051"] Dec 11 03:18:25 crc kubenswrapper[4824]: E1211 03:18:25.184040 4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 8a5e3a54242c369530e676e6e95e4eb46e70c8dc0bc9d47199ef89850a758d0d is running failed: container process not found" containerID="8a5e3a54242c369530e676e6e95e4eb46e70c8dc0bc9d47199ef89850a758d0d" cmd=["grpc_health_probe","-addr=:50051"] Dec 11 03:18:25 crc kubenswrapper[4824]: E1211 03:18:25.187473 4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 8a5e3a54242c369530e676e6e95e4eb46e70c8dc0bc9d47199ef89850a758d0d is running failed: container process not found" containerID="8a5e3a54242c369530e676e6e95e4eb46e70c8dc0bc9d47199ef89850a758d0d" cmd=["grpc_health_probe","-addr=:50051"] Dec 11 03:18:25 crc kubenswrapper[4824]: E1211 03:18:25.187571 4824 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 8a5e3a54242c369530e676e6e95e4eb46e70c8dc0bc9d47199ef89850a758d0d is running failed: container process not found" probeType="Liveness" pod="openshift-marketplace/community-operators-k6z4z" podUID="2b75794c-f41d-442d-8d2b-674841a8f02e" containerName="registry-server" Dec 11 03:18:26 crc kubenswrapper[4824]: I1211 03:18:26.136474 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k6z4z" Dec 11 03:18:26 crc kubenswrapper[4824]: I1211 03:18:26.147757 4824 generic.go:334] "Generic (PLEG): container finished" podID="2b75794c-f41d-442d-8d2b-674841a8f02e" containerID="8a5e3a54242c369530e676e6e95e4eb46e70c8dc0bc9d47199ef89850a758d0d" exitCode=0 Dec 11 03:18:26 crc kubenswrapper[4824]: I1211 03:18:26.147798 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k6z4z" event={"ID":"2b75794c-f41d-442d-8d2b-674841a8f02e","Type":"ContainerDied","Data":"8a5e3a54242c369530e676e6e95e4eb46e70c8dc0bc9d47199ef89850a758d0d"} Dec 11 03:18:26 crc kubenswrapper[4824]: I1211 03:18:26.147828 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k6z4z" event={"ID":"2b75794c-f41d-442d-8d2b-674841a8f02e","Type":"ContainerDied","Data":"348d1da4706eda3a613107965cb6c69df7a3c54a920fe6ca36ac012e71acb60c"} Dec 11 03:18:26 crc kubenswrapper[4824]: I1211 03:18:26.147845 4824 scope.go:117] "RemoveContainer" containerID="8a5e3a54242c369530e676e6e95e4eb46e70c8dc0bc9d47199ef89850a758d0d" Dec 11 03:18:26 crc kubenswrapper[4824]: I1211 03:18:26.147841 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k6z4z" Dec 11 03:18:26 crc kubenswrapper[4824]: I1211 03:18:26.169060 4824 scope.go:117] "RemoveContainer" containerID="483ce8d257588ef9b69792a991f62fccdd4b5bf0eadbc5d47b45614121ab8748" Dec 11 03:18:26 crc kubenswrapper[4824]: I1211 03:18:26.188335 4824 scope.go:117] "RemoveContainer" containerID="6bef3134006b3e8d708c8050a171ff3784987d0975f3de1d62a94684745856b0" Dec 11 03:18:26 crc kubenswrapper[4824]: I1211 03:18:26.245856 4824 scope.go:117] "RemoveContainer" containerID="8a5e3a54242c369530e676e6e95e4eb46e70c8dc0bc9d47199ef89850a758d0d" Dec 11 03:18:26 crc kubenswrapper[4824]: E1211 03:18:26.246565 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a5e3a54242c369530e676e6e95e4eb46e70c8dc0bc9d47199ef89850a758d0d\": container with ID starting with 8a5e3a54242c369530e676e6e95e4eb46e70c8dc0bc9d47199ef89850a758d0d not found: ID does not exist" containerID="8a5e3a54242c369530e676e6e95e4eb46e70c8dc0bc9d47199ef89850a758d0d" Dec 11 03:18:26 crc kubenswrapper[4824]: I1211 03:18:26.246608 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a5e3a54242c369530e676e6e95e4eb46e70c8dc0bc9d47199ef89850a758d0d"} err="failed to get container status \"8a5e3a54242c369530e676e6e95e4eb46e70c8dc0bc9d47199ef89850a758d0d\": rpc error: code = NotFound desc = could not find container \"8a5e3a54242c369530e676e6e95e4eb46e70c8dc0bc9d47199ef89850a758d0d\": container with ID starting with 8a5e3a54242c369530e676e6e95e4eb46e70c8dc0bc9d47199ef89850a758d0d not found: ID does not exist" Dec 11 03:18:26 crc kubenswrapper[4824]: I1211 03:18:26.246642 4824 scope.go:117] "RemoveContainer" containerID="483ce8d257588ef9b69792a991f62fccdd4b5bf0eadbc5d47b45614121ab8748" Dec 11 03:18:26 crc kubenswrapper[4824]: E1211 03:18:26.247087 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"483ce8d257588ef9b69792a991f62fccdd4b5bf0eadbc5d47b45614121ab8748\": container with ID starting with 483ce8d257588ef9b69792a991f62fccdd4b5bf0eadbc5d47b45614121ab8748 not found: ID does not exist" containerID="483ce8d257588ef9b69792a991f62fccdd4b5bf0eadbc5d47b45614121ab8748" Dec 11 03:18:26 crc kubenswrapper[4824]: I1211 03:18:26.247135 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"483ce8d257588ef9b69792a991f62fccdd4b5bf0eadbc5d47b45614121ab8748"} err="failed to get container status \"483ce8d257588ef9b69792a991f62fccdd4b5bf0eadbc5d47b45614121ab8748\": rpc error: code = NotFound desc = could not find container \"483ce8d257588ef9b69792a991f62fccdd4b5bf0eadbc5d47b45614121ab8748\": container with ID starting with 483ce8d257588ef9b69792a991f62fccdd4b5bf0eadbc5d47b45614121ab8748 not found: ID does not exist" Dec 11 03:18:26 crc kubenswrapper[4824]: I1211 03:18:26.247161 4824 scope.go:117] "RemoveContainer" containerID="6bef3134006b3e8d708c8050a171ff3784987d0975f3de1d62a94684745856b0" Dec 11 03:18:26 crc kubenswrapper[4824]: E1211 03:18:26.247560 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6bef3134006b3e8d708c8050a171ff3784987d0975f3de1d62a94684745856b0\": container with ID starting with 6bef3134006b3e8d708c8050a171ff3784987d0975f3de1d62a94684745856b0 not found: ID does not exist" containerID="6bef3134006b3e8d708c8050a171ff3784987d0975f3de1d62a94684745856b0" Dec 11 03:18:26 crc kubenswrapper[4824]: I1211 03:18:26.247598 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6bef3134006b3e8d708c8050a171ff3784987d0975f3de1d62a94684745856b0"} err="failed to get container status \"6bef3134006b3e8d708c8050a171ff3784987d0975f3de1d62a94684745856b0\": rpc error: code = NotFound desc = could not find container \"6bef3134006b3e8d708c8050a171ff3784987d0975f3de1d62a94684745856b0\": container with ID starting with 6bef3134006b3e8d708c8050a171ff3784987d0975f3de1d62a94684745856b0 not found: ID does not exist" Dec 11 03:18:26 crc kubenswrapper[4824]: I1211 03:18:26.301524 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b75794c-f41d-442d-8d2b-674841a8f02e-utilities\") pod \"2b75794c-f41d-442d-8d2b-674841a8f02e\" (UID: \"2b75794c-f41d-442d-8d2b-674841a8f02e\") " Dec 11 03:18:26 crc kubenswrapper[4824]: I1211 03:18:26.301593 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2wrlw\" (UniqueName: \"kubernetes.io/projected/2b75794c-f41d-442d-8d2b-674841a8f02e-kube-api-access-2wrlw\") pod \"2b75794c-f41d-442d-8d2b-674841a8f02e\" (UID: \"2b75794c-f41d-442d-8d2b-674841a8f02e\") " Dec 11 03:18:26 crc kubenswrapper[4824]: I1211 03:18:26.301810 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b75794c-f41d-442d-8d2b-674841a8f02e-catalog-content\") pod \"2b75794c-f41d-442d-8d2b-674841a8f02e\" (UID: \"2b75794c-f41d-442d-8d2b-674841a8f02e\") " Dec 11 03:18:26 crc kubenswrapper[4824]: I1211 03:18:26.302295 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b75794c-f41d-442d-8d2b-674841a8f02e-utilities" (OuterVolumeSpecName: "utilities") pod "2b75794c-f41d-442d-8d2b-674841a8f02e" (UID: "2b75794c-f41d-442d-8d2b-674841a8f02e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 03:18:26 crc kubenswrapper[4824]: I1211 03:18:26.311674 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b75794c-f41d-442d-8d2b-674841a8f02e-kube-api-access-2wrlw" (OuterVolumeSpecName: "kube-api-access-2wrlw") pod "2b75794c-f41d-442d-8d2b-674841a8f02e" (UID: "2b75794c-f41d-442d-8d2b-674841a8f02e"). InnerVolumeSpecName "kube-api-access-2wrlw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 03:18:26 crc kubenswrapper[4824]: I1211 03:18:26.321082 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b75794c-f41d-442d-8d2b-674841a8f02e-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 03:18:26 crc kubenswrapper[4824]: I1211 03:18:26.325558 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2wrlw\" (UniqueName: \"kubernetes.io/projected/2b75794c-f41d-442d-8d2b-674841a8f02e-kube-api-access-2wrlw\") on node \"crc\" DevicePath \"\"" Dec 11 03:18:26 crc kubenswrapper[4824]: I1211 03:18:26.395238 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b75794c-f41d-442d-8d2b-674841a8f02e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2b75794c-f41d-442d-8d2b-674841a8f02e" (UID: "2b75794c-f41d-442d-8d2b-674841a8f02e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 03:18:26 crc kubenswrapper[4824]: I1211 03:18:26.427690 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b75794c-f41d-442d-8d2b-674841a8f02e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 03:18:26 crc kubenswrapper[4824]: I1211 03:18:26.494290 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-k6z4z"] Dec 11 03:18:26 crc kubenswrapper[4824]: I1211 03:18:26.501779 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-k6z4z"] Dec 11 03:18:26 crc kubenswrapper[4824]: I1211 03:18:26.657198 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b75794c-f41d-442d-8d2b-674841a8f02e" path="/var/lib/kubelet/pods/2b75794c-f41d-442d-8d2b-674841a8f02e/volumes" Dec 11 03:18:58 crc kubenswrapper[4824]: I1211 03:18:58.744858 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-6rfpf"] Dec 11 03:18:58 crc kubenswrapper[4824]: E1211 03:18:58.746393 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b75794c-f41d-442d-8d2b-674841a8f02e" containerName="extract-content" Dec 11 03:18:58 crc kubenswrapper[4824]: I1211 03:18:58.746418 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b75794c-f41d-442d-8d2b-674841a8f02e" containerName="extract-content" Dec 11 03:18:58 crc kubenswrapper[4824]: E1211 03:18:58.746458 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b75794c-f41d-442d-8d2b-674841a8f02e" containerName="registry-server" Dec 11 03:18:58 crc kubenswrapper[4824]: I1211 03:18:58.746471 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b75794c-f41d-442d-8d2b-674841a8f02e" containerName="registry-server" Dec 11 03:18:58 crc kubenswrapper[4824]: E1211 03:18:58.746496 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b75794c-f41d-442d-8d2b-674841a8f02e" containerName="extract-utilities" Dec 11 03:18:58 crc kubenswrapper[4824]: I1211 03:18:58.746508 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b75794c-f41d-442d-8d2b-674841a8f02e" containerName="extract-utilities" Dec 11 03:18:58 crc kubenswrapper[4824]: I1211 03:18:58.746831 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b75794c-f41d-442d-8d2b-674841a8f02e" containerName="registry-server" Dec 11 03:18:58 crc kubenswrapper[4824]: I1211 03:18:58.756153 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6rfpf" Dec 11 03:18:58 crc kubenswrapper[4824]: I1211 03:18:58.774818 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f672252a-cc9c-4bf8-bda8-4358b25fa325-catalog-content\") pod \"redhat-operators-6rfpf\" (UID: \"f672252a-cc9c-4bf8-bda8-4358b25fa325\") " pod="openshift-marketplace/redhat-operators-6rfpf" Dec 11 03:18:58 crc kubenswrapper[4824]: I1211 03:18:58.775013 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-868zn\" (UniqueName: \"kubernetes.io/projected/f672252a-cc9c-4bf8-bda8-4358b25fa325-kube-api-access-868zn\") pod \"redhat-operators-6rfpf\" (UID: \"f672252a-cc9c-4bf8-bda8-4358b25fa325\") " pod="openshift-marketplace/redhat-operators-6rfpf" Dec 11 03:18:58 crc kubenswrapper[4824]: I1211 03:18:58.775096 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f672252a-cc9c-4bf8-bda8-4358b25fa325-utilities\") pod \"redhat-operators-6rfpf\" (UID: \"f672252a-cc9c-4bf8-bda8-4358b25fa325\") " pod="openshift-marketplace/redhat-operators-6rfpf" Dec 11 03:18:58 crc kubenswrapper[4824]: I1211 03:18:58.805739 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6rfpf"] Dec 11 03:18:58 crc kubenswrapper[4824]: I1211 03:18:58.877746 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-868zn\" (UniqueName: \"kubernetes.io/projected/f672252a-cc9c-4bf8-bda8-4358b25fa325-kube-api-access-868zn\") pod \"redhat-operators-6rfpf\" (UID: \"f672252a-cc9c-4bf8-bda8-4358b25fa325\") " pod="openshift-marketplace/redhat-operators-6rfpf" Dec 11 03:18:58 crc kubenswrapper[4824]: I1211 03:18:58.877805 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f672252a-cc9c-4bf8-bda8-4358b25fa325-utilities\") pod \"redhat-operators-6rfpf\" (UID: \"f672252a-cc9c-4bf8-bda8-4358b25fa325\") " pod="openshift-marketplace/redhat-operators-6rfpf" Dec 11 03:18:58 crc kubenswrapper[4824]: I1211 03:18:58.877913 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f672252a-cc9c-4bf8-bda8-4358b25fa325-catalog-content\") pod \"redhat-operators-6rfpf\" (UID: \"f672252a-cc9c-4bf8-bda8-4358b25fa325\") " pod="openshift-marketplace/redhat-operators-6rfpf" Dec 11 03:18:58 crc kubenswrapper[4824]: I1211 03:18:58.878518 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f672252a-cc9c-4bf8-bda8-4358b25fa325-utilities\") pod \"redhat-operators-6rfpf\" (UID: \"f672252a-cc9c-4bf8-bda8-4358b25fa325\") " pod="openshift-marketplace/redhat-operators-6rfpf" Dec 11 03:18:58 crc kubenswrapper[4824]: I1211 03:18:58.878570 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f672252a-cc9c-4bf8-bda8-4358b25fa325-catalog-content\") pod \"redhat-operators-6rfpf\" (UID: \"f672252a-cc9c-4bf8-bda8-4358b25fa325\") " pod="openshift-marketplace/redhat-operators-6rfpf" Dec 11 03:18:58 crc kubenswrapper[4824]: I1211 03:18:58.906620 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-868zn\" (UniqueName: \"kubernetes.io/projected/f672252a-cc9c-4bf8-bda8-4358b25fa325-kube-api-access-868zn\") pod \"redhat-operators-6rfpf\" (UID: \"f672252a-cc9c-4bf8-bda8-4358b25fa325\") " pod="openshift-marketplace/redhat-operators-6rfpf" Dec 11 03:18:59 crc kubenswrapper[4824]: I1211 03:18:59.084416 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6rfpf" Dec 11 03:18:59 crc kubenswrapper[4824]: I1211 03:18:59.605806 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6rfpf"] Dec 11 03:19:00 crc kubenswrapper[4824]: I1211 03:19:00.599449 4824 generic.go:334] "Generic (PLEG): container finished" podID="f672252a-cc9c-4bf8-bda8-4358b25fa325" containerID="31337dc0c896baaa6f717f723f26b41e67c9fbffebf01ba876789716417ae476" exitCode=0 Dec 11 03:19:00 crc kubenswrapper[4824]: I1211 03:19:00.599521 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6rfpf" event={"ID":"f672252a-cc9c-4bf8-bda8-4358b25fa325","Type":"ContainerDied","Data":"31337dc0c896baaa6f717f723f26b41e67c9fbffebf01ba876789716417ae476"} Dec 11 03:19:00 crc kubenswrapper[4824]: I1211 03:19:00.599870 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6rfpf" event={"ID":"f672252a-cc9c-4bf8-bda8-4358b25fa325","Type":"ContainerStarted","Data":"47f19d1982d84e4b8fd41f35937373d66d2eed41a09c91b3e659e5785c19cefd"} Dec 11 03:19:00 crc kubenswrapper[4824]: I1211 03:19:00.602439 4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 11 03:19:01 crc kubenswrapper[4824]: I1211 03:19:01.623103 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6rfpf" event={"ID":"f672252a-cc9c-4bf8-bda8-4358b25fa325","Type":"ContainerStarted","Data":"cf848ce37b3c4a626ec8b085b47110d4452b55b8124cb17e0f54377102fb80ff"} Dec 11 03:19:02 crc kubenswrapper[4824]: I1211 03:19:02.637349 4824 generic.go:334] "Generic (PLEG): container finished" podID="f672252a-cc9c-4bf8-bda8-4358b25fa325" containerID="cf848ce37b3c4a626ec8b085b47110d4452b55b8124cb17e0f54377102fb80ff" exitCode=0 Dec 11 03:19:02 crc kubenswrapper[4824]: I1211 03:19:02.648987 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6rfpf" event={"ID":"f672252a-cc9c-4bf8-bda8-4358b25fa325","Type":"ContainerDied","Data":"cf848ce37b3c4a626ec8b085b47110d4452b55b8124cb17e0f54377102fb80ff"} Dec 11 03:19:03 crc kubenswrapper[4824]: I1211 03:19:03.652355 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6rfpf" event={"ID":"f672252a-cc9c-4bf8-bda8-4358b25fa325","Type":"ContainerStarted","Data":"6e8ebef6bbe97e5ccbc8bb9895e688283ed732c14caa3330a56b0cbac3fbfedb"} Dec 11 03:19:03 crc kubenswrapper[4824]: I1211 03:19:03.681624 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-6rfpf" podStartSLOduration=3.07620485 podStartE2EDuration="5.68159866s" podCreationTimestamp="2025-12-11 03:18:58 +0000 UTC" firstStartedPulling="2025-12-11 03:19:00.602222379 +0000 UTC m=+4682.291259758" lastFinishedPulling="2025-12-11 03:19:03.207616179 +0000 UTC m=+4684.896653568" observedRunningTime="2025-12-11 03:19:03.680038109 +0000 UTC m=+4685.369075558" watchObservedRunningTime="2025-12-11 03:19:03.68159866 +0000 UTC m=+4685.370636059" Dec 11 03:19:09 crc kubenswrapper[4824]: I1211 03:19:09.085479 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-6rfpf" Dec 11 03:19:09 crc kubenswrapper[4824]: I1211 03:19:09.086133 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-6rfpf" Dec 11 03:19:10 crc kubenswrapper[4824]: I1211 03:19:10.316397 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-6rfpf" podUID="f672252a-cc9c-4bf8-bda8-4358b25fa325" containerName="registry-server" probeResult="failure" output=< Dec 11 03:19:10 crc kubenswrapper[4824]: timeout: failed to connect service ":50051" within 1s Dec 11 03:19:10 crc kubenswrapper[4824]: > Dec 11 03:19:19 crc kubenswrapper[4824]: I1211 03:19:19.843825 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-6rfpf" Dec 11 03:19:19 crc kubenswrapper[4824]: I1211 03:19:19.913072 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-6rfpf" Dec 11 03:19:20 crc kubenswrapper[4824]: I1211 03:19:20.092261 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6rfpf"] Dec 11 03:19:21 crc kubenswrapper[4824]: I1211 03:19:21.861875 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-6rfpf" podUID="f672252a-cc9c-4bf8-bda8-4358b25fa325" containerName="registry-server" containerID="cri-o://6e8ebef6bbe97e5ccbc8bb9895e688283ed732c14caa3330a56b0cbac3fbfedb" gracePeriod=2 Dec 11 03:19:22 crc kubenswrapper[4824]: I1211 03:19:22.876659 4824 generic.go:334] "Generic (PLEG): container finished" podID="f672252a-cc9c-4bf8-bda8-4358b25fa325" containerID="6e8ebef6bbe97e5ccbc8bb9895e688283ed732c14caa3330a56b0cbac3fbfedb" exitCode=0 Dec 11 03:19:22 crc kubenswrapper[4824]: I1211 03:19:22.876754 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6rfpf" event={"ID":"f672252a-cc9c-4bf8-bda8-4358b25fa325","Type":"ContainerDied","Data":"6e8ebef6bbe97e5ccbc8bb9895e688283ed732c14caa3330a56b0cbac3fbfedb"} Dec 11 03:19:22 crc kubenswrapper[4824]: I1211 03:19:22.877288 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6rfpf" event={"ID":"f672252a-cc9c-4bf8-bda8-4358b25fa325","Type":"ContainerDied","Data":"47f19d1982d84e4b8fd41f35937373d66d2eed41a09c91b3e659e5785c19cefd"} Dec 11 03:19:22 crc kubenswrapper[4824]: I1211 03:19:22.877305 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="47f19d1982d84e4b8fd41f35937373d66d2eed41a09c91b3e659e5785c19cefd" Dec 11 03:19:22 crc kubenswrapper[4824]: I1211 03:19:22.896978 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6rfpf" Dec 11 03:19:23 crc kubenswrapper[4824]: I1211 03:19:23.044756 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f672252a-cc9c-4bf8-bda8-4358b25fa325-utilities\") pod \"f672252a-cc9c-4bf8-bda8-4358b25fa325\" (UID: \"f672252a-cc9c-4bf8-bda8-4358b25fa325\") " Dec 11 03:19:23 crc kubenswrapper[4824]: I1211 03:19:23.044896 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-868zn\" (UniqueName: \"kubernetes.io/projected/f672252a-cc9c-4bf8-bda8-4358b25fa325-kube-api-access-868zn\") pod \"f672252a-cc9c-4bf8-bda8-4358b25fa325\" (UID: \"f672252a-cc9c-4bf8-bda8-4358b25fa325\") " Dec 11 03:19:23 crc kubenswrapper[4824]: I1211 03:19:23.044920 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f672252a-cc9c-4bf8-bda8-4358b25fa325-catalog-content\") pod \"f672252a-cc9c-4bf8-bda8-4358b25fa325\" (UID: \"f672252a-cc9c-4bf8-bda8-4358b25fa325\") " Dec 11 03:19:23 crc kubenswrapper[4824]: I1211 03:19:23.049766 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f672252a-cc9c-4bf8-bda8-4358b25fa325-utilities" (OuterVolumeSpecName: "utilities") pod "f672252a-cc9c-4bf8-bda8-4358b25fa325" (UID: "f672252a-cc9c-4bf8-bda8-4358b25fa325"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 03:19:23 crc kubenswrapper[4824]: I1211 03:19:23.054728 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f672252a-cc9c-4bf8-bda8-4358b25fa325-kube-api-access-868zn" (OuterVolumeSpecName: "kube-api-access-868zn") pod "f672252a-cc9c-4bf8-bda8-4358b25fa325" (UID: "f672252a-cc9c-4bf8-bda8-4358b25fa325"). InnerVolumeSpecName "kube-api-access-868zn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 03:19:23 crc kubenswrapper[4824]: I1211 03:19:23.151276 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f672252a-cc9c-4bf8-bda8-4358b25fa325-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 03:19:23 crc kubenswrapper[4824]: I1211 03:19:23.151324 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-868zn\" (UniqueName: \"kubernetes.io/projected/f672252a-cc9c-4bf8-bda8-4358b25fa325-kube-api-access-868zn\") on node \"crc\" DevicePath \"\"" Dec 11 03:19:23 crc kubenswrapper[4824]: I1211 03:19:23.171345 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f672252a-cc9c-4bf8-bda8-4358b25fa325-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f672252a-cc9c-4bf8-bda8-4358b25fa325" (UID: "f672252a-cc9c-4bf8-bda8-4358b25fa325"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 03:19:23 crc kubenswrapper[4824]: I1211 03:19:23.253291 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f672252a-cc9c-4bf8-bda8-4358b25fa325-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 03:19:23 crc kubenswrapper[4824]: I1211 03:19:23.887209 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6rfpf" Dec 11 03:19:23 crc kubenswrapper[4824]: I1211 03:19:23.926999 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6rfpf"] Dec 11 03:19:23 crc kubenswrapper[4824]: I1211 03:19:23.935756 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-6rfpf"] Dec 11 03:19:24 crc kubenswrapper[4824]: I1211 03:19:24.652564 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f672252a-cc9c-4bf8-bda8-4358b25fa325" path="/var/lib/kubelet/pods/f672252a-cc9c-4bf8-bda8-4358b25fa325/volumes" Dec 11 03:19:37 crc kubenswrapper[4824]: I1211 03:19:37.250879 4824 patch_prober.go:28] interesting pod/machine-config-daemon-gx6xt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 03:19:37 crc kubenswrapper[4824]: I1211 03:19:37.251557 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 03:20:07 crc kubenswrapper[4824]: I1211 03:20:07.251097 4824 patch_prober.go:28] interesting pod/machine-config-daemon-gx6xt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 03:20:07 crc kubenswrapper[4824]: I1211 03:20:07.251856 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gx6xt" podUID="44b156e4-64a4-4d45-aa5e-9b10a862faed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515116434155024452 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015116434156017370 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015116422425016507 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015116422425015457 5ustar corecore